Skip to main content

LLM Bench Marker

AI Utility

LLM Bench Marker - Screenshot 1 - AI Utility project

A benchmarking tool that runs multi‑model sweeps on curated datasets with fixed prompts to identify the best cost/quality trade‑offs.

It logs tokens, latency, and quality scores per run, compares models side‑by‑side, and highlights the most suitable option for a target budget or score.

Includes run versioning and exportable reports (CSV/JSON) for analysis and sharing.

Project Info

Start:August 2025
End:
August 2025
Duration:couple weeks
Tech:4 used
Images:2 available

Technologies Used

4+ technologies
Click any tech to filter projects