A small tool you run on your own computer to measure how fast your local AI models are. Results are shared anonymously with the community.
Pick your platform. No installation needed — just download and run.
On Linux or macOS, make the file executable first: chmod +x llm-benchmark*
You need Ollama, LM Studio, vLLM, or HF TGI running locally with at least one model loaded. The whole process takes about 5–15 minutes.