Model Leaderboard
Compare key performance metrics for LLM APIs.
Updated at: 7/19/2025, 6:03:14 AM
Model | Provider | |||||||
---|---|---|---|---|---|---|---|---|
Gemini 2.0 Flash google | Google AI Studio | #31 | $0.10 | 0.55s | 158.5T/s | 99.99% | ||
DeepSeek V3 0324 deepseek | SambaNova | #14 | $0.26 | 2.28s | 166.7T/s | 99.57% | ||
Gemini 2.5 Flash google | Google AI Studio | #7 | $0.32 | 0.85s | 147.7T/s | 99.94% | ||
Gemini 2.5 Flash Lite Preview 06-17 google | Google AI Studio | - | $0.10 | 0.33s | 298.6T/s | 99.97% | ||
Gemini 1.5 Flash google | Google Vertex | #95 | $0.08 | 0.77s | 141.3T/s | 99.70% | ||
GPT-4o-mini openai | Azure | #14 | $0.15 | 1.14s | 334.2T/s | 99.96% | ||
Mistral Nemo mistralai | Nineteen | - | $0.01 | 0.47s | 230.6T/s | 99.98% | ||
Llama 3.1 8B Instruct meta-llama | Cerebras | #141 | $0.02 | 0.12s | 6064.2T/s | 99.98% | ||
Qwen2.5 7B Instruct qwen | Together | - | $0.04 | 0.26s | 153.7T/s | 99.89% | ||
Llama 3.3 70B Instruct meta-llama | Cerebras | #66 | $0.04 | 0.18s | 2750.0T/s | 99.98% | ||
R1 Distill Llama 70B deepseek | Cerebras | - | $0.05 | 0.71s | 1626.3T/s | 99.97% | ||
Gemini 2.0 Flash Lite google | Google AI Studio | #15 | $0.08 | 0.57s | 135.7T/s | 99.99% | ||
Gemma 3 4B google | DeepInfra | #76 | $0.02 | 0.45s | 357.1T/s | 4.00% | ||
Llama 4 Maverick meta-llama | Cerebras | #48 | $0.00 | 0.14s | 2063.8T/s | 99.99% | ||
GPT-4.1 openai | OpenAI | - | $2.06 | 0.35s | 127.4T/s | 99.66% | ||
Claude Sonnet 4 anthropic | Google Vertex (Europe) | - | $3.12 | 1.95s | 83.9T/s | 99.95% | ||
Mistral Tiny mistralai | Mistral | - | $0.25 | 0.23s | 272.7T/s | 4.00% | ||
GPT-4o openai | Azure | - | $2.58 | 2.15s | 157.3T/s | 99.99% | ||
Llama 3.1 70B Instruct meta-llama | Fireworks | - | $0.10 | 0.52s | 125.9T/s | 99.88% | ||
Llama 3.2 3B Instruct meta-llama | Nineteen | - | $0.00 | 0.45s | 459.8T/s | 99.98% | ||
Gemini 1.5 Flash 8B google | Google AI Studio | #114 | $0.04 | 0.24s | 196.4T/s | 100.00% | ||
R1 (free) deepseek | Chutes | - | $0.00 | 2.13s | 59.2T/s | 99.98% | ||
Claude 3.7 Sonnet anthropic | Google Vertex (Europe) | #25 | $3.12 | 1.43s | 74.7T/s | 99.94% | ||
GPT-4.1 Mini openai | OpenAI | - | $0.41 | 0.48s | 72.0T/s | 99.96% | ||
R1 0528 (free) deepseek | Chutes | #7 | $0.00 | 2.77s | 54.0T/s | 97.17% | ||
Deepseek R1 0528 Qwen3 8B deepseek | Nineteen | - | $0.01 | 8.30s | 299.9T/s | 99.91% | ||
Gemini 2.5 Pro google | Google Vertex (Global) | #1 | $1.33 | 2.48s | 83.6T/s | 99.51% | ||
Kimi K2 moonshotai | Groq | #6 | $0.16 | 3.16s | 352.4T/s | 97.17% |