Model Leaderboard

Compare key performance metrics for LLM APIs.

Updated at: 7/19/2025, 6:03:14 AM

ModelProvider
Gemini 2.0 Flash
google
Google AI Studio
#31
$0.100.55s158.5T/s99.99%
DeepSeek V3 0324
deepseek
SambaNova
#14
$0.262.28s166.7T/s99.57%
Gemini 2.5 Flash
google
Google AI Studio
#7
$0.320.85s147.7T/s99.94%
Gemini 2.5 Flash Lite Preview 06-17
google
Google AI Studio
-
$0.100.33s298.6T/s99.97%
Gemini 1.5 Flash
google
Google Vertex
#95
$0.080.77s141.3T/s99.70%
GPT-4o-mini
openai
Azure
#14
$0.151.14s334.2T/s99.96%
Mistral Nemo
mistralai
Nineteen
-
$0.010.47s230.6T/s99.98%
Llama 3.1 8B Instruct
meta-llama
Cerebras
#141
$0.020.12s6064.2T/s99.98%
Qwen2.5 7B Instruct
qwen
Together
-
$0.040.26s153.7T/s99.89%
Llama 3.3 70B Instruct
meta-llama
Cerebras
#66
$0.040.18s2750.0T/s99.98%
R1 Distill Llama 70B
deepseek
Cerebras
-
$0.050.71s1626.3T/s99.97%
Gemini 2.0 Flash Lite
google
Google AI Studio
#15
$0.080.57s135.7T/s99.99%
Gemma 3 4B
google
DeepInfra
#76
$0.020.45s357.1T/s4.00%
Llama 4 Maverick
meta-llama
Cerebras
#48
$0.000.14s2063.8T/s99.99%
GPT-4.1
openai
OpenAI
-
$2.060.35s127.4T/s99.66%
Claude Sonnet 4
anthropic
Google Vertex (Europe)
-
$3.121.95s83.9T/s99.95%
Mistral Tiny
mistralai
Mistral
-
$0.250.23s272.7T/s4.00%
GPT-4o
openai
Azure
-
$2.582.15s157.3T/s99.99%
Llama 3.1 70B Instruct
meta-llama
Fireworks
-
$0.100.52s125.9T/s99.88%
Llama 3.2 3B Instruct
meta-llama
Nineteen
-
$0.000.45s459.8T/s99.98%
Gemini 1.5 Flash 8B
google
Google AI Studio
#114
$0.040.24s196.4T/s100.00%
R1 (free)
deepseek
Chutes
-
$0.002.13s59.2T/s99.98%
Claude 3.7 Sonnet
anthropic
Google Vertex (Europe)
#25
$3.121.43s74.7T/s99.94%
GPT-4.1 Mini
openai
OpenAI
-
$0.410.48s72.0T/s99.96%
R1 0528 (free)
deepseek
Chutes
#7
$0.002.77s54.0T/s97.17%
Deepseek R1 0528 Qwen3 8B
deepseek
Nineteen
-
$0.018.30s299.9T/s99.91%
Gemini 2.5 Pro
google
Google Vertex (Global)
#1
$1.332.48s83.6T/s99.51%
Kimi K2
moonshotai
Groq
#6
$0.163.16s352.4T/s97.17%