YPerf

Track & Compare LLM API Performance Metrics

Last updated: 1/6/2025, 8:00:26 PM

ModelProvider
Llama 3.2 1B Instruct
meta-llama
Lepton
#147
$0.010.33s323.6T/s100.00%
Gemini Flash 1.5 8B
google
Google AI Studio
#65
$0.040.52s182.0T/s98.30%
MythoMax 13B
gryphe
Fireworks
-
$0.070.23s110.6T/s100.00%
Gemini Flash 1.5
google
Google AI Studio
#26
$0.083.92s174.4T/s100.00%
DeepSeek V3
deepseek
DeepSeek
#8
$0.141.49s54.0T/s99.72%
GPT-4o-mini
openai
OpenAI
#23
$0.150.56s69.1T/s99.33%
Mistral Nemo
mistralai
Mistral
-
$0.040.28s106.3T/s100.00%
Llama 3.1 70B Instruct
meta-llama
Fireworks
#38
$0.120.43s100.2T/s100.00%
Mistral Tiny
mistralai
Mistral
#136
$0.250.21s89.9T/s100.00%
Llama 3.3 70B Instruct
meta-llama
Together
#22
$0.120.85s157.3T/s99.98%
Qwen2.5 72B Instruct
qwen
Fireworks
#37
$0.230.35s85.0T/s100.00%
Ministral 3B
mistralai
Mistral
-
$0.040.18s119.4T/s100.00%
Ministral 8B
mistralai
Mistral
#81
$0.100.22s97.0T/s100.00%
Claude 3.5 Sonnet
anthropic
Anthropic
#8
$3.122.70s62.0T/s99.55%
Claude 3.5 Sonnet (self-moderated)
anthropic
Anthropic
-
$3.122.70s62.0T/s99.96%
Qwen2.5 Coder 32B Instruct
qwen
Fireworks
#56
$0.070.77s93.1T/s100.00%
Command R (08-2024)
cohere
Cohere
#68
$0.150.28s90.2T/s100.00%
Gemma 2 27B
google
Together
#43
$0.270.49s54.9T/s100.00%
GPT-4o-mini (2024-07-18)
openai
OpenAI
#23
$0.150.47s87.3T/s100.00%
Llama 3.1 8B Instruct
meta-llama
Together
#101
$0.020.35s302.6T/s100.00%
Hermes 3 405B Instruct
nousresearch
DeepInfra
#15
$0.810.53s19.1T/s99.90%
OpenChat 3.5 7B
openchat
Lepton
#118
$0.060.44s90.6T/s100.00%
WizardLM-2 8x22B
microsoft
Together
-
$0.500.80s61.1T/s100.00%
Llama 3 70B Instruct
meta-llama
Fireworks
#54
$0.230.44s141.5T/s100.00%
Llama 3 8B Instruct
meta-llama
Fireworks
#91
$0.030.30s216.9T/s100.00%
Gemini Flash 2.0 Experimental (free)
google
Google AI Studio
#4
$0.002.06s138.6T/s60.43%
Grok 2 1212
x-ai
xAI
-
$2.080.52s68.0T/s99.93%
Qwen2-VL 72B Instruct
qwen
Hyperbolic
-
$0.404.66s7.8T/s95.89%
Llama 3.2 1B Instruct (free)
meta-llama
SambaNova
#147
$0.000.41s2137.9T/s98.58%
Llama 3.2 3B Instruct
meta-llama
Fireworks
#119
$0.020.33s217.9T/s100.00%
Rocinante 12B
thedrummer
Infermatic
-
$0.250.76s22.1T/s100.00%
Magnum v4 72B
anthracite-org
Infermatic
-
$1.891.24s31.6T/s99.99%
Claude 3.5 Haiku
anthropic
Google Vertex
#24
$0.832.29s62.1T/s99.99%
Claude 3.5 Haiku (self-moderated)
anthropic
Anthropic
-
$0.832.58s53.2T/s95.94%
Unslopnemo 12b
thedrummer
Infermatic
-
$0.500.63s68.4T/s100.00%
Mistral Large 2411
mistralai
Mistral
#38
$2.050.59s41.6T/s99.99%
GPT-4o (2024-11-20)
openai
OpenAI
#1
$2.580.55s87.5T/s99.97%
Llama 3.1 405B Instruct
meta-llama
Fireworks
-
$0.810.77s58.9T/s99.99%
Hermes 3 70B Instruct
nousresearch
Hyperbolic
#43
$0.120.73s31.7T/s99.97%
Llama 3.1 Euryale 70B v2.2
sao10k
DeepInfra
-
$0.350.63s35.1T/s100.00%
GPT-4
openai
Azure
-
$30.481.23s37.6T/s96.00%
GPT-3.5 Turbo 16k
openai
OpenAI
#98
$0.510.28s113.1T/s95.99%
GPT-3.5 Turbo
openai
OpenAI
#96
$0.510.42s115.0T/s99.95%
MythoMax 13B (nitro)
gryphe
Fireworks
-
$0.200.23s110.6T/s100.00%
Hermes 13B
nousresearch
NovitaAI
-
$0.170.59s81.9T/s100.00%
OpenHermes 2.5 Mistral 7B
teknium
NovitaAI
#120
$0.170.81s65.6T/s100.00%
Mixtral 8x7B Instruct
mistralai
Fireworks
#111
$0.240.39s139.0T/s100.00%
Mistral Small
mistralai
Mistral
-
$0.200.65s60.2T/s100.00%
Claude 3 Haiku
anthropic
Google Vertex
-
$0.261.11s134.0T/s100.00%
Claude 3 Haiku (self-moderated)
anthropic
Anthropic
-
$0.261.09s129.7T/s99.83%
Gemini Pro 1.5
google
Google AI Studio
#9
$1.291.88s49.2T/s99.56%
WizardLM-2 7B
microsoft
Lepton
-
$0.060.29s76.4T/s100.00%
Mixtral 8x22B Instruct
mistralai
Mistral
#86
$0.910.33s79.9T/s100.00%
GPT-4o (extended)
openai
OpenAI
-
$6.140.82s93.6T/s76.00%
GPT-4o
openai
OpenAI
-
$2.581.02s140.9T/s100.00%
Hermes 2 Pro - Llama-3 8B
nousresearch
Lambda
-
$0.030.52s101.3T/s99.88%
Mistral 7B Instruct (nitro)
mistralai
Lepton
-
$0.070.42s87.9T/s52.00%
Mistral 7B Instruct
mistralai
Together
#136
$0.030.30s139.9T/s100.00%
Mistral 7B Instruct (free)
mistralai
Lepton
-
$0.000.42s87.9T/s95.99%
Dolphin 2.9.2 Mixtral 8x22B 🐬
cognitivecomputations
NovitaAI
-
$0.911.82s24.0T/s99.11%
Claude 3.5 Sonnet (2024-06-20)
anthropic
Anthropic
-
$3.121.42s60.1T/s99.99%
Claude 3.5 Sonnet (2024-06-20) (self-moderated)
anthropic
Anthropic
-
$3.121.42s60.1T/s99.98%
Gemma 2 9B
google
DeepInfra
#65
$0.030.47s49.0T/s100.00%
Nova Lite 1.0
amazon
Amazon BedRock
#63
$0.060.53s73.6T/s96.00%
Llama 3.2 11B Vision Instruct
meta-llama
Together
-
$0.060.82s122.1T/s100.00%
Llama 3.1 Nemotron 70B Instruct
nvidia
Infermatic
#38
$0.121.43s30.1T/s95.95%
Qwen2.5 7B Instruct
qwen
Together
-
$0.270.75s130.9T/s96.00%
Nova Pro 1.0
amazon
Amazon BedRock
-
$0.830.56s75.9T/s95.99%
Command R+ (08-2024)
cohere
Cohere
#48
$2.450.62s57.5T/s99.96%
Codestral Mamba
mistralai
Mistral
-
$0.250.43s91.3T/s96.00%
Llama 3.1 70B Instruct (free)
meta-llama
SambaNova
-
$0.001.68s391.9T/s88.29%
Llama 3.1 Sonar 70B Online
perplexity
Perplexity
#130
$1.011.43s55.7T/s99.99%
GPT-4o (2024-08-06)
openai
Azure
#15
$2.5810.44s97.4T/s99.98%
ChatGPT-4o
openai
OpenAI
-
$5.120.69s94.3T/s99.88%
Llama 3.1 Sonar 405B Online
perplexity
Perplexity
-
$5.042.25s19.8T/s97.33%
GPT-4 Turbo (older v1106)
openai
OpenAI
#24
$10.241.18s31.9T/s95.99%
OpenChat 3.5 7B (free)
openchat
Lepton
-
$0.000.51s83.8T/s96.00%
Mistral Large
mistralai
Mistral
#26
$2.050.72s31.2T/s99.95%
Claude 3 Sonnet
anthropic
Anthropic
-
$3.121.49s64.1T/s99.95%
Claude 3 Sonnet (self-moderated)
anthropic
Anthropic
-
$3.121.49s64.1T/s100.00%
Claude 3 Opus
anthropic
Anthropic
-
$15.602.92s26.6T/s99.98%
Llama 3 70B Instruct (nitro)
meta-llama
Fireworks
-
$0.800.44s141.5T/s100.00%
Llama 3 8B Instruct (free)
meta-llama
Together (lite)
-
$0.000.58s181.0T/s100.00%
LlamaGuard 2 8B
meta-llama
Together
-
$0.180.41s45.1T/s100.00%
Llama 3 Euryale 70B v2.1
sao10k
DeepInfra
-
$0.350.17s40.1T/s100.00%
Gemma 2 9B (free)
google
DeepInfra
-
$0.000.53s48.8T/s100.00%
Created by JCdata from OpenRouter and Chatbot Arena