Model | Provider | |||||||
---|---|---|---|---|---|---|---|---|
Llama 3.3 70B Instruct meta-llama | Groq | #41 | $0.12 | 0.44s | 304.6T/s | 99.88% | ||
Gemini Flash 1.5 8B google | Google AI Studio | #78 | $0.04 | 0.43s | 336.5T/s | 99.87% | ||
Gemini Flash 2.0 google | Google Vertex | #13 | $0.10 | 0.46s | 176.7T/s | 99.97% | ||
GPT-4o-mini openai | Azure | #5 | $0.15 | 1.06s | 125.5T/s | 99.86% | ||
Gemini 2.0 Flash Lite google | Google AI Studio | #15 | $0.08 | 0.46s | 148.2T/s | 99.61% | ||
Gemini Flash 1.5 google | Google AI Studio | #55 | $0.08 | 0.83s | 175.1T/s | 99.98% | ||
DeepSeek V3 0324 (free) deepseek | Targon | - | $0.00 | 7.08s | 220.6T/s | 94.68% | ||
Claude 3.7 Sonnet (thinking) anthropic | Amazon Bedrock | #12 | $3.12 | 0.82s | 87.6T/s | 97.45% | ||
R1 (free) deepseek | Chutes | #7 | $0.00 | 9.29s | 68.4T/s | 99.14% | ||
Claude 3.7 Sonnet anthropic | Amazon Bedrock | #11 | $3.12 | 0.82s | 87.6T/s | 96.75% | ||
R1 Distill Llama 70B deepseek | SambaNova | - | $0.24 | 2.25s | 605.8T/s | 98.90% | ||
Claude 3.5 Sonnet anthropic | Google Vertex | - | $3.12 | 1.72s | 60.2T/s | 99.85% | ||
Claude 3.5 Haiku anthropic | Google Vertex | - | $0.83 | 0.75s | 109.1T/s | 99.24% | ||
MythoMax 13B gryphe | Together | - | $0.07 | 0.45s | 120.1T/s | 99.99% | ||
DeepSeek V3 0324 deepseek | SambaNova | - | $0.28 | 5.94s | 222.2T/s | 97.73% | ||
Mistral Small 3 mistralai | Mistral | #74 | $0.07 | 0.30s | 119.2T/s | 100.00% | ||
Claude 3.5 Haiku (self-moderated) anthropic | Anthropic | #44 | $0.83 | 1.30s | 108.8T/s | 98.92% | ||
Mistral 7B Instruct mistralai | Together | #142 | $0.03 | 0.38s | 163.1T/s | 100.00% | ||
GPT-4o-mini (2024-07-18) openai | OpenAI | #42 | $0.15 | 0.33s | 76.9T/s | 99.98% | ||
Command R (08-2024) cohere | Cohere | #85 | $0.15 | 0.37s | 54.7T/s | 99.67% | ||
R1 deepseek | SambaNova | #7 | $0.72 | 8.28s | 337.6T/s | 93.03% | ||
Gemma 3 27B google | Parasail | #15 | $0.10 | 0.74s | 46.5T/s | 99.50% | ||
Qwen2.5 72B Instruct qwen | SambaNova | #56 | $0.13 | 0.39s | 360.2T/s | 99.68% | ||
Llama 3.1 8B Instruct meta-llama | SambaNova | #108 | $0.02 | 0.30s | 816.6T/s | 99.99% | ||
DeepSeek V3 deepseek | Fireworks | #17 | $0.41 | 1.40s | 52.9T/s | 99.96% | ||
DeepSeek V3 (free) deepseek | Chutes | #16 | $0.00 | 1.55s | 38.7T/s | 99.51% | ||
Qwen2.5 Coder 32B Instruct qwen | SambaNova | #68 | $0.07 | 0.52s | 526.7T/s | 99.90% | ||
LFM 3B liquid | Liquid | - | $0.02 | 0.40s | 36.1T/s | 100.00% | ||
Claude 3.5 Sonnet (self-moderated) anthropic | Anthropic | - | $3.12 | 1.51s | 54.0T/s | 93.79% | ||
WizardLM-2 8x22B microsoft | Together | - | $0.50 | 0.52s | 76.0T/s | 99.99% | ||
GPT-4o openai | OpenAI | - | $2.58 | 0.46s | 73.1T/s | 98.68% | ||
Hermes 2 Pro - Llama-3 8B nousresearch | Lambda | - | $0.03 | 0.50s | 144.5T/s | 100.00% | ||
Mistral Nemo mistralai | Parasail | - | $0.04 | 0.74s | 135.8T/s | 99.99% | ||
Llama 3.1 70B Instruct meta-llama | SambaNova | - | $0.12 | 0.57s | 281.3T/s | 99.98% | ||
Llama 3.1 405B Instruct meta-llama | SambaNova | - | $0.81 | 2.55s | 90.2T/s | 99.86% | ||
Hermes 3 405B Instruct nousresearch | Lambda | - | $0.81 | 1.16s | 27.9T/s | 99.90% | ||
Gemini Pro 2.5 Experimental (free) google | Google Vertex | - | $0.00 | 12.82s | 120.2T/s | 48.84% | ||
Llama 3 8B Instruct meta-llama | Groq | #100 | $0.03 | 0.28s | 941.0T/s | 99.95% | ||
Claude 3 Haiku anthropic | Anthropic | #84 | $0.26 | 0.66s | 155.4T/s | 99.77% | ||
Phi 4 microsoft | Nebius AI Studio | #85 | $0.07 | 0.14s | 115.3T/s | 99.81% | ||
Gemma 2 9B (free) google | Chutes | #79 | $0.00 | 1.52s | 114.3T/s | 99.93% | ||
Gemma 2 9B google | Groq | #79 | $0.03 | 0.29s | 550.3T/s | 100.00% | ||
QwQ 32B qwen | SambaNova | #126 | $0.12 | 4.62s | 428.5T/s | 99.50% | ||
GPT-4.5 (Preview) openai | OpenAI | #2 | $76.20 | 1.22s | 9.4T/s | 99.55% | ||
o3 Mini openai | OpenAI | #18 | $1.14 | 6.51s | 117.5T/s | 99.93% | ||
Mixtral 8x7B Instruct mistralai | Groq | #115 | $0.24 | 0.36s | 648.5T/s | 100.00% | ||
Claude 3.7 Sonnet (self-moderated) anthropic | Anthropic | #9 | $3.12 | 1.34s | 56.9T/s | 92.06% | ||
Claude 3 Haiku (self-moderated) anthropic | Anthropic | #83 | $0.26 | 0.66s | 155.4T/s | 98.39% | ||
Llama 3.2 3B Instruct meta-llama | SambaNova | #121 | $0.02 | 0.33s | 1401.8T/s | 100.00% | ||
Claude 3.5 Haiku (2024-10-22) anthropic | Google Vertex | #40 | $0.83 | 1.35s | 56.5T/s | 85.69% | ||
Claude 3.5 Haiku (2024-10-22) (self-moderated) anthropic | Anthropic | #40 | $0.83 | 1.40s | 52.6T/s | 99.01% | ||
Gemini Pro 2.0 Experimental (free) google | Google Vertex | #4 | $0.00 | 1.59s | 47.3T/s | 92.93% | ||
Gemini Flash 2.0 Experimental (free) google | Google Vertex | #8 | $0.00 | 0.79s | 206.0T/s | 93.84% | ||
Grok 2 1212 x-ai | xAI | - | $2.08 | 0.16s | 50.9T/s | 99.92% | ||
Grok 2 Vision 1212 x-ai | xAI | - | $2.08 | 0.37s | 26.1T/s | 99.38% | ||
Llama 3.3 Euryale 70B sao10k | DeepInfra | - | $0.71 | 0.62s | 36.4T/s | 99.91% | ||
Codestral 2501 mistralai | Mistral | - | $0.31 | 0.25s | 179.1T/s | 99.97% | ||
MiniMax-01 minimax | Minimax | - | $0.21 | 1.23s | 25.4T/s | 99.95% | ||
LFM 7B liquid | Liquid | - | $0.01 | 0.69s | 58.9T/s | 99.99% | ||
R1 Distill Qwen 32B deepseek | Groq | - | $0.12 | 0.67s | 137.3T/s | 99.81% | ||
Qwen-Max qwen | Alibaba | - | $1.65 | 0.90s | 36.6T/s | 99.66% | ||
Qwen-Turbo qwen | Alibaba | - | $0.05 | 0.87s | 104.1T/s | 99.63% | ||
Gemini Flash Lite 2.0 Preview (free) google | Google Vertex | #13 | $0.00 | 0.76s | 144.3T/s | 98.55% | ||
Mistral Small 3.1 24B mistralai | Mistral | - | $0.10 | 0.47s | 109.2T/s | 76.80% | ||
GPT-4o (2024-11-20) openai | OpenAI | - | $2.58 | 0.71s | 37.5T/s | 99.81% | ||
Rocinante 12B thedrummer | Infermatic | - | $0.25 | 0.65s | 24.8T/s | 99.92% | ||
Qwen2.5 7B Instruct qwen | Together | - | $0.03 | 0.36s | 107.1T/s | 100.00% | ||
Unslopnemo 12B thedrummer | Infermatic | - | $0.50 | 0.55s | 68.9T/s | 99.68% | ||
Mistral Tiny mistralai | Mistral | - | $0.25 | 0.31s | 127.3T/s | 100.00% | ||
Gemini Pro 1.5 google | Google Vertex | - | $1.29 | 1.13s | 76.4T/s | 99.96% | ||
Llama 3 Lumimaid 8B neversleep | Mancer (private) | - | $0.20 | 0.57s | 62.1T/s | 99.98% | ||
Llama 3 Lumimaid 8B (extended) neversleep | Mancer (private) | - | $0.20 | 0.57s | 62.1T/s | 100.00% | ||
GPT-4o (2024-08-06) openai | OpenAI | #19 | $2.58 | 0.45s | 85.9T/s | 99.93% | ||
Hermes 3 70B Instruct nousresearch | Hyperbolic | - | $0.12 | 1.04s | 31.8T/s | 99.91% | ||
Llama 3.1 Euryale 70B v2.2 sao10k | DeepInfra | - | $0.71 | 0.45s | 35.6T/s | 99.95% | ||
Llama 3 70B Instruct meta-llama | Groq | #63 | $0.23 | 0.18s | 316.7T/s | 100.00% | ||
GPT-3.5 Turbo openai | OpenAI | #105 | $0.51 | 0.33s | 117.8T/s | 99.52% | ||
QwQ 32B (free) qwen | Nineteen | #126 | $0.00 | 3.23s | 128.5T/s | 98.03% | ||
o1-mini openai | OpenAI | #25 | $1.14 | 0.65s | 162.2T/s | 99.19% | ||
Llama 3.2 1B Instruct meta-llama | SambaNova | #150 | $0.01 | 0.49s | 1931.0T/s | 100.00% | ||
Mistral Small 3.1 24B (free) mistralai | Chutes | - | $0.00 | 1.33s | 62.0T/s | 98.32% | ||
Command R+ (08-2024) cohere | Cohere | #65 | $2.45 | 0.51s | 41.7T/s | 99.59% | ||
o3 Mini High openai | OpenAI | #13 | $1.14 | 9.32s | 119.2T/s | 98.71% | ||
Llama 3.3 70B Instruct (free) meta-llama | Together | #39 | $0.00 | 1.18s | 89.5T/s | 99.60% | ||
Gemma 3 27B (free) google | Chutes | #13 | $0.00 | 1.18s | 41.7T/s | 96.76% | ||
Nova Lite 1.0 amazon | Amazon Bedrock | #77 | $0.06 | 0.37s | 128.7T/s | 98.94% | ||
R1 Distill Llama 70B (free) deepseek | Together | - | $0.00 | 1.10s | 49.4T/s | 98.41% | ||
Sonar perplexity | Perplexity | - | $1.01 | 1.91s | 37.6T/s | 99.56% | ||
Qwen2.5 VL 72B Instruct qwen | Parasail | - | $0.71 | 0.71s | 33.5T/s | 99.60% | ||
R1 Distill Llama 8B deepseek | NovitaAI | - | $0.04 | 0.84s | 55.1T/s | 97.94% | ||
Ministral 3B mistralai | Mistral | - | $0.04 | 0.19s | 197.8T/s | 99.95% | ||
Magnum v4 72B anthracite-org | Infermatic | - | $1.89 | 0.53s | 29.9T/s | 99.98% | ||
Llama 3.2 11B Vision Instruct meta-llama | SambaNova | - | $0.06 | 0.70s | 776.1T/s | 99.18% | ||
ReMM SLERP 13B undi95 | Mancer | - | $0.81 | 0.77s | 45.7T/s | 100.00% | ||
OpenChat 3.5 7B openchat | Lepton | - | $0.06 | 0.39s | 108.3T/s | 99.94% | ||
Dolphin 2.9.2 Mixtral 8x22B 🐬 cognitivecomputations | NovitaAI | - | $0.91 | 4.06s | 8.9T/s | 98.49% | ||
ChatGPT-4o openai | OpenAI | #2 | $5.12 | 0.62s | 95.4T/s | 99.75% |
Created by JC, data from OpenRouter and Chatbot Arena