AI Stats
255 LLM models tracked with pricing & performance benchmarks.
Data sourced from ArtificialAnalysis.com & Epoch.ai
Leaderboard
Switch metrics to see different top-5 rankings.
- 1Gemini 3 Pro Preview (high)Google89.8%
- 2Claude Opus 4.5 (Reasoning)Anthropic89.5%
- 3Gemini 3 Pro Preview (low)Google89.5%
- 4Claude Opus 4.5 (Non-reasoning)Anthropic88.9%
- 5Claude 4.1 Opus (Reasoning)Anthropic88.0%
- 1Gemini 3 Pro Preview (high)Google90.8%
- 2Gemini 3 Pro Preview (low)Google88.7%
- 3Grok 4xAI87.7%
- 4GPT-5.1 (high)OpenAI87.3%
- 5Claude Opus 4.5 (Reasoning)Anthropic86.6%
- 1Gemini 3 Pro Preview (high)Google37.2%
- 2Claude Opus 4.5 (Reasoning)Anthropic28.4%
- 3Gemini 3 Pro Preview (low)Google27.6%
- 4GPT-5.1 (high)OpenAI26.5%
- 5GPT-5 (high)OpenAI26.5%
- 1GPT-5 (high)OpenAI95.7%
- 2Grok 4xAI94.3%
- 3o4-mini (high)OpenAI94.0%
- 4Qwen3 235B A22B 2507 (Reasoning)Alibaba94.0%
- 5Grok 3 mini Reasoning (high)xAI93.3%
- 1Gemini 3 Pro Preview (high)Google56.1%
- 2Gemini 3 Pro Preview (low)Google49.9%
- 3Claude Opus 4.5 (Reasoning)Anthropic49.5%
- 4Claude Opus 4.5 (Non-reasoning)Anthropic47.0%
- 5o4-mini (high)OpenAI46.5%
- 1GPT-5 (high)OpenAI99.4%
- 2o3OpenAI99.2%
- 3Grok 3 mini Reasoning (high)xAI99.2%
- 4GPT-5 (medium)OpenAI99.1%
- 5Claude 4 Sonnet (Reasoning)Anthropic99.1%
- 1Gemini 3 Pro Preview (high)Google91.7%
- 2gpt-oss-120B (high)OpenAI87.8%
- 3Claude Opus 4.5 (Reasoning)Anthropic87.1%
- 4GPT-5.1 (high)OpenAI86.8%
- 5o4-mini (high)OpenAI85.9%
- 1GPT-5 Codex (high)OpenAI98.7%
- 2Gemini 3 Pro Preview (high)Google95.7%
- 3Kimi K2 ThinkingMoonshot AI94.7%
- 4GPT-5 (high)OpenAI94.3%
- 5GPT-5.1 (high)OpenAI94.0%
- 1Gemini 3 Pro Preview (high)Google62.3%
- 2Claude Opus 4.5 (Reasoning)Anthropic60.2%
- 3GPT-5.1 (high)OpenAI57.5%
- 4Gemini 3 Pro Preview (low)Google55.8%
- 5Grok 4xAI55.1%
- 1Gemini 3 Pro Preview (high)Google72.8%
- 2Claude Opus 4.5 (Reasoning)Anthropic69.8%
- 3GPT-5.1 (high)OpenAI69.7%
- 4GPT-5 Codex (high)OpenAI68.5%
- 5GPT-5 (high)OpenAI68.5%
- 1gemini 3 pro previewGoogle DeepMind92.6%
- 2gpt 5.1 highOpenAI87.6%
- 3gpt 5 2025 08 07 highOpenAI86.2%
- 4claude opus 4 5 20251101 32KAnthropic86.1%
- 5claude opus 4 5 20251101 16KAnthropic85.5%
- 1o3 2025 04 16 highOpenAI31.8%
- 2gpt 5.1 highOpenAI31.2%
- 3gemini 2.5 proGoogle DeepMind25.6%
- 4o4 mini 2025 04 16 highOpenAI21.4%
- 5claude sonnet 4 5 20250929 32KAnthropic18.5%
- 1claude sonnet 4 20250514Anthropic72.8%
- 2claude 3 7 sonnet 20250219Anthropic62.3%
- 3gpt 5 mini 2025 08 07 highOpenAI61.6%
- 4claude haiku 4 5 20251001Anthropic60.6%
- 5DeepSeek V3 0324DeepSeek54.8%
- 1o3 2025 04 16 highOpenAI96.8%
- 2gpt 5.1 highOpenAI95.8%
- 3DeepSeek R1 0528DeepSeek94.5%
- 4gemini 2.5 proGoogle DeepMind91.2%
- 5o3 mini 2025 01 31 highOpenAI91.2%
- 1o3 2025 04 16 highOpenAI87.6%
- 2gpt 5.1 highOpenAI84.2%
- 3o3 2025 04 16 mediumOpenAI75.2%
- 4gemini 2.5 proGoogle DeepMind32.5%
- 5claude sonnet 4 5 20250929 32KAnthropic21.8%
- 1gpt 5.1 highOpenAI93.2%
- 2gemini 3 pro previewGoogle DeepMind92.8%
- 3gpt 5 2025 08 07 highOpenAI92.1%
- 4gemini 2.5 proGoogle DeepMind91.5%
- 5o3 2025 04 16 highOpenAI90.8%
- 1gpt 5.1 highOpenAI97.8%
- 2o3 2025 04 16 highOpenAI97.5%
- 3DeepSeek R1 0528DeepSeek97.2%
- 4gemini 2.5 proGoogle DeepMind96.8%
- 5claude sonnet 4 5 20250929 32KAnthropic96.5%
- 1claude sonnet 4 5 20250929Anthropic85.6%
- 2gpt 5.1 highOpenAI84.2%
- 3gemini 2.5 proGoogle DeepMind82.1%
- 4DeepSeek V3 0324DeepSeek78.8%
- 5gpt 4.1 2025 04 14OpenAI72.4%
All models
Search and scan every tracked model.
Showing 20 of 255 models
No models found
Try adjusting your search or filters.
Gemini 3 Pro Preview (high)
Claude Opus 4.5 (Reasoning)
Anthropic
GPT-5.1 (high)
OpenAI
GPT-5 Codex (high)
OpenAI
GPT-5 (high)
OpenAI
Kimi K2 Thinking
Moonshot AI
GPT-5 (medium)
OpenAI
o3
OpenAI
Grok 4
xAI
o3-pro
OpenAI
Gemini 3 Pro Preview (low)
GPT-5 mini (high)
OpenAI
Grok 4.1 Fast (Reasoning)
xAI
Claude 4.5 Sonnet (Reasoning)
Anthropic
GPT-5 (low)
OpenAI
MiniMax-M2
MiniMax
GPT-5 mini (medium)
OpenAI
gpt-oss-120B (high)
OpenAI
Grok 4 Fast (Reasoning)
xAI
Claude Opus 4.5 (Non-reasoning)
Anthropic
Gemini 2.5 Pro
o4-mini (high)
OpenAI
Claude 4.1 Opus (Reasoning)
Anthropic
DeepSeek V3.1 Terminus (Reasoning)
DeepSeek
Qwen3 235B A22B 2507 (Reasoning)
Alibaba
Grok 3 mini Reasoning (high)
xAI
Doubao Seed Code
ByteDance Seed
DeepSeek V3.2 Exp (Reasoning)
DeepSeek
Claude 4 Sonnet (Reasoning)
Anthropic
GLM-4.6 (Reasoning)
Z AI
Qwen3 Max Thinking
Alibaba
Qwen3 Max
Alibaba
Claude 4.5 Haiku (Reasoning)
Anthropic
Gemini 2.5 Flash Preview (Sep '25) (Reasoning)
Qwen3 VL 235B A22B (Reasoning)
Alibaba
Qwen3 Next 80B A3B (Reasoning)
Alibaba
Claude 4 Opus (Reasoning)
Anthropic
Gemini 2.5 Pro Preview (Mar' 25)
DeepSeek V3.1 (Reasoning)
DeepSeek
Gemini 2.5 Pro Preview (May' 25)
gpt-oss-20B (high)
OpenAI
Magistral Medium 1.2
Mistral
DeepSeek R1 0528 (May '25)
DeepSeek
Qwen3 VL 32B (Reasoning)
Alibaba
Seed-OSS-36B-Instruct
ByteDance Seed
GLM-4.5 (Reasoning)
Z AI
Gemini 2.5 Flash (Reasoning)
GPT-5 nano (high)
OpenAI
o3-mini (high)
OpenAI
Kimi K2 0905
Moonshot AI
Claude 3.7 Sonnet (Reasoning)
Anthropic
Claude 4.5 Sonnet (Non-reasoning)
Anthropic
GPT-5 nano (medium)
OpenAI
GLM-4.5-Air
Z AI
Grok Code Fast 1
xAI
Qwen3 Max (Preview)
Alibaba
Kimi K2
Moonshot AI
o3-mini
OpenAI
o1-pro
OpenAI
Gemini 2.5 Flash-Lite Preview (Sep '25) (Reasoning)
gpt-oss-120B (low)
OpenAI
o1
OpenAI
Gemini 2.5 Flash Preview (Sep '25) (Non-reasoning)
Qwen3 30B A3B 2507 (Reasoning)
Alibaba
DeepSeek V3.2 Exp (Non-reasoning)
DeepSeek
MiniMax M1 80k
MiniMax
DeepSeek V3.1 Terminus (Non-reasoning)
DeepSeek
Qwen3 235B A22B 2507 Instruct
Alibaba
Grok 3
xAI
Qwen3 VL 30B A3B (Reasoning)
Alibaba
Llama Nemotron Super 49B v1.5 (Reasoning)
NVIDIA
o1-preview
OpenAI
Qwen3 Next 80B A3B Instruct
Alibaba
DeepSeek V3.1 (Non-reasoning)
DeepSeek
Ling-1T
InclusionAI
GLM-4.6 (Non-reasoning)
Z AI
Claude 4.1 Opus (Non-reasoning)
Anthropic
Claude 4 Sonnet (Non-reasoning)
Anthropic
gpt-oss-20B (low)
OpenAI
Qwen3 VL 235B A22B Instruct
Alibaba
DeepSeek R1 (Jan '25)
DeepSeek
GPT-5 (minimal)
OpenAI
GPT-4.1
OpenAI
Magistral Small 1.2
Mistral
GPT-5.1 (Non-reasoning)
OpenAI
EXAONE 4.0 32B (Reasoning)
LG AI Research
GPT-4.1 mini
OpenAI
Claude 4 Opus (Non-reasoning)
Anthropic
Qwen3 Coder 480B A35B Instruct
Alibaba
GPT-5 (ChatGPT)
OpenAI
Ring-1T
InclusionAI
Qwen3 235B A22B (Reasoning)
Alibaba
Claude 4.5 Haiku (Non-reasoning)
Anthropic
GPT-5 mini (minimal)
OpenAI
Gemini 2.5 Flash-Lite Preview (Sep '25) (Non-reasoning)
Hermes 4 - Llama-3.1 405B (Reasoning)
Nous Research
DeepSeek V3 0324
DeepSeek
Claude 3.7 Sonnet (Non-reasoning)
Anthropic
Qwen3 VL 32B Instruct
Alibaba
Gemini 2.5 Flash (Non-reasoning)
Gemini 2.5 Flash-Lite (Reasoning)
MiniMax M1 40k
MiniMax
Qwen3 Omni 30B A3B (Reasoning)
Alibaba
Ring-flash-2.0
InclusionAI
Hermes 4 - Llama-3.1 70B (Reasoning)
Nous Research
Qwen3 32B (Reasoning)
Alibaba
Grok 4 Fast (Non-reasoning)
xAI
Qwen3 VL 30B A3B Instruct
Alibaba
Llama 3.1 Nemotron Ultra 253B v1 (Reasoning)
NVIDIA
Grok 4.1 Fast (Non-reasoning)
xAI
Ling-flash-2.0
InclusionAI
QwQ 32B
Alibaba
Solar Pro 2 (Reasoning)
Upstage
NVIDIA Nemotron Nano 9B V2 (Reasoning)
NVIDIA
GLM-4.5V (Reasoning)
Z AI
Qwen3 30B A3B 2507 Instruct
Alibaba
Qwen3 30B A3B (Reasoning)
Alibaba
NVIDIA Nemotron Nano 9B V2 (Non-reasoning)
NVIDIA
Qwen3 14B (Reasoning)
Alibaba
Llama 4 Maverick
Meta
GPT-4o (March 2025, chatgpt-4o-latest)
OpenAI
Mistral Medium 3.1
Mistral
Sonar Reasoning
Perplexity
Gemini 2.0 Flash (Feb '25)
Mistral Medium 3
Mistral
Qwen3 Coder 30B A3B Instruct
Alibaba
Magistral Medium 1
Mistral
ERNIE 4.5 300B A47B
Baidu
DeepSeek R1 Distill Qwen 32B
DeepSeek
Hermes 4 - Llama-3.1 405B (Non-reasoning)
Nous Research
DeepSeek V3 (Dec '24)
DeepSeek
Nova Premier
Amazon
Qwen3 VL 8B (Reasoning)
Alibaba
Magistral Small 1
Mistral
DeepSeek R1 0528 Qwen3 8B
DeepSeek
Qwen2.5 Max
Alibaba
EXAONE 4.0 32B (Non-reasoning)
LG AI Research
Solar Pro 2 (Non-reasoning)
Upstage
Qwen3 Omni 30B A3B Instruct
Alibaba
Gemini 2.5 Flash-Lite (Non-reasoning)
Qwen3 235B A22B (Non-reasoning)
Alibaba
DeepSeek R1 Distill Llama 70B
DeepSeek
Claude 3.5 Sonnet (Oct '24)
Anthropic
DeepSeek R1 Distill Qwen 14B
DeepSeek
Qwen3 14B (Non-reasoning)
Alibaba
Mistral Small 3.2
Mistral
GPT-5 nano (minimal)
OpenAI
GPT-4o (Aug '24)
OpenAI
Sonar
Perplexity
Qwen3 8B (Reasoning)
Alibaba
MiniMax-Text-01
MiniMax
Sonar Pro
Perplexity
Llama 3.1 Instruct 405B
Meta
Llama 4 Scout
Meta
QwQ 32B-Preview
Alibaba
Llama 3.3 Instruct 70B
Meta
Devstral Medium
Mistral
Ling-mini-2.0
InclusionAI
GPT-4.1 nano
OpenAI
Devstral Small (Jul '25)
Mistral
Qwen3 VL 8B Instruct
Alibaba
GPT-4o (Nov '24)
OpenAI
Command A
Cohere
Mistral Large 2 (Nov '24)
Mistral
Gemini 2.0 Flash-Lite (Feb '25)
Llama Nemotron Super 49B v1.5 (Non-reasoning)
NVIDIA
Qwen3 30B A3B (Non-reasoning)
Alibaba
Qwen3 32B (Non-reasoning)
Alibaba
GPT-4o (May '24)
OpenAI
Gemini 2.0 Flash-Lite (Preview)
GLM-4.5V (Non-reasoning)
Z AI
Reka Flash 3
Reka AI
Qwen3 4B (Reasoning)
Alibaba
Claude 3.5 Sonnet (June '24)
Anthropic
GPT-4o (ChatGPT)
OpenAI
Pixtral Large
Mistral
Nova Pro
Amazon
Mistral Small 3.1
Mistral
Grok 2 (Dec '24)
xAI
GPT-4 Turbo
OpenAI
Hermes 4 - Llama-3.1 70B (Non-reasoning)
Nous Research
Llama 3.1 Nemotron Instruct 70B
NVIDIA
Qwen3 8B (Non-reasoning)
Alibaba
Granite 4.0 H Small
IBM
Phi-4
Microsoft Azure
Llama 3.1 Instruct 70B
Meta
Qwen3 1.7B (Reasoning)
Alibaba
Mistral Large 2 (Jul '24)
Mistral
CompactifAI Llama 4 Scout Slim
Multiverse Computing
Qwen2.5 Coder Instruct 32B
Alibaba
Nova Lite
Amazon
GPT-4
OpenAI
Mistral Small 3
Mistral
GPT-4o mini
OpenAI
Jamba 1.7 Large
AI21 Labs
Qwen3 4B (Non-reasoning)
Alibaba
Claude 3 Opus
Anthropic
Claude 3.5 Haiku
Anthropic
Codestral (Jan '25)
Mistral
Devstral Small (May '25)
Mistral
Reka Core
Reka AI
Qwen2.5 Turbo
Alibaba
Reka Flash (Sep '24)
Reka AI
Solar Mini
Upstage
Llama 3.2 Instruct 90B (Vision)
Meta
Reka Flash (Feb '24)
Reka AI
Reka Edge
Reka AI
Nova Micro
Amazon
Llama 3.1 Instruct 8B
Meta
CompactifAI Mistral Small 3.1 Slim
Multiverse Computing
CompactifAI Llama 3.3 70B Slim
Multiverse Computing
Llama 3.2 Instruct 11B (Vision)
Meta
Gemma 3n E4B Instruct
Granite 3.3 8B (Non-reasoning)
IBM
Jamba 1.7 Mini
AI21 Labs
Jamba 1.5 Large
AI21 Labs
Hermes 3 - Llama-3.1 70B
Nous Research
OLMo 2 32B
Allen Institute for AI
Phi-3 Medium Instruct 14B
Microsoft Azure
Qwen3 1.7B (Non-reasoning)
Alibaba
Jamba 1.6 Large
AI21 Labs
Qwen3 0.6B (Reasoning)
Alibaba
Aya Expanse 32B
Cohere
Claude 3 Sonnet
Anthropic
Llama 3 Instruct 70B
Meta
Mistral Small (Sep '24)
Mistral
Phi-3 Mini Instruct 3.8B
Microsoft Azure
Ministral 8B
Mistral
Mistral Large (Feb '24)
Mistral
Llama 2 Chat 7B
Meta
CompactifAI Llama 3.1 8B Slim
Multiverse Computing
Llama 3.2 Instruct 3B
Meta
Qwen3 0.6B (Non-reasoning)
Alibaba
Ministral 3B
Mistral
Aya Expanse 8B
Cohere
Claude 3 Haiku
Anthropic
Llama 3.2 Instruct 1B
Meta
Pixtral 12B (2409)
Mistral
Mistral Small (Feb '24)
Mistral
Mistral Medium
Mistral
GPT-3.5 Turbo
OpenAI
Gemma 2 9B
Command-R+ (Aug '24)
Cohere
Llama 3 Instruct 8B
Meta
Command-R+ (Apr '24)
Cohere
Mistral NeMo
Mistral
Jamba 1.5 Mini
AI21 Labs
Jamba 1.6 Mini
AI21 Labs
Mixtral 8x7B Instruct
Mistral
Command-R (Mar '24)
Cohere
Command-R (Aug '24)
Cohere
Mistral 7B Instruct
Mistral
Cogito v2.1 (Reasoning)
Deep Cogito
DeepSeek-OCR
DeepSeek
Grok 3 mini Reasoning (Low)
xAI
| Model | Provider | Input | Output | MMLU | GPQA | TPS |
|---|---|---|---|---|---|---|
| Gemini 3 Pro Preview (high) | $2.000 | $12.000 | 89.8% | 90.8% | - | |
| Claude Opus 4.5 (Reasoning) | Anthropic | $5.000 | $25.000 | 89.5% | 86.6% | 44.7 |
| GPT-5.1 (high) | OpenAI | $1.250 | $10.000 | 87.0% | 87.3% | 147.0 |
| GPT-5 Codex (high) | OpenAI | $1.250 | $10.000 | 86.5% | 83.7% | 232.9 |
| GPT-5 (high) | OpenAI | $1.250 | $10.000 | 87.1% | 85.4% | - |
| Kimi K2 Thinking | Moonshot AI | $0.600 | $2.500 | 84.8% | 83.8% | 89.8 |
| GPT-5 (medium) | OpenAI | $1.250 | $10.000 | 86.7% | 84.2% | - |
| o3 | OpenAI | $2.000 | $8.000 | 85.3% | 82.7% | 212.4 |
| Grok 4 | xAI | $3.000 | $15.000 | 86.6% | 87.7% | 54.3 |
| o3-pro | OpenAI | $20.000 | $80.000 | — | 84.5% | 49.9 |
| Gemini 3 Pro Preview (low) | $2.000 | $12.000 | 89.5% | 88.7% | - | |
| GPT-5 mini (high) | OpenAI | $0.250 | $2.000 | 83.7% | 82.8% | 73.2 |
| Grok 4.1 Fast (Reasoning) | xAI | $0.200 | $0.500 | 85.4% | 85.3% | 84.4 |
| Claude 4.5 Sonnet (Reasoning) | Anthropic | $3.000 | $15.000 | 87.5% | 83.4% | 69.5 |
| GPT-5 (low) | OpenAI | $1.250 | $10.000 | 86.0% | 80.8% | 129.5 |
| MiniMax-M2 | MiniMax | $0.300 | $1.200 | 82.0% | 77.7% | 108.4 |
| GPT-5 mini (medium) | OpenAI | $0.250 | $2.000 | 82.8% | 80.3% | 79.0 |
| gpt-oss-120B (high) | OpenAI | $0.150 | $0.600 | 80.8% | 78.2% | 353.7 |
| Grok 4 Fast (Reasoning) | xAI | $0.200 | $0.500 | 85.0% | 84.7% | 226.6 |
| Claude Opus 4.5 (Non-reasoning) | Anthropic | $5.000 | $25.000 | 88.9% | 81.0% | 65.3 |
| Gemini 2.5 Pro | $1.250 | $10.000 | 86.2% | 84.4% | 41.8 | |
| o4-mini (high) | OpenAI | $1.100 | $4.400 | 83.2% | 78.4% | - |
| Claude 4.1 Opus (Reasoning) | Anthropic | $15.000 | $75.000 | 88.0% | 80.9% | 38.9 |
| DeepSeek V3.1 Terminus (Reasoning) | DeepSeek | $0.400 | $2.000 | 85.1% | 79.2% | - |
| Qwen3 235B A22B 2507 (Reasoning) | Alibaba | $0.700 | $8.400 | 84.3% | 79.0% | 80.8 |
| Grok 3 mini Reasoning (high) | xAI | $0.300 | $0.500 | 82.8% | 79.1% | - |
| Doubao Seed Code | ByteDance Seed | $0.170 | $1.120 | 85.4% | 76.4% | - |
| DeepSeek V3.2 Exp (Reasoning) | DeepSeek | $0.280 | $0.420 | 85.0% | 79.7% | 29.1 |
| Claude 4 Sonnet (Reasoning) | Anthropic | $3.000 | $15.000 | 84.2% | 77.7% | 65.7 |
| GLM-4.6 (Reasoning) | Z AI | $0.600 | $2.200 | 82.9% | 78.0% | 112.0 |
| Qwen3 Max Thinking | Alibaba | $1.200 | $6.000 | 82.4% | 77.6% | 36.1 |
| Qwen3 Max | Alibaba | $1.200 | $6.000 | 84.1% | 76.4% | 27.9 |
| Claude 4.5 Haiku (Reasoning) | Anthropic | $1.000 | $5.000 | 76.0% | 67.2% | 83.9 |
| Gemini 2.5 Flash Preview (Sep '25) (Reasoning) | $0.300 | $2.500 | 84.2% | 79.3% | 142.3 | |
| Qwen3 VL 235B A22B (Reasoning) | Alibaba | $0.700 | $8.400 | 83.6% | 77.2% | 43.8 |
| Qwen3 Next 80B A3B (Reasoning) | Alibaba | $0.500 | $6.000 | 82.4% | 75.9% | - |
| Claude 4 Opus (Reasoning) | Anthropic | $15.000 | $75.000 | 87.3% | 79.6% | 41.2 |
| Gemini 2.5 Pro Preview (Mar' 25) | $1.250 | $10.000 | 85.8% | 83.6% | 39.8 | |
| DeepSeek V3.1 (Reasoning) | DeepSeek | $0.425 | $1.340 | 85.1% | 77.9% | - |
| Gemini 2.5 Pro Preview (May' 25) | $1.250 | $10.000 | 83.7% | 82.2% | 40.7 | |
| gpt-oss-20B (high) | OpenAI | $0.070 | $0.200 | 74.8% | 68.8% | 236.6 |
| Magistral Medium 1.2 | Mistral | $2.000 | $5.000 | 81.5% | 73.9% | 102.6 |
| DeepSeek R1 0528 (May '25) | DeepSeek | $1.350 | $4.000 | 84.9% | 81.3% | - |
| Qwen3 VL 32B (Reasoning) | Alibaba | $0.700 | $8.400 | 81.8% | 73.3% | 51.1 |
| Seed-OSS-36B-Instruct | ByteDance Seed | $0.210 | $0.570 | 81.5% | 72.6% | 27.1 |
| GLM-4.5 (Reasoning) | Z AI | $0.575 | $2.195 | 83.5% | 78.2% | 51.3 |
| Gemini 2.5 Flash (Reasoning) | $0.300 | $2.500 | 83.2% | 79.0% | 146.5 | |
| GPT-5 nano (high) | OpenAI | $0.050 | $0.400 | 78.0% | 67.6% | - |
| o3-mini (high) | OpenAI | $1.100 | $4.400 | 80.2% | 77.3% | 143.0 |
| Kimi K2 0905 | Moonshot AI | $0.990 | $2.500 | 81.9% | 76.7% | 93.9 |
| Claude 3.7 Sonnet (Reasoning) | Anthropic | $3.000 | $15.000 | 83.7% | 77.2% | - |
| Claude 4.5 Sonnet (Non-reasoning) | Anthropic | $3.000 | $15.000 | 86.0% | 72.7% | 72.4 |
| GPT-5 nano (medium) | OpenAI | $0.050 | $0.400 | 77.2% | 67.0% | - |
| GLM-4.5-Air | Z AI | $0.200 | $1.100 | 81.5% | 73.3% | 104.8 |
| Grok Code Fast 1 | xAI | $0.200 | $1.500 | 79.3% | 72.7% | 209.0 |
| Qwen3 Max (Preview) | Alibaba | $1.200 | $6.000 | 83.8% | 76.4% | 30.1 |
| Kimi K2 | Moonshot AI | $0.600 | $2.500 | 82.4% | 76.6% | 55.9 |
| o3-mini | OpenAI | $1.100 | $4.400 | 79.1% | 74.8% | 132.2 |
| o1-pro | OpenAI | $150.000 | $600.000 | — | — | - |
| Gemini 2.5 Flash-Lite Preview (Sep '25) (Reasoning) | $0.100 | $0.400 | 80.8% | 70.9% | 6.2 | |
| gpt-oss-120B (low) | OpenAI | $0.150 | $0.595 | 77.5% | 67.2% | 337.9 |
| o1 | OpenAI | $15.000 | $60.000 | 84.1% | 74.7% | 189.5 |
| Gemini 2.5 Flash Preview (Sep '25) (Non-reasoning) | $0.300 | $2.500 | 83.6% | 76.6% | 231.1 | |
| Qwen3 30B A3B 2507 (Reasoning) | Alibaba | $0.200 | $2.400 | 80.5% | 70.7% | 187.6 |
| DeepSeek V3.2 Exp (Non-reasoning) | DeepSeek | $0.280 | $0.420 | 83.6% | 73.8% | 26.8 |
| MiniMax M1 80k | MiniMax | $0.400 | $2.100 | 81.6% | 69.7% | - |
| DeepSeek V3.1 Terminus (Non-reasoning) | DeepSeek | $0.400 | $1.680 | 83.6% | 75.1% | - |
| Qwen3 235B A22B 2507 Instruct | Alibaba | $0.700 | $2.800 | 82.8% | 75.3% | 49.5 |
| Grok 3 | xAI | $3.000 | $15.000 | 79.9% | 69.3% | 53.3 |
| Qwen3 VL 30B A3B (Reasoning) | Alibaba | $0.200 | $2.400 | 80.7% | 72.0% | 106.5 |
| Llama Nemotron Super 49B v1.5 (Reasoning) | NVIDIA | $0.100 | $0.400 | 81.4% | 74.8% | 77.7 |
| o1-preview | OpenAI | $16.500 | $66.000 | — | — | - |
| Qwen3 Next 80B A3B Instruct | Alibaba | $0.500 | $2.000 | 81.9% | 73.8% | 158.6 |
| DeepSeek V3.1 (Non-reasoning) | DeepSeek | $0.560 | $1.660 | 83.3% | 73.5% | - |
| Ling-1T | InclusionAI | $0.570 | $2.280 | 82.2% | 71.9% | - |
| GLM-4.6 (Non-reasoning) | Z AI | $0.600 | $2.200 | 78.4% | 63.2% | 44.2 |
| Claude 4.1 Opus (Non-reasoning) | Anthropic | $15.000 | $75.000 | — | — | 38.7 |
| Claude 4 Sonnet (Non-reasoning) | Anthropic | $3.000 | $15.000 | 83.7% | 68.3% | 60.0 |
| gpt-oss-20B (low) | OpenAI | $0.070 | $0.200 | 71.8% | 61.1% | 234.0 |
| Qwen3 VL 235B A22B Instruct | Alibaba | $0.700 | $2.800 | 82.3% | 71.2% | 34.5 |
| DeepSeek R1 (Jan '25) | DeepSeek | $1.350 | $4.000 | 84.4% | 70.8% | - |
| GPT-5 (minimal) | OpenAI | $1.250 | $10.000 | 80.6% | 67.3% | 81.1 |
| GPT-4.1 | OpenAI | $2.000 | $8.000 | 80.6% | 66.6% | 123.2 |
| Magistral Small 1.2 | Mistral | $0.500 | $1.500 | 76.8% | 66.3% | 194.8 |
| GPT-5.1 (Non-reasoning) | OpenAI | $1.250 | $10.000 | 80.1% | 64.3% | 89.8 |
| EXAONE 4.0 32B (Reasoning) | LG AI Research | $0.600 | $1.000 | 81.8% | 73.9% | 106.4 |
| GPT-4.1 mini | OpenAI | $0.400 | $1.600 | 78.1% | 66.4% | 77.8 |
| Claude 4 Opus (Non-reasoning) | Anthropic | $15.000 | $75.000 | 86.0% | 70.1% | 40.7 |
| Qwen3 Coder 480B A35B Instruct | Alibaba | $1.500 | $7.500 | 78.8% | 61.8% | 49.3 |
| GPT-5 (ChatGPT) | OpenAI | $1.250 | $10.000 | 82.0% | 68.6% | 165.8 |
| Ring-1T | InclusionAI | $0.570 | $2.280 | 80.6% | 59.5% | - |
| Qwen3 235B A22B (Reasoning) | Alibaba | $0.700 | $8.400 | 82.8% | 70.0% | 59.9 |
| Claude 4.5 Haiku (Non-reasoning) | Anthropic | $1.000 | $5.000 | 80.0% | 64.6% | 86.6 |
| GPT-5 mini (minimal) | OpenAI | $0.250 | $2.000 | 77.5% | 68.7% | 72.2 |
| Gemini 2.5 Flash-Lite Preview (Sep '25) (Non-reasoning) | $0.100 | $0.400 | 79.6% | 65.1% | 487.7 | |
| Hermes 4 - Llama-3.1 405B (Reasoning) | Nous Research | $1.000 | $3.000 | 82.9% | 72.7% | 36.1 |
| DeepSeek V3 0324 | DeepSeek | $1.140 | $1.250 | 81.9% | 65.5% | - |
| Claude 3.7 Sonnet (Non-reasoning) | Anthropic | $3.000 | $15.000 | 80.3% | 65.6% | - |
| Qwen3 VL 32B Instruct | Alibaba | $0.700 | $2.800 | 79.1% | 67.1% | 44.7 |
| Gemini 2.5 Flash (Non-reasoning) | $0.300 | $2.500 | 80.9% | 68.3% | 253.5 | |
| Gemini 2.5 Flash-Lite (Reasoning) | $0.100 | $0.400 | 75.9% | 62.5% | - | |
| MiniMax M1 40k | MiniMax | $0.400 | $2.100 | 80.8% | 68.2% | - |
| Qwen3 Omni 30B A3B (Reasoning) | Alibaba | $0.250 | $0.970 | 79.2% | 72.6% | 96.7 |
| Ring-flash-2.0 | InclusionAI | $0.140 | $0.570 | 79.3% | 72.5% | 52.6 |
| Hermes 4 - Llama-3.1 70B (Reasoning) | Nous Research | $0.130 | $0.400 | 81.1% | 69.9% | 75.0 |
| Qwen3 32B (Reasoning) | Alibaba | $0.700 | $8.400 | 79.8% | 66.8% | 98.3 |
| Grok 4 Fast (Non-reasoning) | xAI | $0.200 | $0.500 | 73.0% | 60.6% | 216.5 |
| Qwen3 VL 30B A3B Instruct | Alibaba | $0.200 | $0.800 | 76.4% | 69.5% | 92.4 |
| Llama 3.1 Nemotron Ultra 253B v1 (Reasoning) | NVIDIA | $0.600 | $1.800 | 82.5% | 72.8% | 38.4 |
| Grok 4.1 Fast (Non-reasoning) | xAI | $0.200 | $0.500 | 74.3% | 63.7% | 68.7 |
| Ling-flash-2.0 | InclusionAI | $0.140 | $0.570 | 77.7% | 65.7% | 54.6 |
| QwQ 32B | Alibaba | $0.430 | $0.600 | 76.4% | 59.3% | 28.4 |
| Solar Pro 2 (Reasoning) | Upstage | $0.500 | $0.500 | 80.5% | 68.7% | 111.3 |
| NVIDIA Nemotron Nano 9B V2 (Reasoning) | NVIDIA | $0.040 | $0.160 | 74.2% | 57.0% | 98.3 |
| GLM-4.5V (Reasoning) | Z AI | $0.550 | $1.750 | 78.8% | 68.4% | 74.8 |
| Qwen3 30B A3B 2507 Instruct | Alibaba | $0.200 | $0.800 | 77.7% | 65.9% | 61.1 |
| Qwen3 30B A3B (Reasoning) | Alibaba | $0.200 | $2.400 | 77.7% | 61.6% | 79.5 |
| NVIDIA Nemotron Nano 9B V2 (Non-reasoning) | NVIDIA | $0.040 | $0.160 | 73.9% | 55.7% | 96.8 |
| Qwen3 14B (Reasoning) | Alibaba | $0.350 | $4.200 | 77.4% | 60.4% | 58.6 |
| Llama 4 Maverick | Meta | $0.270 | $0.850 | 80.9% | 67.1% | 130.0 |
| GPT-4o (March 2025, chatgpt-4o-latest) | OpenAI | $5.000 | $15.000 | 80.3% | 65.5% | 272.6 |
| Mistral Medium 3.1 | Mistral | $0.400 | $2.000 | 68.3% | 58.8% | 59.6 |
| Sonar Reasoning | Perplexity | $1.000 | $5.000 | — | 62.3% | 69.1 |
| Gemini 2.0 Flash (Feb '25) | $0.100 | $0.400 | 77.9% | 62.3% | 181.8 | |
| Mistral Medium 3 | Mistral | $0.400 | $2.000 | 76.0% | 57.8% | 33.3 |
| Qwen3 Coder 30B A3B Instruct | Alibaba | $0.450 | $2.250 | 70.6% | 51.6% | 88.7 |
| Magistral Medium 1 | Mistral | $2.000 | $5.000 | 75.3% | 67.9% | 148.3 |
| ERNIE 4.5 300B A47B | Baidu | $0.280 | $1.100 | 77.6% | 81.1% | 24.4 |
| DeepSeek R1 Distill Qwen 32B | DeepSeek | $0.285 | $0.285 | 73.9% | 61.5% | 92.1 |
| Hermes 4 - Llama-3.1 405B (Non-reasoning) | Nous Research | $1.000 | $3.000 | 72.9% | 53.6% | 32.2 |
| DeepSeek V3 (Dec '24) | DeepSeek | $0.400 | $0.890 | 75.2% | 55.7% | - |
| Nova Premier | Amazon | $2.500 | $12.500 | 73.3% | 56.9% | 57.2 |
| Qwen3 VL 8B (Reasoning) | Alibaba | $0.180 | $2.100 | 74.9% | 57.9% | 63.7 |
| Magistral Small 1 | Mistral | $0.500 | $1.500 | 74.6% | 64.1% | 259.0 |
| DeepSeek R1 0528 Qwen3 8B | DeepSeek | $0.060 | $0.090 | 73.9% | 61.2% | 78.6 |
| Qwen2.5 Max | Alibaba | $1.600 | $6.400 | 76.2% | 58.7% | 28.3 |
| EXAONE 4.0 32B (Non-reasoning) | LG AI Research | $0.600 | $1.000 | 76.8% | 62.8% | 87.7 |
| Solar Pro 2 (Non-reasoning) | Upstage | $0.500 | $0.500 | 75.0% | 56.1% | 107.6 |
| Qwen3 Omni 30B A3B Instruct | Alibaba | $0.250 | $0.970 | 72.5% | 62.0% | 91.1 |
| Gemini 2.5 Flash-Lite (Non-reasoning) | $0.100 | $0.400 | 72.4% | 47.4% | 263.0 | |
| Qwen3 235B A22B (Non-reasoning) | Alibaba | $0.700 | $2.800 | 76.2% | 61.3% | 55.9 |
| DeepSeek R1 Distill Llama 70B | DeepSeek | $0.800 | $1.050 | 79.5% | 40.2% | 102.2 |
| Claude 3.5 Sonnet (Oct '24) | Anthropic | $3.000 | $15.000 | 77.2% | 59.9% | - |
| DeepSeek R1 Distill Qwen 14B | DeepSeek | $0.150 | $0.150 | 74.0% | 48.4% | 41.9 |
| Qwen3 14B (Non-reasoning) | Alibaba | $0.350 | $1.400 | 67.5% | 47.0% | 53.5 |
| Mistral Small 3.2 | Mistral | $0.100 | $0.300 | 68.1% | 50.5% | 134.6 |
| GPT-5 nano (minimal) | OpenAI | $0.050 | $0.400 | 55.6% | 42.8% | 118.8 |
| GPT-4o (Aug '24) | OpenAI | $2.500 | $10.000 | — | 52.1% | 118.0 |
| Sonar | Perplexity | $1.000 | $1.000 | 68.9% | 47.1% | 75.6 |
| Qwen3 8B (Reasoning) | Alibaba | $0.180 | $2.100 | 74.3% | 58.9% | 86.6 |
| MiniMax-Text-01 | MiniMax | $0.200 | $1.100 | 75.9% | 57.8% | 27.6 |
| Sonar Pro | Perplexity | $3.000 | $15.000 | 75.5% | 57.8% | 85.8 |
| Llama 3.1 Instruct 405B | Meta | $3.750 | $6.750 | 73.2% | 51.5% | 24.4 |
| Llama 4 Scout | Meta | $0.140 | $0.545 | 75.2% | 58.7% | 136.6 |
| QwQ 32B-Preview | Alibaba | $0.120 | $0.180 | 64.8% | 55.7% | 108.7 |
| Llama 3.3 Instruct 70B | Meta | $0.540 | $0.710 | 71.3% | 49.8% | 99.2 |
| Devstral Medium | Mistral | $0.400 | $2.000 | 70.8% | 49.2% | 111.5 |
| Ling-mini-2.0 | InclusionAI | $0.070 | $0.280 | 67.1% | 56.2% | 150.3 |
| GPT-4.1 nano | OpenAI | $0.100 | $0.400 | 65.7% | 51.2% | 149.8 |
| Devstral Small (Jul '25) | Mistral | $0.100 | $0.300 | 62.2% | 41.4% | 241.5 |
| Qwen3 VL 8B Instruct | Alibaba | $0.180 | $0.700 | 68.6% | 42.7% | 90.0 |
| GPT-4o (Nov '24) | OpenAI | $2.500 | $10.000 | 74.8% | 54.3% | 189.6 |
| Command A | Cohere | $2.500 | $10.000 | 71.2% | 52.7% | 55.0 |
| Mistral Large 2 (Nov '24) | Mistral | $2.000 | $6.000 | 69.7% | 48.6% | 38.1 |
| Gemini 2.0 Flash-Lite (Feb '25) | $0.075 | $0.300 | 72.4% | 53.5% | 188.1 | |
| Llama Nemotron Super 49B v1.5 (Non-reasoning) | NVIDIA | $0.100 | $0.400 | 69.2% | 48.1% | 68.7 |
| Qwen3 30B A3B (Non-reasoning) | Alibaba | $0.200 | $0.800 | 71.0% | 51.5% | 72.6 |
| Qwen3 32B (Non-reasoning) | Alibaba | $0.700 | $2.800 | 72.7% | 53.5% | 90.8 |
| GPT-4o (May '24) | OpenAI | $5.000 | $15.000 | 74.0% | 52.6% | 115.6 |
| Gemini 2.0 Flash-Lite (Preview) | $0.075 | $0.300 | — | 54.2% | 184.5 | |
| GLM-4.5V (Non-reasoning) | Z AI | $0.600 | $1.800 | 75.1% | 57.3% | 74.3 |
| Reka Flash 3 | Reka AI | $0.200 | $0.800 | 66.9% | 52.9% | 51.8 |
| Qwen3 4B (Reasoning) | Alibaba | $0.110 | $1.260 | 69.6% | 52.2% | 84.0 |
| Claude 3.5 Sonnet (June '24) | Anthropic | $3.000 | $15.000 | 75.1% | 56.0% | - |
| GPT-4o (ChatGPT) | OpenAI | $5.000 | $15.000 | 77.3% | 51.1% | 266.8 |
| Pixtral Large | Mistral | $2.000 | $6.000 | 70.1% | 50.5% | 36.7 |
| Nova Pro | Amazon | $0.800 | $3.200 | 69.1% | 49.9% | - |
| Mistral Small 3.1 | Mistral | $0.100 | $0.300 | 65.9% | 45.4% | 160.5 |
| Grok 2 (Dec '24) | xAI | $2.000 | $10.000 | 70.9% | 51.0% | 94.6 |
| GPT-4 Turbo | OpenAI | $10.000 | $30.000 | 69.4% | — | 38.4 |
| Hermes 4 - Llama-3.1 70B (Non-reasoning) | Nous Research | $0.130 | $0.400 | 66.4% | 49.1% | 66.8 |
| Llama 3.1 Nemotron Instruct 70B | NVIDIA | $0.600 | $0.600 | 69.0% | 46.5% | 38.7 |
| Qwen3 8B (Non-reasoning) | Alibaba | $0.180 | $0.700 | 64.3% | 45.2% | 81.6 |
| Granite 4.0 H Small | IBM | $0.060 | $0.250 | 62.4% | 41.6% | 395.0 |
| Phi-4 | Microsoft Azure | $0.125 | $0.500 | 71.4% | 57.5% | 11.5 |
| Llama 3.1 Instruct 70B | Meta | $0.560 | $0.560 | 67.6% | 40.9% | 35.1 |
| Qwen3 1.7B (Reasoning) | Alibaba | $0.110 | $1.260 | 57.0% | 35.6% | 123.6 |
| Mistral Large 2 (Jul '24) | Mistral | $2.000 | $6.000 | 68.3% | 47.2% | - |
| CompactifAI Llama 4 Scout Slim | Multiverse Computing | $0.070 | $0.100 | 70.3% | 42.6% | 116.7 |
| Qwen2.5 Coder Instruct 32B | Alibaba | $0.130 | $0.175 | 63.5% | 41.7% | 54.1 |
| Nova Lite | Amazon | $0.060 | $0.240 | 59.0% | 43.3% | 150.5 |
| GPT-4 | OpenAI | $30.000 | $60.000 | — | — | 30.1 |
| Mistral Small 3 | Mistral | $0.100 | $0.300 | 65.2% | 46.2% | 176.8 |
| GPT-4o mini | OpenAI | $0.150 | $0.600 | 64.8% | 42.6% | 54.5 |
| Jamba 1.7 Large | AI21 Labs | $2.000 | $8.000 | 57.7% | 39.0% | 34.7 |
| Qwen3 4B (Non-reasoning) | Alibaba | $0.110 | $0.420 | 58.6% | 39.8% | 80.1 |
| Claude 3 Opus | Anthropic | $15.000 | $75.000 | 69.6% | 48.9% | - |
| Claude 3.5 Haiku | Anthropic | $0.800 | $4.000 | 63.4% | 40.8% | 48.2 |
| Codestral (Jan '25) | Mistral | $0.300 | $0.900 | 44.6% | 31.2% | 235.0 |
| Devstral Small (May '25) | Mistral | $0.100 | $0.300 | 63.2% | 43.4% | 203.9 |
| Reka Core | Reka AI | $2.000 | $2.000 | — | — | 40.6 |
| Qwen2.5 Turbo | Alibaba | $0.050 | $0.200 | 63.3% | 41.0% | 74.2 |
| Reka Flash (Sep '24) | Reka AI | $0.200 | $0.800 | — | — | 68.8 |
| Solar Mini | Upstage | $0.150 | $0.150 | — | — | 79.2 |
| Llama 3.2 Instruct 90B (Vision) | Meta | $0.720 | $0.720 | 67.1% | 43.2% | 36.8 |
| Reka Flash (Feb '24) | Reka AI | $0.200 | $0.800 | — | — | 69.9 |
| Reka Edge | Reka AI | $0.100 | $0.100 | — | — | 63.3 |
| Nova Micro | Amazon | $0.035 | $0.140 | 53.1% | 35.8% | 263.9 |
| Llama 3.1 Instruct 8B | Meta | $0.100 | $0.100 | 47.6% | 25.9% | 174.3 |
| CompactifAI Mistral Small 3.1 Slim | Multiverse Computing | $0.050 | $0.080 | 53.8% | 32.9% | 121.6 |
| CompactifAI Llama 3.3 70B Slim | Multiverse Computing | $0.160 | $0.310 | 57.1% | 35.5% | 130.7 |
| Llama 3.2 Instruct 11B (Vision) | Meta | $0.160 | $0.160 | 46.4% | 22.1% | 69.3 |
| Gemma 3n E4B Instruct | $0.020 | $0.040 | 48.8% | 29.6% | 41.3 | |
| Granite 3.3 8B (Non-reasoning) | IBM | $0.030 | $0.250 | 46.8% | 33.8% | 458.0 |
| Jamba 1.7 Mini | AI21 Labs | $0.200 | $0.400 | 38.8% | 32.2% | 153.5 |
| Jamba 1.5 Large | AI21 Labs | $2.000 | $8.000 | 57.2% | 42.7% | - |
| Hermes 3 - Llama-3.1 70B | Nous Research | $0.300 | $0.300 | 57.1% | 40.1% | 37.8 |
| OLMo 2 32B | Allen Institute for AI | $0.200 | $0.350 | 51.1% | 32.8% | - |
| Phi-3 Medium Instruct 14B | Microsoft Azure | $0.170 | $0.680 | 54.3% | 32.6% | 42.3 |
| Qwen3 1.7B (Non-reasoning) | Alibaba | $0.110 | $0.420 | 41.1% | 28.3% | 115.3 |
| Jamba 1.6 Large | AI21 Labs | $2.000 | $8.000 | 56.5% | 38.7% | 34.9 |
| Qwen3 0.6B (Reasoning) | Alibaba | $0.110 | $1.260 | 34.7% | 23.9% | 201.3 |
| Aya Expanse 32B | Cohere | $0.500 | $1.500 | 37.7% | 23.0% | 41.9 |
| Claude 3 Sonnet | Anthropic | $3.000 | $15.000 | 57.9% | 40.0% | - |
| Llama 3 Instruct 70B | Meta | $0.650 | $0.880 | 57.4% | 37.9% | 40.5 |
| Mistral Small (Sep '24) | Mistral | $0.200 | $0.600 | 52.9% | 38.1% | 84.9 |
| Phi-3 Mini Instruct 3.8B | Microsoft Azure | $0.130 | $0.520 | 43.5% | 31.9% | 69.1 |
| Ministral 8B | Mistral | $0.100 | $0.100 | 38.9% | 27.6% | 195.9 |
| Mistral Large (Feb '24) | Mistral | $4.000 | $12.000 | 51.5% | 35.1% | - |
| Llama 2 Chat 7B | Meta | $0.050 | $0.250 | 16.4% | 22.7% | 112.2 |
| CompactifAI Llama 3.1 8B Slim | Multiverse Computing | $0.050 | $0.070 | 32.1% | 22.1% | 228.4 |
| Llama 3.2 Instruct 3B | Meta | $0.060 | $0.060 | 34.7% | 25.5% | 114.4 |
| Qwen3 0.6B (Non-reasoning) | Alibaba | $0.110 | $0.420 | 23.1% | 23.1% | 191.3 |
| Ministral 3B | Mistral | $0.040 | $0.040 | 33.9% | 26.0% | 272.6 |
| Aya Expanse 8B | Cohere | $0.500 | $1.500 | 31.2% | 24.7% | 80.9 |
| Claude 3 Haiku | Anthropic | $0.250 | $1.250 | — | — | 116.9 |
| Llama 3.2 Instruct 1B | Meta | $0.053 | $0.055 | 20.0% | 19.6% | 74.6 |
| Pixtral 12B (2409) | Mistral | $0.150 | $0.150 | 47.3% | 34.3% | 144.9 |
| Mistral Small (Feb '24) | Mistral | $1.000 | $3.000 | 41.9% | 30.2% | 163.9 |
| Mistral Medium | Mistral | $2.750 | $8.100 | 49.1% | 34.9% | 64.3 |
| GPT-3.5 Turbo | OpenAI | $0.500 | $1.500 | 46.2% | 29.7% | 84.9 |
| Gemma 2 9B | $0.030 | $0.090 | 49.5% | 31.1% | - | |
| Command-R+ (Aug '24) | Cohere | $2.500 | $10.000 | 42.7% | 33.7% | 20.4 |
| Llama 3 Instruct 8B | Meta | $0.045 | $0.155 | 40.5% | 29.6% | 67.6 |
| Command-R+ (Apr '24) | Cohere | $3.000 | $15.000 | 43.2% | 32.3% | - |
| Mistral NeMo | Mistral | $0.150 | $0.150 | 39.9% | 31.4% | 188.3 |
| Jamba 1.5 Mini | AI21 Labs | $0.200 | $0.400 | 37.1% | 30.2% | - |
| Jamba 1.6 Mini | AI21 Labs | $0.200 | $0.400 | 36.7% | 30.0% | 150.6 |
| Mixtral 8x7B Instruct | Mistral | $0.540 | $0.600 | 38.7% | 29.2% | - |
| Command-R (Mar '24) | Cohere | $0.500 | $1.500 | 33.8% | 28.4% | - |
| Command-R (Aug '24) | Cohere | $0.150 | $0.600 | 33.7% | 28.9% | 58.9 |
| Mistral 7B Instruct | Mistral | $0.250 | $0.250 | 24.5% | 17.7% | 119.4 |
| Cogito v2.1 (Reasoning) | Deep Cogito | $1.250 | $1.250 | 84.9% | 76.8% | 75.3 |
| DeepSeek-OCR | DeepSeek | $0.030 | $0.100 | — | — | 305.1 |
| Grok 3 mini Reasoning (Low) | xAI | $0.300 | $0.500 | — | — | 110.8 |