Every major AI model ranked across benchmark quality, inference speed, agentic capability, programming aptitude, and cost efficiency — updated continuously from published evaluation data.
296
Tracked models
27
Providers
253
Benchmarked
13.4
Avg. index
296 models
| Rank | Model | Provider | Score | Benchmarks | Inference | Agentic | Programming | Value | Price |
|---|---|---|---|---|---|---|---|---|---|
| 101 | Gemini Diffusion gemini-diffusion codeprogrammingtool use | Google | 1.7 Programming | 7.0 | 0.0 | 0.0 | 1.7 | 0.0 | N/A |
| 102 | DeepSeek-V2.5 deepseek-v2.5 codeprogrammingtool use | DeepSeek | 0.9 Programming | 0.0 | 46.5 | 0.0 | 0.9 | 79.7 | $0.14 in / $0.28 out |
| 103 | ChatGPT-4o Latest chatgpt-4o-latest multimodalvisionmulti-input reasoning | OpenAI | 0.0 Programming | 56.0 | 63.8 | 0.0 | 0.0 | 32.0 | |
| 104 | Claude 3.5 Sonnet claude-3-5-sonnet-20240620 multimodalvisionmulti-input reasoning | Anthropic | 0.0 Programming | 25.4 | 68.2 | 0.0 | 0.0 | 24.6 | |
| 105 | Claude 3 Haiku claude-3-haiku-20240307 multimodalvisionmulti-input reasoning | Anthropic | 0.0 Programming | 5.8 | 61.8 | 0.0 | 0.0 | 57.9 | |
| 106 | Claude 3 Opus claude-3-opus-20240229 multimodalvisionmulti-input reasoning | Anthropic | 0.0 Programming | 19.3 | 71.7 | 0.0 | 0.0 | 19.5 | |
| 107 | Claude 3 Sonnet claude-3-sonnet-20240229 multimodalvisionmulti-input reasoning | Anthropic | 0.0 Programming | 10.0 | 30.5 | 0.0 | 0.0 | 13.3 | |
| 108 | Codestral-22B codestral-22b textinference | Mistral AI | 0.0 Programming | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 109 | Command R+ command-r-plus-04-2024 textinference | Cohere | 0.0 Programming | 0.0 | 32.5 | 0.0 | 0.0 | 55.4 | $0.25 in / $1 out |
| 110 | DeepSeek-V3.2 (Non-thinking) deepseek-chat textinference | DeepSeek | 0.0 Programming | 0.0 | 58.0 | 0.0 | 0.0 | 70.2 | $0.28 in / $0.42 out |
| 111 | DeepSeek-R1 deepseek-r1 textinference | DeepSeek | 0.0 Programming | 0.0 | 14.3 | 0.0 | 0.0 | 35.1 | $0.55 in / $2.19 out |
| 112 | DeepSeek R1 Distill Llama 70B deepseek-r1-distill-llama-70b textinference | DeepSeek | 0.0 Programming | 28.8 | 16.6 | 0.0 | 0.0 | 66.6 | $0.1 in / $0.4 out |
| 113 | DeepSeek R1 Distill Llama 8B deepseek-r1-distill-llama-8b textinference | DeepSeek | 0.0 Programming | 17.8 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 114 | DeepSeek R1 Distill Qwen 14B deepseek-r1-distill-qwen-14b textinference | DeepSeek | 0.0 Programming | 24.7 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 115 | DeepSeek R1 Distill Qwen 1.5B deepseek-r1-distill-qwen-1.5b textinference | DeepSeek | 0.0 Programming | 6.1 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 116 | DeepSeek R1 Distill Qwen 32B deepseek-r1-distill-qwen-32b textinference | DeepSeek | 0.0 Programming | 26.6 | 16.6 | 0.0 | 0.0 | 75.9 | $0.12 in / $0.18 out |
| 117 | DeepSeek R1 Distill Qwen 7B deepseek-r1-distill-qwen-7b textinference | DeepSeek | 0.0 Programming | 18.3 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 118 | DeepSeek R1 Zero deepseek-r1-zero textinference | DeepSeek | 0.0 Programming | 39.4 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
| 119 | DeepSeek-V3 0324 deepseek-v3-0324 textinference | DeepSeek | 0.0 Programming | 32.8 | 39.8 | 0.0 | 0.0 | 57.7 | $0.28 in / $1.14 out |
| 120 | DeepSeek VL2 deepseek-vl2 multimodalvisionmulti-input reasoning | DeepSeek | 0.0 Programming | 6.9 | 0.0 | 0.0 | 0.0 | 0.0 | N/A |
Gemini Diffusion
1.7
N/A
DeepSeek-V2.5
DeepSeek
0.9
$0.14 in / $0.28 out
ChatGPT-4o Latest
OpenAI
0.0
$2.5 in / $10 out
Want benchmark charts, model comparison, and pricing analytics?
Sign in to access the full interactive leaderboard with deep benchmark breakdowns and model comparison tools.
Open full leaderboardRankings are based on multi-dimensional evaluation across benchmark quality, inference efficiency, and cost-per-output. Scores are updated continuously and may differ from individual third-party benchmarks.
| $2.5 in / $10 out |
| $3 in / $15 out |
| $0.25 in / $1.25 out |
| $15 in / $75 out |
| $3 in / $15 out |
Claude 3.5 Sonnet
Anthropic
0.0
$3 in / $15 out
Claude 3 Haiku
Anthropic
0.0
$0.25 in / $1.25 out
Claude 3 Opus
Anthropic
0.0
$15 in / $75 out
Claude 3 Sonnet
Anthropic
0.0
$3 in / $15 out
Codestral-22B
Mistral AI
0.0
N/A
Command R+
Cohere
0.0
$0.25 in / $1 out
DeepSeek-V3.2 (Non-thinking)
DeepSeek
0.0
$0.28 in / $0.42 out
DeepSeek-R1
DeepSeek
0.0
$0.55 in / $2.19 out
DeepSeek R1 Distill Llama 70B
DeepSeek
0.0
$0.1 in / $0.4 out
DeepSeek R1 Distill Llama 8B
DeepSeek
0.0
N/A
DeepSeek R1 Distill Qwen 14B
DeepSeek
0.0
N/A
DeepSeek R1 Distill Qwen 1.5B
DeepSeek
0.0
N/A
DeepSeek R1 Distill Qwen 32B
DeepSeek
0.0
$0.12 in / $0.18 out
DeepSeek R1 Distill Qwen 7B
DeepSeek
0.0
N/A
DeepSeek R1 Zero
DeepSeek
0.0
N/A
DeepSeek-V3 0324
DeepSeek
0.0
$0.28 in / $1.14 out
DeepSeek VL2
DeepSeek
0.0
N/A