Skytells
HomeModelsCLIChangelog
  • Home
  • Models
  • CLI
  • Changelog
Skytells

Addressing the world's greatest challenges with AI. Enterprise research, foundation models, and infrastructure trusted by organizations worldwide since 2012.

Get Started

  • Console
  • Learn
  • Documentation
  • API Reference
  • Pricing
  • ModelsNew

Platform

  • Cloud AgentsNew
  • AI Solutions
  • Infrastructure
  • Edge Network
  • Trust Center
  • CLI

Resources

  • Blog
  • Changelog
  • AI Leaderboard
  • Research
  • Status

Company

  • About
  • Careers
  • Legal
  • Privacy Policy

© 2012–2026 Skytells, Inc. All rights reserved.

Live rankings

AI Model Leaderboard

Every major AI model ranked across benchmark quality, inference speed, agentic capability, programming aptitude, and cost efficiency — updated continuously from published evaluation data.

Explore full leaderboardBrowse model catalog

296

Tracked models

27

Providers

253

Benchmarked

34.7

Avg. index

OverallBenchmarksInferenceAgenticProgrammingValue / Price

296 models

RankModelProviderScoreBenchmarksInferenceAgenticProgrammingValuePrice
221

DeepSeek-R1

deepseek-r1

textinference
DeepSeek

22.3

overall

0.014.30.00.035.1$0.55 in / $2.19 out
222

Magistral Medium

magistral-medium

multimodalvisionmulti-input reasoning
Mistral AI

22.2

overall

22.20.00.00.00.0
223

Mistral Large 3 (675B Base)

mistral-large-3-675b-base-2512

multimodalvisionmulti-input reasoning
Mistral AI

22.2

overall

22.20.00.00.00.0
224

Mistral Large 3 (675B Instruct 2512 Eagle)

mistral-large-3-675B-instruct-2512-eagle

multimodalvisionmulti-input reasoning
Mistral AI

22.2

overall

22.20.00.00.00.0
225

Mistral Large 3 (675B Instruct 2512 NVFP4)

mistral-large-3-675b-instruct-2512-nvfp4

multimodalvisionmulti-input reasoning
Mistral AI

22.2

overall

22.20.00.00.00.0
226

Llama 3.2 90B Instruct

llama-3.2-90b-instruct

multimodalvisionmulti-input reasoning
MMeta

22.0

overall

16.311.30.00.054.9$0.35 in / $0.4 out
227

Sarvam-30B

sarvam-30b

codeprogrammingtool use
SSarvam AI

21.7

overall

46.40.08.25.20.0N/A
228

Phi 4 Mini Reasoning

phi-4-mini-reasoning

textinference
MMicrosoft

21.7

overall

21.70.00.00.00.0N/A
229

DeepSeek-R1-0528

deepseek-r1-0528

codeprogrammingtool use
DeepSeek

21.3

overall

50.114.30.06.635.1$0.55 in / $2.19 out
230

GPT-3.5 Turbo

gpt-3.5-turbo-0125

multimodalvisionmulti-input reasoning
OpenAI

21.3

overall

2.536.70.00.049.4
231

Mistral Small

mistral-small-2409

textinference
Mistral AI

21.3

overall

0.02.10.00.051.9$0.2 in / $0.6 out
232

Pixtral Large

pixtral-large

multimodalvisionmulti-input reasoning
Mistral AI

20.6

overall

27.87.00.00.022.4
233

GPT-5 nano

gpt-5-nano-2025-08-07

multimodalvisionmulti-input reasoning
OpenAI

19.9

overall

26.30.00.011.80.0
234

Pixtral-12B

pixtral-12b-2409

multimodalvisionmulti-input reasoning
Mistral AI

19.8

overall

8.17.00.00.073.0
235

Gemma 4 E4B

gemma-4-e4b-it

multimodalvisionmulti-input reasoning
Google

19.2

overall

19.20.00.00.00.0N/A
236

Mistral Small 3.2 24B Instruct

mistral-small-3.2-24b-instruct-2506

multimodalvisionmulti-input reasoning
Mistral AI

19.1

overall

19.10.00.00.00.0
237

Phi-3.5-mini-instruct

phi-3.5-mini-instruct

multimodalvisionmulti-input reasoning
MMicrosoft

18.9

overall

2.710.80.00.077.2$0.1 in / $0.1 out
238

Jamba 1.5 Large

jamba-1.5-large

textinference
AAI21 Labs

18.8

overall

8.133.60.00.025.2$2 in / $8 out
239

Qwen2.5 32B Instruct

qwen-2.5-32b-instruct

textinference
AAlibaba Cloud / Qwen Team

18.6

overall

18.60.00.00.00.0N/A
240

DeepSeek R1 Distill Qwen 7B

deepseek-r1-distill-qwen-7b

textinference
DeepSeek

18.3

overall

18.30.00.00.00.0N/A
221

DeepSeek-R1

DeepSeek

22.3

$0.55 in / $2.19 out

222

Magistral Medium

Mistral AI

22.2

N/A

223

Mistral Large 3 (675B Base)

Mistral AI

22.2

N/A

224

Page 12 of 15 · 296 models

PreviousNext

Want benchmark charts, model comparison, and pricing analytics?

Sign in to access the full interactive leaderboard with deep benchmark breakdowns and model comparison tools.

Open full leaderboard

Rankings are based on multi-dimensional evaluation across benchmark quality, inference efficiency, and cost-per-output. Scores are updated continuously and may differ from individual third-party benchmarks.

N/A
N/A
N/A
N/A
$0.5 in / $1.5 out
$2 in / $6 out
N/A
$0.15 in / $0.15 out
N/A

Mistral Large 3 (675B Instruct 2512 Eagle)

Mistral AI

22.2

N/A

225

Mistral Large 3 (675B Instruct 2512 NVFP4)

Mistral AI

22.2

N/A

226
M

Llama 3.2 90B Instruct

Meta

22.0

$0.35 in / $0.4 out

227
S

Sarvam-30B

Sarvam AI

21.7

N/A

228
M

Phi 4 Mini Reasoning

Microsoft

21.7

N/A

229

DeepSeek-R1-0528

DeepSeek

21.3

$0.55 in / $2.19 out

230

GPT-3.5 Turbo

OpenAI

21.3

$0.5 in / $1.5 out

231

Mistral Small

Mistral AI

21.3

$0.2 in / $0.6 out

232

Pixtral Large

Mistral AI

20.6

$2 in / $6 out

233

GPT-5 nano

OpenAI

19.9

N/A

234

Pixtral-12B

Mistral AI

19.8

$0.15 in / $0.15 out

235

Gemma 4 E4B

Google

19.2

N/A

236

Mistral Small 3.2 24B Instruct

Mistral AI

19.1

N/A

237
M

Phi-3.5-mini-instruct

Microsoft

18.9

$0.1 in / $0.1 out

238
A

Jamba 1.5 Large

AI21 Labs

18.8

$2 in / $8 out

239
A

Qwen2.5 32B Instruct

Alibaba Cloud / Qwen Team

18.6

N/A

240

DeepSeek R1 Distill Qwen 7B

DeepSeek

18.3

N/A