DataLearner logoDataLearnerAI
Latest AI Insights
Model Leaderboards
Benchmarks
Model Directory
Model Comparison
Resource Center
Tools
LanguageEnglish
DataLearner logoDataLearner AI

A knowledge platform focused on LLM benchmarking, datasets, and practical instruction with continuously updated capability maps.

Products

  • Leaderboards
  • Model comparison
  • Datasets

Resources

  • Tutorials
  • Editorial
  • Tool directory

Company

  • About
  • Privacy policy
  • Data methodology
  • Contact

© 2026 DataLearner AI. DataLearner curates industry data and case studies so researchers, enterprises, and developers can rely on trustworthy intelligence.

Privacy policyTerms of service

AI Model Leaderboards

Live rankings across ARC-AGI-2, HLE, AIME 2025, SWE-bench Verified, and more — browse composite scores or drill into math, coding, and agent categories.

View benchmark detailsUpdated on 2026-05-02 07:14:49

As of 2026-05, AA Intelligence Index leaders include GPT-5.5 (xhigh), GPT-5.5 (high), Opus 4.7 (max), based on 10 standardized capability benchmarks.

On the user-preference side, LMArena Text Generation currently ranks Opus 4.7 (thinking), Claude Opus 4.6 (thinking), Claude Opus 4.6 near the top via anonymous A/B voting.

Scroll down for per-benchmark breakdowns in math, coding, and agent categories. See Data Methodology for scoring details, or browse LLM Blogs for in-depth commentary.

Composite Rankings

There is no single, universally agreed-upon comprehensive AI model ranking, so we selected two representative leaderboards that approach the question from different angles. Artificial Analysis Intelligence Index aggregates scores from 10 standardized benchmarks (coding, math, reasoning, etc.) to measure objective capability. LMArena (formerly Chatbot Arena) ranks models by Elo ratings derived from anonymous crowd-sourced A/B voting, reflecting real-world user preference. Together they offer both an objective and a subjective perspective.

AA Intelligence Index

Full ranking

Composite of 10 standardized benchmarks across coding, math, science, reasoning, and agentic tasks.

Updated 2026-05-10

#ModelScore
1
OpenAI
GPT-5.5 (xhigh)OpenAI
60
2
OpenAI
GPT-5.5 (high)OpenAI
59
3
Anthropic
Opus 4.7 (max)Anthropic
57
4
Google Deep Mind
Gemini 3.1 Pro PreviewGoogle Deep Mind
57
5
OpenAI
GPT-5.5 (medium)OpenAI
57
6
Moonshot AI
Kimi K2.6Moonshot AI
54
7
X
MiMo-V2.5-ProXiaomi
54
8
OpenAI
GPT-5.3 Codex (xhigh)OpenAI
54
9
xAI
Grok 4.3xAI
53
10
F
Muse SparkFacebook AI研究实验室
52
Source: Artificial Analysis

LMArena Text Generation

Full ranking

Elo ratings from anonymous crowdsourced A/B voting, reflecting real user preference for response quality.

Updated 2026-05-07

#ModelElo
1
Anthropic
Opus 4.7 (thinking)Anthropic
1503
2
Anthropic
Claude Opus 4.6 (thinking)Anthropic
1502
3
Anthropic
Claude Opus 4.6Anthropic
1498
4
Google Deep Mind
Gemini 3.1 Pro PreviewGoogle Deep Mind
1492
5
Anthropic
Opus 4.7Anthropic
1491
6
F
Muse SparkFacebook AI研究实验室
1490
7
Google Deep Mind
Gemini 3.0 Pro (Preview 11-2025)Google Deep Mind
1486
8
OpenAI
gpt-5.5-highOpenAI
1484
9
xAI
grok-4.20-beta1xAI
1480
10
OpenAI
gpt-5.2-chat-latest-20260210OpenAI
1477
Source: LMArena

Per-Benchmark Rankings

Filter by math, coding, agent, and more. Switch benchmarks below or jump into a category leaderboard for the full ranking. View all benchmarks.

Benchmark Tracks
Overall
ARC-AGI-2HLEMMLU ProOpen Benchmark Directory
Math
AIME 2025FrontierMathMATH-500Open Math Leaderboard
Coding
SWE-bench VerifiedLiveCodeBenchSWE-Bench ProOpen Coding Leaderboard
Agent
τ²-BenchTerminal Bench 2.0Aider-PolyglotOpen Agent Leaderboard
Model Size:All3B and below7B13B34B65B100B and above
Model Type:AllReasoning ModelsFoundation ModelsInstruction/Chat ModelsCoding Models
Source:AllOpen SourceClosed Source
Origin:AllChina

LLM Performance Results

Data source: DataLearnerAI
Scores shown are the best result across all evaluation modes. Click a model name for the full breakdown.
RankModelLicense
OpenAI
GPT-5.5
OpenAI
52.2085.0035.40——Proprietary
Anthropic
Claude Mythos Preview
Anthropic
64.70——93.90—Proprietary
OpenAI
GPT-5.3 Codex
OpenAI
—————Proprietary
4
OpenAI
GPT-5.4
OpenAI
52.1077.1027.10——Proprietary
5
Anthropic
Opus 4.7
Anthropic
54.7075.8022.9087.60—Proprietary
6
Google Deep Mind
Gemini 3.1 Pro Preview
Google Deep Mind
51.4077.1016.7080.6090.80Proprietary
7
DeepSeek-AI
DeepSeek-V4-Pro
DeepSeek-AI
48.20——80.60—Free commercial
8
Moonshot AI
Kimi K2.6
Moonshot AI
54.00——80.20—Free commercial
9
Anthropic
Claude Opus 4.6
Anthropic
53.0066.3022.9080.8491.89Proprietary
10
阿里巴巴
Qwen3.6-Max-Preview
阿里巴巴
—————Proprietary
11
智谱AI
GLM 5.1
智谱AI
52.30————Free commercial
12
Cursor
Composer 2
Cursor
—————Proprietary
13
阿里巴巴
Qwen 3.6 Plus Preview
阿里巴巴
50.60——78.80—Proprietary
14
智谱AI
GLM-5
智谱AI
50.404.902.1077.8089.70Free commercial
15
OpenAI
GPT-5.4 mini
OpenAI
41.50—2.10——Proprietary
16
Anthropic
Opus 4.5
Anthropic
43.2037.604.2080.9081.99Proprietary
17
阿里巴巴
Qwen3.6-27B
阿里巴巴
24.00——77.20—Free commercial
18
Anthropic
Claude Sonnet 4.6
Anthropic
49.0058.308.3079.60—Proprietary
19
Facebook AI研究实验室
Muse Spark
Facebook AI研究实验室
58.0042.5014.6077.40—Proprietary
20
DeepSeek-AI
DeepSeek-V4-Flash
DeepSeek-AI
45.10——79.00—Free commercial
21
Google Deep Mind
Gemini 3.0 Pro (Preview 11-2025)
Google Deep Mind
45.8045.1018.8076.2085.40Proprietary
22
阿里巴巴
Qwen3.5-397B-A17B
阿里巴巴
48.30——76.4086.70Free commercial
23
MiniMaxAI
MiniMax M2.5
MiniMaxAI
19.404.90—80.20—Free commercial
24
阿里巴巴
Qwen3.6-35B-A3B
阿里巴巴
21.40——73.40—Free commercial
25
StepFunAI
Step 3.5 Flash
StepFunAI
———74.4088.20Free commercial
26
Moonshot AI
Kimi K2.5
Moonshot AI
50.2011.804.2076.80—Free commercial
27
MiniMaxAI
M2.1
MiniMaxAI
22.00——74.80—Free commercial
28
Cursor
Composer 1.5
Cursor
—————Proprietary
29
Google Deep Mind
Gemini 3.0 Flash
Google Deep Mind
43.5033.604.2068.7090.20Proprietary
30
OpenAI
GPT-5.1
OpenAI
42.7017.6012.5076.30—Proprietary
31
DeepSeek-AI
DeepSeek V3.2
DeepSeek-AI
25.104.002.1073.1080.30Free commercial
32
OpenAI
GPT-5.4 nano
OpenAI
37.70—6.30——Proprietary
33
Anthropic
Claude Sonnet 4.5
Anthropic
33.6013.604.2082.0084.70Proprietary
34
阿里巴巴
Qwen3.5-27B
阿里巴巴
48.50——72.4079.00Free commercial
35
智谱AI
GLM-4.7
智谱AI
42.80—2.1073.8087.40Free commercial
36
Cursor
Composer 1
Cursor
—————Proprietary
37
阿里巴巴
Qwen3-Coder-Next
阿里巴巴
———70.60—Free commercial
38
Google Deep Mind
Gemini 2.5-Pro
Google Deep Mind
21.604.902.1067.20—Proprietary
39
阿里巴巴
Qwen3-30B-A3B-2507
阿里巴巴
9.80——22.0049.00Free commercial
40
Anthropic
Haiku 4.5
Anthropic
9.704.502.1073.3033.00Proprietary
41
Anthropic
Claude Sonnet 4
Anthropic
9.605.90—80.2052.00Proprietary
42
OpenAI
OpenAI o1
OpenAI
9.10——48.90—Proprietary
43
DeepMind
Gemma 4 26B A4B
DeepMind
17.20———68.20Free commercial
44
DeepSeek-AI
DeepSeek V3.2-Exp
DeepSeek-AI
20.30——67.8066.70Free commercial
45
Google Deep Mind
Gemini 2.5 Flash
Google Deep Mind
11.00—4.2050.00—Proprietary
46
MiniMaxAI
MiniMax-M1-80k
MiniMaxAI
8.40——56.00—Free commercial
47
阿里巴巴
Qwen3-235B-A22B
阿里巴巴
7.60——34.4034.40Free commercial
48
MiniMaxAI
MiniMax-M1-40k
MiniMaxAI
7.20——55.60—Free commercial
49
Google Deep Mind
Gemini 2.5 Flash-Lite
Google Deep Mind
6.90——27.60—Proprietary
50
OpenAI
GPT-5
OpenAI
35.209.9012.5072.8080.00Proprietary
GPT-5.5
OpenAI
HLE52.20
ARC-AGI-285.00
FrontierMath - Tier 435.40
SWE-bench Verified—
τ²-Bench—
Proprietary
Claude Mythos Preview
Anthropic
HLE64.70
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified93.90
τ²-Bench—
Proprietary
GPT-5.3 Codex
OpenAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
4
GPT-5.4
OpenAI
HLE52.10
ARC-AGI-277.10
FrontierMath - Tier 427.10
SWE-bench Verified—
τ²-Bench—
Proprietary
5
Opus 4.7
Anthropic
HLE54.70
ARC-AGI-275.80
FrontierMath - Tier 422.90
SWE-bench Verified87.60
τ²-Bench—
Proprietary
6
Gemini 3.1 Pro Preview
Google Deep Mind
HLE51.40
ARC-AGI-277.10
FrontierMath - Tier 416.70
SWE-bench Verified80.60
τ²-Bench90.80
Proprietary
7
DeepSeek-V4-Pro
DeepSeek-AI
HLE48.20
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified80.60
τ²-Bench—
Free commercial
8
Kimi K2.6
Moonshot AI
HLE54.00
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified80.20
τ²-Bench—
Free commercial
9
Claude Opus 4.6
Anthropic
HLE53.00
ARC-AGI-266.30
FrontierMath - Tier 422.90
SWE-bench Verified80.84
τ²-Bench91.89
Proprietary
10
Qwen3.6-Max-Preview
阿里巴巴
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
11
GLM 5.1
智谱AI
HLE52.30
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Free commercial
12
Composer 2
Cursor
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
13
Qwen 3.6 Plus Preview
阿里巴巴
HLE50.60
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified78.80
τ²-Bench—
Proprietary
14
GLM-5
智谱AI
HLE50.40
ARC-AGI-24.90
FrontierMath - Tier 42.10
SWE-bench Verified77.80
τ²-Bench89.70
Free commercial
15
GPT-5.4 mini
OpenAI
HLE41.50
ARC-AGI-2—
FrontierMath - Tier 42.10
SWE-bench Verified—
τ²-Bench—
Proprietary
16
Opus 4.5
Anthropic
HLE43.20
ARC-AGI-237.60
FrontierMath - Tier 44.20
SWE-bench Verified80.90
τ²-Bench81.99
Proprietary
17
Qwen3.6-27B
阿里巴巴
HLE24.00
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified77.20
τ²-Bench—
Free commercial
18
Claude Sonnet 4.6
Anthropic
HLE49.00
ARC-AGI-258.30
FrontierMath - Tier 48.30
SWE-bench Verified79.60
τ²-Bench—
Proprietary
19
Muse Spark
Facebook AI研究实验室
HLE58.00
ARC-AGI-242.50
FrontierMath - Tier 414.60
SWE-bench Verified77.40
τ²-Bench—
Proprietary
20
DeepSeek-V4-Flash
DeepSeek-AI
HLE45.10
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified79.00
τ²-Bench—
Free commercial
21
Gemini 3.0 Pro (Preview 11-2025)
Google Deep Mind
HLE45.80
ARC-AGI-245.10
FrontierMath - Tier 418.80
SWE-bench Verified76.20
τ²-Bench85.40
Proprietary
22
Qwen3.5-397B-A17B
阿里巴巴
HLE48.30
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified76.40
τ²-Bench86.70
Free commercial
23
MiniMax M2.5
MiniMaxAI
HLE19.40
ARC-AGI-24.90
FrontierMath - Tier 4—
SWE-bench Verified80.20
τ²-Bench—
Free commercial
24
Qwen3.6-35B-A3B
阿里巴巴
HLE21.40
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified73.40
τ²-Bench—
Free commercial
25
Step 3.5 Flash
StepFunAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified74.40
τ²-Bench88.20
Free commercial
26
Kimi K2.5
Moonshot AI
HLE50.20
ARC-AGI-211.80
FrontierMath - Tier 44.20
SWE-bench Verified76.80
τ²-Bench—
Free commercial
27
M2.1
MiniMaxAI
HLE22.00
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified74.80
τ²-Bench—
Free commercial
28
Composer 1.5
Cursor
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
29
Gemini 3.0 Flash
Google Deep Mind
HLE43.50
ARC-AGI-233.60
FrontierMath - Tier 44.20
SWE-bench Verified68.70
τ²-Bench90.20
Proprietary
30
GPT-5.1
OpenAI
HLE42.70
ARC-AGI-217.60
FrontierMath - Tier 412.50
SWE-bench Verified76.30
τ²-Bench—
Proprietary
31
DeepSeek V3.2
DeepSeek-AI
HLE25.10
ARC-AGI-24.00
FrontierMath - Tier 42.10
SWE-bench Verified73.10
τ²-Bench80.30
Free commercial
32
GPT-5.4 nano
OpenAI
HLE37.70
ARC-AGI-2—
FrontierMath - Tier 46.30
SWE-bench Verified—
τ²-Bench—
Proprietary
33
Claude Sonnet 4.5
Anthropic
HLE33.60
ARC-AGI-213.60
FrontierMath - Tier 44.20
SWE-bench Verified82.00
τ²-Bench84.70
Proprietary
34
Qwen3.5-27B
阿里巴巴
HLE48.50
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified72.40
τ²-Bench79.00
Free commercial
35
GLM-4.7
智谱AI
HLE42.80
ARC-AGI-2—
FrontierMath - Tier 42.10
SWE-bench Verified73.80
τ²-Bench87.40
Free commercial
36
Composer 1
Cursor
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
37
Qwen3-Coder-Next
阿里巴巴
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified70.60
τ²-Bench—
Free commercial
38
Gemini 2.5-Pro
Google Deep Mind
HLE21.60
ARC-AGI-24.90
FrontierMath - Tier 42.10
SWE-bench Verified67.20
τ²-Bench—
Proprietary
39
Qwen3-30B-A3B-2507
阿里巴巴
HLE9.80
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified22.00
τ²-Bench49.00
Free commercial
40
Haiku 4.5
Anthropic
HLE9.70
ARC-AGI-24.50
FrontierMath - Tier 42.10
SWE-bench Verified73.30
τ²-Bench33.00
Proprietary
41
Claude Sonnet 4
Anthropic
HLE9.60
ARC-AGI-25.90
FrontierMath - Tier 4—
SWE-bench Verified80.20
τ²-Bench52.00
Proprietary
42
OpenAI o1
OpenAI
HLE9.10
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified48.90
τ²-Bench—
Proprietary
43
Gemma 4 26B A4B
DeepMind
HLE17.20
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench68.20
Free commercial
44
DeepSeek V3.2-Exp
DeepSeek-AI
HLE20.30
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified67.80
τ²-Bench66.70
Free commercial
45
Gemini 2.5 Flash
Google Deep Mind
HLE11.00
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified50.00
τ²-Bench—
Proprietary
46
MiniMax-M1-80k
MiniMaxAI
HLE8.40
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified56.00
τ²-Bench—
Free commercial
47
Qwen3-235B-A22B
阿里巴巴
HLE7.60
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified34.40
τ²-Bench34.40
Free commercial
48
MiniMax-M1-40k
MiniMaxAI
HLE7.20
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified55.60
τ²-Bench—
Free commercial
49
Gemini 2.5 Flash-Lite
Google Deep Mind
HLE6.90
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified27.60
τ²-Bench—
Proprietary
50
GPT-5
OpenAI
HLE35.20
ARC-AGI-29.90
FrontierMath - Tier 412.50
SWE-bench Verified72.80
τ²-Bench80.00
Proprietary
Sort by:
Showing 50 of 204 modelsView Terminal Bench 2.0 benchmark page

Leaderboard FAQ

01

Where does the leaderboard data come from?

Scores are aggregated from primary sources: official model cards, technical reports, papers, vendor blog posts, and reproducible third-party evaluations. Each row links back to the underlying model detail page where the source is cited.

02

Why do scores for the same model differ across benchmarks?

Each benchmark measures a different capability — reasoning (HLE, ARC-AGI-2), math (AIME, FrontierMath), coding (SWE-bench Verified), agent use (τ²-Bench), and so on. A model tuned for one capability may perform very differently on another, which is exactly why we surface per-benchmark scores rather than a single number.

03

How often is the leaderboard updated?

Data is revalidated every 5 minutes, and new models or evaluation results are added as soon as they are published. The "Updated on" indicator at the top of the page reflects the most recent data refresh.

04

How should I read the composite ranking?

The composite view aggregates a model's standing across multiple core benchmarks. It is a useful first filter, but for production decisions you should drill into the specific benchmark closest to your workload — for example, SWE-bench Verified for coding agents, or τ²-Bench for tool-use scenarios.

05

How do I compare an open-source model with a closed API model?

Use the license filter at the top to mix open and closed models in the same view, then look at the same benchmark column for both. Beyond raw scores, consider total cost of ownership: API pricing for closed models vs. self-hosting cost for open weights.