DataLearner logoDataLearnerAI
Latest AI Insights
Model Leaderboards
Benchmarks
Model Directory
Model Comparison
Resource Center
Tools
LanguageEnglish
DataLearner logoDataLearner AI

A knowledge platform focused on LLM benchmarking, datasets, and practical instruction with continuously updated capability maps.

Products

  • Leaderboards
  • Model comparison
  • Datasets

Resources

  • Tutorials
  • Editorial
  • Tool directory

Company

  • About
  • Privacy policy
  • Data methodology
  • Contact

© 2026 DataLearner AI. DataLearner curates industry data and case studies so researchers, enterprises, and developers can rely on trustworthy intelligence.

Privacy policyTerms of service

AI Model Leaderboards

Live rankings across ARC-AGI-2, HLE, AIME 2025, SWE-bench Verified, and more — browse composite scores or drill into math, coding, and agent categories.

View benchmark detailsUpdated on 2026-05-02 07:14:49

As of 2026-05, AA Intelligence Index leaders include GPT-5.5 (xhigh), GPT-5.5 (high), Opus 4.7 (max), based on 10 standardized capability benchmarks.

On the user-preference side, LMArena Text Generation currently ranks Opus 4.7 (thinking), Claude Opus 4.6 (thinking), Claude Opus 4.6 near the top via anonymous A/B voting.

Scroll down for per-benchmark breakdowns in math, coding, and agent categories. See Data Methodology for scoring details, or browse LLM Blogs for in-depth commentary.

Composite Rankings

There is no single, universally agreed-upon comprehensive AI model ranking, so we selected two representative leaderboards that approach the question from different angles. Artificial Analysis Intelligence Index aggregates scores from 10 standardized benchmarks (coding, math, reasoning, etc.) to measure objective capability. LMArena (formerly Chatbot Arena) ranks models by Elo ratings derived from anonymous crowd-sourced A/B voting, reflecting real-world user preference. Together they offer both an objective and a subjective perspective.

AA Intelligence Index

Full ranking

Composite of 10 standardized benchmarks across coding, math, science, reasoning, and agentic tasks.

Updated 2026-05-10

#ModelScore
1
OpenAI
GPT-5.5 (xhigh)OpenAI
60
2
OpenAI
GPT-5.5 (high)OpenAI
59
3
Anthropic
Opus 4.7 (max)Anthropic
57
4
Google Deep Mind
Gemini 3.1 Pro PreviewGoogle Deep Mind
57
5
OpenAI
GPT-5.5 (medium)OpenAI
57
6
Moonshot AI
Kimi K2.6Moonshot AI
54
7
X
MiMo-V2.5-ProXiaomi
54
8
OpenAI
GPT-5.3 Codex (xhigh)OpenAI
54
9
xAI
Grok 4.3xAI
53
10
F
Muse SparkFacebook AI研究实验室
52
Source: Artificial Analysis

LMArena Text Generation

Full ranking

Elo ratings from anonymous crowdsourced A/B voting, reflecting real user preference for response quality.

Updated 2026-05-07

#ModelElo
1
Anthropic
Opus 4.7 (thinking)Anthropic
1503
2
Anthropic
Claude Opus 4.6 (thinking)Anthropic
1502
3
Anthropic
Claude Opus 4.6Anthropic
1498
4
Google Deep Mind
Gemini 3.1 Pro PreviewGoogle Deep Mind
1492
5
Anthropic
Opus 4.7Anthropic
1491
6
F
Muse SparkFacebook AI研究实验室
1490
7
Google Deep Mind
Gemini 3.0 Pro (Preview 11-2025)Google Deep Mind
1486
8
OpenAI
gpt-5.5-highOpenAI
1484
9
xAI
grok-4.20-beta1xAI
1480
10
OpenAI
gpt-5.2-chat-latest-20260210OpenAI
1477
Source: LMArena

Per-Benchmark Rankings

Filter by math, coding, agent, and more. Switch benchmarks below or jump into a category leaderboard for the full ranking. View all benchmarks.

Benchmark Tracks
Overall
ARC-AGI-2HLEMMLU ProOpen Benchmark Directory
Math
AIME 2025FrontierMathMATH-500Open Math Leaderboard
Coding
SWE-bench VerifiedLiveCodeBenchSWE-Bench ProOpen Coding Leaderboard
Agent
τ²-BenchTerminal Bench 2.0Aider-PolyglotOpen Agent Leaderboard
Model Size:All3B and below7B13B34B65B100B and above
Model Type:AllReasoning ModelsFoundation ModelsInstruction/Chat ModelsCoding Models
Source:AllOpen SourceClosed Source
Origin:AllChina

LLM Performance Results

Data source: DataLearnerAI
Scores shown are the best result across all evaluation modes. Click a model name for the full breakdown.
RankModelLicense
Facebook AI研究实验室
Muse Spark
Facebook AI研究实验室
58.0042.5014.6077.40—Proprietary
OpenAI
GPT-5.5 Pro
OpenAI
57.2084.6039.60——Proprietary
Anthropic
Opus 4.7
Anthropic
54.7075.8022.9087.60—Proprietary
4
Anthropic
Claude Opus 4.6
Anthropic
53.0066.3022.9080.8491.89Proprietary
5
OpenAI
GPT-5.5
OpenAI
52.2085.0035.40——Proprietary
6
OpenAI
GPT-5.2 Pro
OpenAI
50.0054.2031.30——Proprietary
7
阿里巴巴
Qwen3-Max-Thinking
阿里巴巴
49.80——75.3082.10Proprietary
8
Google Deep Mind
Gemini 3 Deep Think - 2620
Google Deep Mind
48.4084.60———Proprietary
9
Anthropic
Opus 4.5
Anthropic
43.2037.604.2080.9081.99Proprietary
10
OpenAI
GPT-5.1
OpenAI
42.7017.6012.5076.30—Proprietary
11
OpenAI
GPT-5-Pro
OpenAI
42.0018.0014.60——Proprietary
12
OpenAI
GPT-5.4 mini
OpenAI
41.50—2.10——Proprietary
13
xAI
Grok 4
xAI
38.6015.902.1058.60—Proprietary
14
Google Deep Mind
Gemini 2.5-Pro
Google Deep Mind
21.604.902.1067.20—Proprietary
15
Google Deep Mind
Gemini-2.5-Pro-Preview-05-06
Google Deep Mind
21.60—2.1063.20—Proprietary
16
OpenAI
o3-pro
OpenAI
21.00——75.00—Proprietary
17
OpenAI
OpenAI o3
OpenAI
20.326.502.1069.10—Proprietary
18
Google Deep Mind
Gemini 2.5 Pro Experimental 03-25
Google Deep Mind
18.80—4.2063.80—Proprietary
19
OpenAI
OpenAI o4 - mini
OpenAI
17.70—6.3068.1056.90Proprietary
20
xAI
Grok 4.1 Fast
xAI
17.60———82.71Proprietary
21
OpenAI
OpenAI o3-mini
OpenAI
13.40—4.2040.80—Proprietary
22
Google Deep Mind
Gemini 2.5 Flash
Google Deep Mind
11.00—4.2050.00—Proprietary
23
Anthropic
Claude Opus 4
Anthropic
10.708.604.2072.5072.50Proprietary
24
Anthropic
Claude Sonnet 4
Anthropic
9.605.90—80.2052.00Proprietary
25
OpenAI
OpenAI o1
OpenAI
9.10——48.90—Proprietary
26
Google Deep Mind
Gemini 2.5 Flash-Lite
Google Deep Mind
6.90——27.60—Proprietary
27
OpenAI
GPT-5-Nano
OpenAI
——2.10——Proprietary
28
Microsoft Azure
Phi-4-instruct (reasoning-trained)
Microsoft Azure
—————Proprietary
29
Google Deep Mind
Gemini 2.5 Pro Deep Think
Google Deep Mind
——10.40——Proprietary
30
xAI
Grok-3 - Reasoning Beta
xAI
—————Proprietary
31
Moonshot AI
Kimi-k1.6-IOI-high
Moonshot AI
—————Proprietary
32
OpenAI
OpenAI o3-mini (medium)
OpenAI
—————Proprietary
33
Moonshot AI
Kimi-k1.6-IOI
Moonshot AI
—————Proprietary
34
Cursor
Composer 1.5
Cursor
—————Proprietary
35
MistralAI
Magistral-Medium-2506
MistralAI
—————Proprietary
36
xAI
Grok 3.5
xAI
—————Proprietary
37
OpenAI
GPT-5.1 Instant
OpenAI
—0.00———Proprietary
38
Moonshot AI
Kimi k1.5 (Short-CoT)
Moonshot AI
—————Proprietary
39
Anthropic
Claude Sonnet 3.7-64K Extended Thinking
Anthropic
—————Proprietary
40
普林斯顿大学
Kimi k1.5 (Long-CoT)
普林斯顿大学
—————Proprietary
41
OpenAI
OpenAI o3-mini (high)
OpenAI
——4.2049.30—Proprietary
42
xAI
Grok 4.1
xAI
———54.60—Proprietary
43
Anthropic
Opus 4.1
Anthropic
——4.2074.50—Proprietary
44
OpenAI
OpenAI o1-mini
OpenAI
—————Proprietary
45
腾讯AI实验室
Hunyuan-TurboS
腾讯AI实验室
—————Proprietary
46
腾讯AI实验室
Hunyuan-T1
腾讯AI实验室
—————Proprietary
Muse Spark
Facebook AI研究实验室
HLE58.00
ARC-AGI-242.50
FrontierMath - Tier 414.60
SWE-bench Verified77.40
τ²-Bench—
Proprietary
GPT-5.5 Pro
OpenAI
HLE57.20
ARC-AGI-284.60
FrontierMath - Tier 439.60
SWE-bench Verified—
τ²-Bench—
Proprietary
Opus 4.7
Anthropic
HLE54.70
ARC-AGI-275.80
FrontierMath - Tier 422.90
SWE-bench Verified87.60
τ²-Bench—
Proprietary
4
Claude Opus 4.6
Anthropic
HLE53.00
ARC-AGI-266.30
FrontierMath - Tier 422.90
SWE-bench Verified80.84
τ²-Bench91.89
Proprietary
5
GPT-5.5
OpenAI
HLE52.20
ARC-AGI-285.00
FrontierMath - Tier 435.40
SWE-bench Verified—
τ²-Bench—
Proprietary
6
GPT-5.2 Pro
OpenAI
HLE50.00
ARC-AGI-254.20
FrontierMath - Tier 431.30
SWE-bench Verified—
τ²-Bench—
Proprietary
7
Qwen3-Max-Thinking
阿里巴巴
HLE49.80
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified75.30
τ²-Bench82.10
Proprietary
8
Gemini 3 Deep Think - 2620
Google Deep Mind
HLE48.40
ARC-AGI-284.60
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
9
Opus 4.5
Anthropic
HLE43.20
ARC-AGI-237.60
FrontierMath - Tier 44.20
SWE-bench Verified80.90
τ²-Bench81.99
Proprietary
10
GPT-5.1
OpenAI
HLE42.70
ARC-AGI-217.60
FrontierMath - Tier 412.50
SWE-bench Verified76.30
τ²-Bench—
Proprietary
11
GPT-5-Pro
OpenAI
HLE42.00
ARC-AGI-218.00
FrontierMath - Tier 414.60
SWE-bench Verified—
τ²-Bench—
Proprietary
12
GPT-5.4 mini
OpenAI
HLE41.50
ARC-AGI-2—
FrontierMath - Tier 42.10
SWE-bench Verified—
τ²-Bench—
Proprietary
13
Grok 4
xAI
HLE38.60
ARC-AGI-215.90
FrontierMath - Tier 42.10
SWE-bench Verified58.60
τ²-Bench—
Proprietary
14
Gemini 2.5-Pro
Google Deep Mind
HLE21.60
ARC-AGI-24.90
FrontierMath - Tier 42.10
SWE-bench Verified67.20
τ²-Bench—
Proprietary
15
Gemini-2.5-Pro-Preview-05-06
Google Deep Mind
HLE21.60
ARC-AGI-2—
FrontierMath - Tier 42.10
SWE-bench Verified63.20
τ²-Bench—
Proprietary
16
o3-pro
OpenAI
HLE21.00
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified75.00
τ²-Bench—
Proprietary
17
OpenAI o3
OpenAI
HLE20.32
ARC-AGI-26.50
FrontierMath - Tier 42.10
SWE-bench Verified69.10
τ²-Bench—
Proprietary
18
Gemini 2.5 Pro Experimental 03-25
Google Deep Mind
HLE18.80
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified63.80
τ²-Bench—
Proprietary
19
OpenAI o4 - mini
OpenAI
HLE17.70
ARC-AGI-2—
FrontierMath - Tier 46.30
SWE-bench Verified68.10
τ²-Bench56.90
Proprietary
20
Grok 4.1 Fast
xAI
HLE17.60
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench82.71
Proprietary
21
OpenAI o3-mini
OpenAI
HLE13.40
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified40.80
τ²-Bench—
Proprietary
22
Gemini 2.5 Flash
Google Deep Mind
HLE11.00
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified50.00
τ²-Bench—
Proprietary
23
Claude Opus 4
Anthropic
HLE10.70
ARC-AGI-28.60
FrontierMath - Tier 44.20
SWE-bench Verified72.50
τ²-Bench72.50
Proprietary
24
Claude Sonnet 4
Anthropic
HLE9.60
ARC-AGI-25.90
FrontierMath - Tier 4—
SWE-bench Verified80.20
τ²-Bench52.00
Proprietary
25
OpenAI o1
OpenAI
HLE9.10
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified48.90
τ²-Bench—
Proprietary
26
Gemini 2.5 Flash-Lite
Google Deep Mind
HLE6.90
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified27.60
τ²-Bench—
Proprietary
27
GPT-5-Nano
OpenAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 42.10
SWE-bench Verified—
τ²-Bench—
Proprietary
28
Phi-4-instruct (reasoning-trained)
Microsoft Azure
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
29
Gemini 2.5 Pro Deep Think
Google Deep Mind
HLE—
ARC-AGI-2—
FrontierMath - Tier 410.40
SWE-bench Verified—
τ²-Bench—
Proprietary
30
Grok-3 - Reasoning Beta
xAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
31
Kimi-k1.6-IOI-high
Moonshot AI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
32
OpenAI o3-mini (medium)
OpenAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
33
Kimi-k1.6-IOI
Moonshot AI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
34
Composer 1.5
Cursor
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
35
Magistral-Medium-2506
MistralAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
36
Grok 3.5
xAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
37
GPT-5.1 Instant
OpenAI
HLE—
ARC-AGI-20.00
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
38
Kimi k1.5 (Short-CoT)
Moonshot AI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
39
Claude Sonnet 3.7-64K Extended Thinking
Anthropic
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
40
Kimi k1.5 (Long-CoT)
普林斯顿大学
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
41
OpenAI o3-mini (high)
OpenAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified49.30
τ²-Bench—
Proprietary
42
Grok 4.1
xAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified54.60
τ²-Bench—
Proprietary
43
Opus 4.1
Anthropic
HLE—
ARC-AGI-2—
FrontierMath - Tier 44.20
SWE-bench Verified74.50
τ²-Bench—
Proprietary
44
OpenAI o1-mini
OpenAI
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
45
Hunyuan-TurboS
腾讯AI实验室
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
46
Hunyuan-T1
腾讯AI实验室
HLE—
ARC-AGI-2—
FrontierMath - Tier 4—
SWE-bench Verified—
τ²-Bench—
Proprietary
Sort by:

Leaderboard FAQ

01

Where does the leaderboard data come from?

Scores are aggregated from primary sources: official model cards, technical reports, papers, vendor blog posts, and reproducible third-party evaluations. Each row links back to the underlying model detail page where the source is cited.

02

Why do scores for the same model differ across benchmarks?

Each benchmark measures a different capability — reasoning (HLE, ARC-AGI-2), math (AIME, FrontierMath), coding (SWE-bench Verified), agent use (τ²-Bench), and so on. A model tuned for one capability may perform very differently on another, which is exactly why we surface per-benchmark scores rather than a single number.

03

How often is the leaderboard updated?

Data is revalidated every 5 minutes, and new models or evaluation results are added as soon as they are published. The "Updated on" indicator at the top of the page reflects the most recent data refresh.

04

How should I read the composite ranking?

The composite view aggregates a model's standing across multiple core benchmarks. It is a useful first filter, but for production decisions you should drill into the specific benchmark closest to your workload — for example, SWE-bench Verified for coding agents, or τ²-Bench for tool-use scenarios.

05

How do I compare an open-source model with a closed API model?

Use the license filter at the top to mix open and closed models in the same view, then look at the same benchmark column for both. Beyond raw scores, consider total cost of ownership: API pricing for closed models vs. self-hosting cost for open weights.