Opus 4.7vsClaude Opus 4.6
Across 12 shared benchmarks, Opus 4.7 leads overall: Opus 4.7 wins 9, Claude Opus 4.6 wins 1, with 2 ties and an average score difference of +2.54.
Opus 4.7
Anthropic · 2026-04-16 · Reasoning model
Across 12 shared benchmarks, Opus 4.7 leads overall: Opus 4.7 wins 9, Claude Opus 4.6 wins 1, with 2 ties and an average score difference of +2.54.
Anthropic · 2026-04-16 · Reasoning model
Anthropic · 2026-02-05 · Reasoning model
Grouped by capability, sorted by largest gap within each. 12 shared benchmarks.
| Benchmark | Opus 4.7 | Claude Opus 4.6 | Diff |
|---|---|---|---|
| ARC-AGI-2 | 75.809 / 58最高(无工具) | 66.3014 / 58Extended (no tools) | +9.50 |
| GPQA Diamond | 94.204 / 175Extended (no tools) | 91.3112 / 175Extended (no tools) | +2.89 |
| HLE | 54.706 / 149Extended (with tools) | 538 / 149Extended (with tools, internet) | +1.70 |
| ARC-AGI | 93.509 / 65Thinking High (No Tools) | 9211 / 65Extended (no tools) | +1.50 |
| MMLU | 91.506 / 65Normal (No Tools) | 91.057 / 65Extended (no tools) | +0.45 |
| ARC-AGI-3 | 05 / 6Thinking High (No Tools) | 01 / 6最高(无工具) | — |
| Benchmark | Opus 4.7 | Claude Opus 4.6 | Diff |
|---|---|---|---|
| OSWorld-Verified | 783 / 14Extended (with tools) | 72.706 / 14Extended (with tools) | +5.30 |
| Terminal Bench 2.0 | 69.405 / 43Extended (with tools) | 65.409 / 43Extended (with tools) | +4 |
| Benchmark | Opus 4.7 | Claude Opus 4.6 | Diff |
|---|---|---|---|
| FrontierMath | 43.806 / 60极高强度思考(无工具) | 40.707 / 60最高(无工具) | +3.10 |
| FrontierMath - Tier 4 | 22.9012 / 80极高强度思考(无工具) | 22.9012 / 80最高(无工具) | — |
| Benchmark | Opus 4.7 | Claude Opus 4.6 | Diff |
|---|---|---|---|
| BrowseComp | 79.3011 / 43Extended (with tools) | 846 / 43Thinking (With Tools + Internet) | -4.70 |
| Benchmark | Opus 4.7 | Claude Opus 4.6 | Diff |
|---|---|---|---|
| SWE-bench Verified | 87.602 / 103Extended (with tools) | 80.846 / 103Extended (with tools) | +6.76 |
| Field | Opus 4.7 | Claude Opus 4.6 |
|---|---|---|
| Publisher | Anthropic | Anthropic |
| Release date | 2026-04-16 | 2026-02-05 |
| Model type | Reasoning model | Reasoning model |
| Architecture | Dense | Dense |
| Parameters | 0.0 | 0.0 |
| Context length | 1000K | 1000K |
| Max output | 131072 | 65536 |
Prices use DataLearner records when available; missing fields are not inferred.
| Item | Opus 4.7 | Claude Opus 4.6 |
|---|---|---|
| Text input | $5 / 1M tokens | $0.5 / 1M tokens |
| Text output | $25 / 1M tokens | $25 / 1M tokens |
| Cache read | $0.5 / 1M tokens | $0.5 / 1M tokens |
| Cache write | $6.25 / 1M tokens | $10 / 1M tokens |
On average across the 12 shared benchmarks, Opus 4.7 scores 2.54 higher.
Largest single-benchmark gap: ARC-AGI-2 — Opus 4.7 75.80 vs Claude Opus 4.6 66.30 (+9.50).
Page generated from structured model, pricing and benchmark records. No real-time LLM is used to write the prose.