- Gemini 3.1 Pro Previewleads in:General Knowledge (3/5), AI Agent - Information Search (1/1), AI Agent - Tool Usage (1/1), Multimodal Understanding (1/1)
- Claude Opus 4.6leads in:Math and Reasoning (2/2), Claw-style Agent Evaluation (1/1)
- Tied in:Agent Level Benchmark, Coding and Software Engineer
On average across the 15 shared benchmarks, Gemini 3.1 Pro Preview scores 1.71 higher.
Largest single-benchmark gap: LiveCodeBench — Gemini 3.1 Pro Preview 91.70 vs Claude Opus 4.6 76 (+15.70).
Page generated from structured model, pricing and benchmark records. No real-time LLM is used to write the prose.