加载中...
加载中...
Quickly view LLM performance across benchmarks like MMLU Pro, HLE, SWE-Bench, and more. Compare models across general knowledge, coding, and reasoning capabilities. Customize your comparison by selecting specific models and benchmarks.
Detailed benchmark descriptions available at:LLM Benchmark List & Guide
Benchmark switcher
Pick the leaderboard to sync both chart and table
Data source: DataLearnerAI
| 0.00 |
| 0.00 |
| 33.30 |
| 3 | Hunyuan-A13B-Instruct | 67.23 | 71.20 | 0.00 | 0.00 | 87.30 | 63.90 |
| 4 | Llama3.1-70B-Instruct | 66.40 | 48.00 | 0.00 | 0.00 | 0.00 | 33.30 |
| 5 | Qwen3-Next | 66.05 | 0.00 | 0.00 | 0.00 | 0.00 | 56.60 |
| 6 | Qwen2.5-72B | 58.10 | 45.90 | 0.00 | 0.00 | 0.00 | 0.00 |
| 7 | Llama3-70B-Instruct | 56.20 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
| 8 | Llama3-70B | 52.78 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
| 9 | Llama3.1-70B | 52.47 | 0.00 | 0.00 | 0.00 | 0.00 | 0.00 |
| 10 | DeepSeek-R1-Distill-Llama-70B | 0.00 | 65.20 | 0.00 | 94.50 | 0.00 | 0.00 |