加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
Data source: HuggingFace
| Model | Type | Parameters (B) | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | Architecture |
|---|---|---|---|---|---|---|---|---|---|---|
| Platypus2-70B-instruct | Fine Tuned Models | 689.8 | 69.3 | 71.84 | 87.94 | 70.48 | 62.26 | 82.72 | 40.56 | LlamaForCausalLM |
| Yi-34B-200K-rawrr1-LORA-DPO-experimental-r3 | Fine Tuned Models | 340 | 69.29 | 64.85 | 84.77 |
Data is for reference only. Official sources are authoritative. Click model names to view DataLearner model profiles.
| 76.0 |
| 45.35 |
| 83.11 |
| 61.64 |
| ? |
| BeagleLake-7B-Toxic | Fine Tuned Models | 70 | 69.24 | 65.19 | 83.83 | 62.82 | 57.67 | 82.32 | 63.61 | MistralForCausalLM |
| LaseredHermes-7B-v1 | Fine Tuned Models | 72.4 | 69.2 | 66.98 | 85.22 | 63.6 | 59.01 | 78.3 | 62.09 | MistralForCausalLM |
| Pallas-0.5-LASER-0.3 | Fine Tuned Models | 343.9 | 69.17 | 64.76 | 83.17 | 74.66 | 55.43 | 80.9 | 56.1 | LlamaForCausalLM |
| Yi-34b-200K-rawrr-v2-run-0902-LoRA | Fine Tuned Models | 340 | 69.15 | 64.68 | 84.5 | 75.76 | 46.66 | 81.14 | 62.17 | ? |
| airoboros-l2-70b-2.2.1 | Chat Models | 700 | 69.13 | 69.71 | 87.95 | 69.79 | 59.49 | 82.95 | 44.88 | LlamaForCausalLM |
| LaseredHermes-7B-v1 | Fine Tuned Models | 72.4 | 69.09 | 66.89 | 85.21 | 63.58 | 59.09 | 78.45 | 61.33 | MistralForCausalLM |
| openchat-nectar-0.14 | Fine Tuned Models | 72.4 | 69.09 | 65.61 | 83.02 | 64.58 | 50.09 | 82.0 | 69.22 | MistralForCausalLM |
| Mixtral-8x7B-v0.1-top3 | Merged Models or MoE Models | 467 | 69.09 | 67.41 | 86.63 | 71.98 | 48.58 | 82.4 | 57.54 | MixtralForCausalLM |
| Optimus-7B | Fine Tuned Models | 72.4 | 69.09 | 65.44 | 85.41 | 63.61 | 55.79 | 78.77 | 65.5 | MistralForCausalLM |
| loyal-piano-m7-cdpo | Chat Models | 72.4 | 69.08 | 67.15 | 85.39 | 64.52 | 61.53 | 79.4 | 56.48 | MistralForCausalLM |
| Mistral-CatMacaroni-slerp-gradient | Fine Tuned Models | 72.4 | 69.08 | 65.53 | 85.66 | 61.53 | 64.1 | 80.03 | 57.62 | Unknown |
| Neural-una-cybertron-7b | Fine Tuned Models | 72.4 | 69.05 | 69.03 | 84.51 | 62.79 | 64.99 | 80.66 | 52.31 | Unknown |
| orca_mini_v3_70b | Fine Tuned Models | 687.2 | 69.02 | 71.25 | 87.85 | 70.18 | 61.27 | 82.72 | 40.86 | Unknown |
| loyal-piano-m7-cdpo | Fine Tuned Models | 72.4 | 69.0 | 67.06 | 85.42 | 64.54 | 61.54 | 79.08 | 56.33 | MistralForCausalLM |
| servile-harpsichord-cdpo | Chat Models | 72.4 | 68.98 | 67.32 | 85.18 | 64.54 | 60.61 | 79.16 | 57.09 | MistralForCausalLM |
| LeoScorpius-GreenNode-Platypus-7B-v1 | Fine Tuned Models | 70 | 68.96 | 66.04 | 86.53 | 62.06 | 52.78 | 82.16 | 64.22 | MistralForCausalLM |
| LHK_44 | Fine Tuned Models | 107.3 | 68.95 | 66.55 | 84.86 | 65.37 | 59.58 | 80.9 | 56.41 | LlamaForCausalLM |
| MegaDolphin-120b | Fine Tuned Models | 1203.2 | 68.91 | 69.03 | 87.8 | 69.26 | 59.28 | 81.85 | 46.25 | LlamaForCausalLM |
| openchat-3.5-1210 | Fine Tuned Models | 72.4 | 68.89 | 64.93 | 84.92 | 64.62 | 52.15 | 80.74 | 65.96 | MistralForCausalLM |
| Mixtral-8x7B-peft-v0.1 | Chat Models | 70 | 68.87 | 67.24 | 86.03 | 68.59 | 59.54 | 80.43 | 51.4 | Unknown |
| FT | Fine Tuned Models | 343.9 | 68.85 | 63.05 | 82.78 | 69.69 | 59.88 | 79.64 | 58.07 | Unknown |
| kellemar-DPO-7B-d | Fine Tuned Models | 72.4 | 68.84 | 66.89 | 85.16 | 62.77 | 56.88 | 79.32 | 62.02 | MistralForCausalLM |
| mixtral_7bx4_moe | Fine Tuned Models | 241.5 | 68.83 | 65.27 | 85.28 | 62.84 | 59.85 | 77.66 | 62.09 | Unknown |
| Marcoroni-70B-v1 | Fine Tuned Models | 687.2 | 68.83 | 73.55 | 87.62 | 70.67 | 64.41 | 83.43 | 33.28 | Unknown |
| FT | Fine Tuned Models | 343.9 | 68.81 | 63.14 | 82.78 | 69.5 | 59.8 | 79.4 | 58.23 | Unknown |
| Crunchy-onion | Fine Tuned Models | 467 | 68.75 | 67.15 | 86.19 | 70.02 | 63.88 | 73.24 | 52.01 | MixtralForCausalLM |
| LHK | Chat Models | 107.3 | 68.74 | 66.38 | 84.49 | 65.13 | 59.12 | 80.98 | 56.33 | LlamaForCausalLM |
| A11P | Fine Tuned Models | 0 | 68.73 | 62.54 | 82.53 | 70.56 | 56.44 | 79.87 | 60.42 | Unknown |
| pic_7B_mistral_Full_v0.2 | Fine Tuned Models | 70 | 68.72 | 65.36 | 84.03 | 64.51 | 59.2 | 79.48 | 59.74 | MistralForCausalLM |
| SOLAR-10.7B-dpo-instruct-tuned-v0.1 | Chat Models | 107.3 | 68.68 | 65.19 | 86.09 | 66.25 | 51.81 | 83.98 | 58.76 | LlamaForCausalLM |
| Yi-34B-AEZAKMI-v1 | Chat Models | 343.9 | 68.67 | 64.33 | 84.31 | 73.91 | 55.73 | 80.82 | 52.92 | LlamaForCausalLM |
| loyal-piano-m7 | Chat Models | 72.4 | 68.67 | 66.72 | 85.03 | 64.43 | 60.03 | 79.08 | 56.71 | MistralForCausalLM |
| A12P | Fine Tuned Models | 0 | 68.64 | 64.42 | 82.32 | 69.97 | 62.22 | 79.64 | 53.3 | Unknown |
| agiin-13.6B-v0.0 | Fine Tuned Models | 137.8 | 68.63 | 69.45 | 86.59 | 61.94 | 67.4 | 78.69 | 47.69 | LlamaForCausalLM |
| spicyboros-70b-2.2 | Fine Tuned Models | 700 | 68.62 | 70.73 | 87.58 | 70.32 | 58.31 | 83.82 | 40.94 | LlamaForCausalLM |
| Maya_Hermes-2.5-Mistral-7B | Fine Tuned Models | 72.4 | 68.6 | 66.3 | 85.07 | 63.23 | 55.89 | 78.85 | 62.24 | MistralForCausalLM |
| MixtralRPChat-ZLoss | Chat Models | 467 | 68.59 | 68.6 | 86.1 | 70.44 | 53.85 | 82.0 | 50.57 | MixtralForCausalLM |
| model_007 | Fine Tuned Models | 687.2 | 68.56 | 71.08 | 87.65 | 69.04 | 63.12 | 83.35 | 37.15 | Unknown |
| SpellBlade | Fine Tuned Models | 689.8 | 68.54 | 69.28 | 87.31 | 70.5 | 47.1 | 83.19 | 53.83 | LlamaForCausalLM |
| model_009 | Fine Tuned Models | 687.2 | 68.53 | 71.59 | 87.7 | 69.43 | 60.72 | 82.32 | 39.42 | Unknown |
| stealth-v1.3 | Fine Tuned Models | 72.4 | 68.53 | 65.19 | 84.44 | 62.7 | 59.12 | 78.61 | 61.11 | MistralForCausalLM |
| Chupacabra-7B-v2.04 | Fine Tuned Models | 72.4 | 68.52 | 66.3 | 85.7 | 60.94 | 67.76 | 78.93 | 51.48 | MistralForCausalLM |
| Mixtral-8x7B-v0.1 | Fine Tuned Models | 467 | 68.47 | 66.38 | 86.46 | 71.88 | 46.81 | 81.69 | 57.62 | MixtralForCausalLM |
| model_101 | Fine Tuned Models | 687.2 | 68.46 | 68.69 | 86.42 | 69.92 | 58.85 | 82.08 | 44.81 | Unknown |
| ds_diasum_md_mixtral | Chat Models | 0 | 68.42 | 66.3 | 85.45 | 69.51 | 55.72 | 80.35 | 53.22 | Unknown |
| Mixtral-8x7B-v0.1 | Pretrained Models | 467 | 68.42 | 66.04 | 86.49 | 71.82 | 46.78 | 81.93 | 57.47 | MixtralForCausalLM |
| NeuralHermes-2.5-Mistral-7B-distilabel | Fine Tuned Models | 72.4 | 68.4 | 65.78 | 84.97 | 63.63 | 55.86 | 78.69 | 61.49 | MistralForCausalLM |
| agiin-13.6B-v0.1 | Chat Models | 137.8 | 68.4 | 69.45 | 86.64 | 61.15 | 67.97 | 78.69 | 46.47 | MistralForCausalLM |
| xDAN-L1-Chat-RL-v1 | Fine Tuned Models | 72.4 | 68.38 | 66.3 | 85.81 | 63.21 | 56.7 | 78.85 | 59.44 | MistralForCausalLM |
| PlatYi-34B-Llama | Chat Models | 343.9 | 68.37 | 67.83 | 85.35 | 78.26 | 53.46 | 82.87 | 42.46 | Unknown |
| kellemar-DPO-7B-v1.01 | Fine Tuned Models | 72.4 | 68.32 | 65.78 | 85.04 | 63.24 | 55.54 | 78.69 | 61.64 | MistralForCausalLM |
| PlatYi-34B-Llama-Q-FastChat | Chat Models | 343.9 | 68.31 | 66.13 | 85.25 | 78.37 | 53.62 | 82.16 | 44.35 | Unknown |
| neural-chat-7b-v3-2 | Fine Tuned Models | 70 | 68.29 | 67.49 | 83.92 | 63.55 | 59.68 | 79.95 | 55.12 | MistralForCausalLM |
| Pallas-0.5-LASER-0.4 | Fine Tuned Models | 343.9 | 68.28 | 63.31 | 82.74 | 74.32 | 55.25 | 80.58 | 53.45 | LlamaForCausalLM |
| WordWoven-13B | Merged Models or MoE Models | 128.8 | 68.25 | 66.13 | 85.81 | 64.06 | 54.45 | 78.93 | 60.12 | MixtralForCausalLM |
| llama2_70b_mmlu | Fine Tuned Models | 689.8 | 68.24 | 65.61 | 87.37 | 71.89 | 49.15 | 82.4 | 52.99 | LlamaForCausalLM |
| NeuralHermes-2.5-Mistral-7B | Fine Tuned Models | 72.4 | 68.22 | 66.55 | 84.9 | 63.32 | 54.93 | 78.3 | 61.33 | MistralForCausalLM |
| OrionStar-Yi-34B-Chat-Llama | Fine Tuned Models | 343.9 | 68.17 | 64.93 | 84.34 | 73.67 | 53.35 | 78.85 | 53.9 | LlamaForCausalLM |
| Sensualize-Solar-10.7B | Fine Tuned Models | 107.3 | 68.17 | 65.02 | 84.55 | 65.27 | 53.63 | 83.98 | 56.56 | LlamaForCausalLM |
| blossom-v3_1-yi-34b | Fine Tuned Models | 340 | 68.16 | 65.36 | 84.24 | 74.37 | 56.06 | 82.08 | 46.85 | LlamaForCausalLM |
| AZG | Fine Tuned Models | 0 | 68.16 | 62.88 | 82.02 | 70.29 | 53.84 | 79.95 | 59.97 | Unknown |
| CapybaraHermes-2.5-Mistral-7B | Chat Models | 72.4 | 68.14 | 65.78 | 85.45 | 63.13 | 56.91 | 78.3 | 59.29 | MistralForCausalLM |
| agiin-11.1B-v0.0 | Fine Tuned Models | 111.7 | 68.1 | 67.32 | 86.35 | 64.99 | 67.67 | 78.85 | 43.44 | LlamaForCausalLM |
| PlatYi-34B-LoRA | Chat Models | 343.9 | 68.1 | 67.15 | 85.37 | 78.46 | 53.32 | 83.66 | 40.64 | LlamaForCausalLM |
| Merged-DPO-7B | Chat Models | 70 | 68.06 | 68.94 | 87.75 | 55.35 | 72.76 | 78.37 | 45.19 | Unknown |
| lil-c3po | Chat Models | 72.4 | 68.03 | 65.02 | 84.45 | 62.36 | 68.73 | 79.16 | 48.45 | Unknown |
| bagel-dpo-7b-v0.1 | Fine Tuned Models | 72.4 | 67.95 | 66.72 | 84.16 | 64.24 | 64.05 | 80.9 | 47.61 | MistralForCausalLM |
| Pallas-0.5-LASER-exp2-0.1 | Fine Tuned Models | 343.9 | 67.92 | 62.97 | 82.11 | 74.66 | 55.24 | 79.79 | 52.77 | LlamaForCausalLM |
| ThetaWave-7B-sft | Fine Tuned Models | 72.4 | 67.92 | 63.14 | 84.42 | 63.78 | 59.74 | 79.64 | 56.79 | MistralForCausalLM |
| PlatYi-34B-Llama-Q-v2 | Chat Models | 343.9 | 67.88 | 61.09 | 85.09 | 76.59 | 52.65 | 82.79 | 49.05 | LlamaForCausalLM |
| Einstein-openchat-7B | Fine Tuned Models | 72.4 | 67.87 | 65.1 | 83.57 | 64.01 | 54.51 | 79.16 | 60.88 | MistralForCausalLM |
| OpenAGI-7B-v0.1 | Chat Models | 72.4 | 67.87 | 68.26 | 85.06 | 61.6 | 59.4 | 79.79 | 53.07 | MistralForCausalLM |
| PlatYi-34B-200k-Q-FastChat | Chat Models | 340 | 67.85 | 64.93 | 84.46 | 77.13 | 48.38 | 80.74 | 51.48 | LlamaForCausalLM |
| falcon-180B | Pretrained Models | 1795.2 | 67.85 | 69.45 | 88.86 | 70.5 | 45.47 | 86.9 | 45.94 | FalconForCausalLM |
| OpenHermes-2.5-neural-chat-7b-v3-1-7B | Fine Tuned Models | 72.4 | 67.84 | 66.55 | 84.47 | 63.34 | 61.22 | 78.37 | 53.07 | MistralForCausalLM |
| Mixtral-Orca-v0.1 | Chat Models | 467 | 67.82 | 69.71 | 88.88 | 66.06 | 63.85 | 81.14 | 37.3 | MixtralForCausalLM |
| SauerkrautLM-Mixtral-8x7B | Fine Tuned Models | 467 | 67.8 | 68.86 | 86.01 | 66.69 | 57.2 | 80.51 | 47.54 | MixtralForCausalLM |
| stealth-rag-v1.1 | Fine Tuned Models | 72.4 | 67.79 | 62.12 | 83.83 | 64.06 | 49.64 | 79.32 | 67.78 | MistralForCausalLM |
| Xwin-Math-70B-V1.0 | Fine Tuned Models | 700 | 67.78 | 64.51 | 84.88 | 66.2 | 51.58 | 81.53 | 58.0 | LlamaForCausalLM |
| DistilHermes-2.5-Mistral-7B | Fine Tuned Models | 72.4 | 67.76 | 65.87 | 84.78 | 63.65 | 54.24 | 78.22 | 59.82 | MistralForCausalLM |
| Chupacabra-7B | Fine Tuned Models | 72.4 | 67.76 | 66.81 | 83.52 | 62.68 | 52.31 | 79.08 | 62.17 | MistralForCausalLM |
| Bumblebee-7B | Fine Tuned Models | 72.4 | 67.73 | 63.4 | 84.16 | 64.0 | 50.96 | 78.22 | 65.66 | MistralForCausalLM |
| Voldemort-10B-DPO | Fine Tuned Models | 107.3 | 67.69 | 65.7 | 84.79 | 62.82 | 61.33 | 77.27 | 54.21 | MistralForCausalLM |
| Voldemort-10B-DPO | Fine Tuned Models | 107.3 | 67.68 | 66.04 | 84.84 | 62.88 | 61.44 | 77.03 | 53.83 | MistralForCausalLM |
| DPOpenHermes-7B | Fine Tuned Models | 72.4 | 67.63 | 65.96 | 85.9 | 63.98 | 56.92 | 78.22 | 54.81 | MistralForCausalLM |
| CCK-v2.0-DPO | Fine Tuned Models | 108.6 | 67.62 | 65.87 | 86.81 | 62.1 | 69.33 | 82.16 | 39.42 | LlamaForCausalLM |
| ORCA_LLaMA_70B_QLoRA | Fine Tuned Models | 700 | 67.6 | 72.27 | 87.74 | 70.23 | 63.37 | 83.66 | 28.35 | LlamaForCausalLM |
| DPOpenHermes-7B | Chat Models | 72.4 | 67.58 | 65.7 | 85.96 | 63.89 | 56.95 | 78.61 | 54.36 | MistralForCausalLM |
| SeaLLM-7B-v2 | Chat Models | 70 | 67.57 | 62.03 | 82.32 | 61.89 | 51.11 | 79.08 | 68.99 | MistralForCausalLM |
| MoMo-70B-LoRA-V1.1 | Chat Models | 700 | 67.53 | 66.64 | 87.16 | 66.76 | 54.98 | 83.35 | 46.32 | Unknown |
| BigWeave-v6-90b | Merged Models or MoE Models | 878 | 67.47 | 65.36 | 87.21 | 68.04 | 57.96 | 81.69 | 44.58 | LlamaForCausalLM |
| FashionGPT-70B-V1 | Fine Tuned Models | 700 | 67.47 | 71.08 | 87.32 | 70.7 | 63.92 | 83.66 | 28.13 | LlamaForCausalLM |
| juanako-7b-UNA | Fine Tuned Models | 72.4 | 67.46 | 68.17 | 85.34 | 62.47 | 65.13 | 78.85 | 44.81 | MistralForCausalLM |
| UNA-dolphin-2.6-mistral-7b-dpo-laser | Fine Tuned Models | 72.4 | 67.43 | 67.15 | 86.31 | 63.36 | 64.15 | 79.24 | 44.35 | MistralForCausalLM |
| Samantha-1.1-70b | Chat Models | 687.2 | 67.43 | 68.77 | 87.46 | 68.6 | 64.85 | 83.27 | 31.61 | Unknown |
| Moe-2x7b-QA-Code | Fine Tuned Models | 128.8 | 67.42 | 65.19 | 85.36 | 61.71 | 65.23 | 77.35 | 49.66 | MixtralForCausalLM |
| CodeNinja-1.0-OpenChat-7B | Fine Tuned Models | 72.4 | 67.4 | 63.48 | 83.65 | 63.77 | 47.16 | 79.79 | 66.57 | MistralForCausalLM |
| test_42_70b | Fine Tuned Models | 687.2 | 67.38 | 68.26 | 87.65 | 70.0 | 48.76 | 83.66 | 45.94 | Unknown |