加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
数据来源: HuggingFace
模型名称 | 模型类型 | 参数大小(亿) | 平均分 | ARC分数 | HellaSwag分数 | MMLU分数 | TruthfulQA分数 | Winogrande分数 | GSM8K分数 | 模型架构 |
|---|---|---|---|---|---|---|---|---|---|---|
| Smaug-72B-v0.1 | Fine Tuned Models | 722.9 | 80.48 | 76.02 | 89.27 | 77.15 | 76.67 | 85.08 | 78.7 | LlamaForCausalLM |
| alpaca-dragon-72b-v1 | Fine Tuned Models | 722.9 | 79.3 | 73.89 | 88.16 | 77.4 | 72.69 | 86.03 | 77.63 | LlamaForCausalLM |
| MoMo-72B-lora-1.8.7-DPO | Chat Models | 722.9 | 78.55 | 70.82 | 85.96 | 77.13 | 74.71 | 84.06 | 78.62 | LlamaForCausalLM |
| TomGrc_FusionNet_34Bx2_MoE_v0.1_DPO_f16 | Fine Tuned Models | 608.1 | 77.91 | 74.06 | 86.74 | 76.65 | 72.24 | 83.35 | 74.45 | MixtralForCausalLM |
| TomGrc_FusionNet_34Bx2_MoE_v0.1_full_linear_DPO | Fine Tuned Models | 608.1 | 77.52 | 74.06 | 86.67 | 76.69 | 71.32 | 83.43 | 72.93 | MixtralForCausalLM |
| Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B | Chat Models | 128.8 | 77.44 | 74.91 | 89.3 | 64.67 | 78.02 | 88.24 | 69.52 | MixtralForCausalLM |
| CCK_Asura_v1 | Fine Tuned Models | 689.8 | 77.43 | 73.89 | 89.07 | 75.44 | 71.75 | 86.35 | 68.08 | LlamaForCausalLM |
| LHK_DPO_v1 | Fine Tuned Models | 128.8 | 77.43 | 74.74 | 89.37 | 64.87 | 79.88 | 88.16 | 67.55 | MixtralForCausalLM |
| UNA-SimpleSmaug-34b-v1beta | Fine Tuned Models | 343.9 | 77.41 | 74.57 | 86.74 | 76.68 | 70.17 | 83.82 | 72.48 | LlamaForCausalLM |
| FusionNet_34Bx2_MoE_v0.1 | Fine Tuned Models | 608.1 | 77.38 | 73.72 | 86.46 | 76.72 | 71.01 | 83.35 | 73.01 | MixtralForCausalLM |
| Tess-72B-v1.5b | Fine Tuned Models | 722.9 | 77.3 | 71.25 | 85.53 | 76.63 | 71.99 | 81.45 | 76.95 | LlamaForCausalLM |
| MoMo-72B-lora-1.8.6-DPO | Chat Models | 722.9 | 77.29 | 70.14 | 86.03 | 77.4 | 69.0 | 84.37 | 76.8 | LlamaForCausalLM |
| Smaug-34B-v0.1 | Fine Tuned Models | 343.9 | 77.29 | 74.23 | 86.76 | 76.66 | 70.22 | 83.66 | 72.18 | LlamaForCausalLM |
| Truthful_DPO_TomGrc_FusionNet_34Bx2_MoE | Fine Tuned Models | 608.1 | 77.28 | 72.87 | 86.52 | 76.96 | 73.28 | 83.19 | 70.89 | MixtralForCausalLM |
| DARE_TIES_13B | Fine Tuned Models | 128.8 | 77.1 | 74.32 | 89.5 | 64.47 | 78.66 | 88.08 | 67.55 | MixtralForCausalLM |
| 13B_MATH_DPO | Fine Tuned Models | 128.8 | 77.08 | 74.66 | 89.51 | 64.53 | 78.63 | 88.08 | 67.1 | MixtralForCausalLM |
| FusionNet_34Bx2_MoE | Fine Tuned Models | 608.1 | 77.07 | 72.95 | 86.22 | 77.05 | 71.31 | 83.98 | 70.89 | MixtralForCausalLM |
| MoE_13B_DPO | Fine Tuned Models | 128.8 | 77.05 | 74.32 | 89.39 | 64.48 | 78.47 | 88.0 | 67.63 | MixtralForCausalLM |
| 4bit_quant_TomGrc_FusionNet_34Bx2_MoE_v0.1_DPO | Fine Tuned Models | 318 | 76.95 | 73.21 | 86.11 | 75.44 | 72.78 | 82.95 | 71.19 | MixtralForCausalLM |
| QuartetAnemoi-70B-t0.0001 | Merged Models or MoE Models | 689.8 | 76.86 | 73.38 | 88.9 | 75.42 | 69.53 | 85.32 | 68.61 | LlamaForCausalLM |
| MixTAO-7Bx2-MoE-Instruct-v7.0 | Chat Models | 128.8 | 76.55 | 74.23 | 89.37 | 64.54 | 74.26 | 87.77 | 69.14 | MixtralForCausalLM |
| Truthful_DPO_cloudyu_Mixtral_34Bx2_MoE_60B | Fine Tuned Models | 608.1 | 76.48 | 71.25 | 85.24 | 77.28 | 66.74 | 84.29 | 74.07 | MixtralForCausalLM |
| CCK_Asura_v2.1 | Fine Tuned Models | 689.8 | 76.41 | 72.53 | 88.75 | 74.96 | 67.33 | 85.87 | 68.99 | LlamaForCausalLM |
| MoMo-72B-lora-1.8.4-DPO | Chat Models | 722.9 | 76.23 | 69.62 | 85.35 | 77.33 | 64.64 | 84.14 | 76.27 | LlamaForCausalLM |
| FusionNet_7Bx2_MoE_v0.1 | Fine Tuned Models | 128.8 | 76.16 | 74.06 | 88.9 | 65.0 | 71.2 | 87.53 | 70.28 | MixtralForCausalLM |
| NeuralTrix-7B-dpo | Fine Tuned Models | 72.4 | 76.15 | 72.27 | 88.91 | 64.06 | 79.06 | 84.61 | 68.01 | MistralForCausalLM |
| MBX-7B-v3-DPO | Fine Tuned Models | 72.4 | 76.13 | 73.55 | 89.11 | 64.91 | 74.0 | 85.56 | 69.67 | MistralForCausalLM |
| CarbonBeagle-11B-truthy | Chat Models | 107.3 | 76.1 | 72.27 | 89.31 | 66.55 | 78.55 | 83.82 | 66.11 | MistralForCausalLM |
| OmniBeagleSquaredMBX-v3-7B-v2 | Merged Models or MoE Models | 72.4 | 75.98 | 74.06 | 88.93 | 64.53 | 72.93 | 85.56 | 69.9 | MistralForCausalLM |
| NeuralOmniBeagleMBX-v3-7B | Merged Models or MoE Models | 72.4 | 75.93 | 73.38 | 88.91 | 64.99 | 73.1 | 84.21 | 70.96 | MistralForCausalLM |
| FusionNet_7Bx2_MoE_14B | Fine Tuned Models | 128.8 | 75.91 | 73.55 | 88.84 | 64.68 | 69.6 | 88.16 | 70.66 | MixtralForCausalLM |
| MiquMaid-v2-2x70B-DPO | Fine Tuned Models | 1253.5 | 75.89 | 72.53 | 88.36 | 75.31 | 66.5 | 85.32 | 67.32 | MixtralForCausalLM |
| AiMaven-Prometheus | Fine Tuned Models | 72.4 | 75.74 | 73.98 | 88.83 | 65.17 | 72.22 | 85.16 | 69.07 | MistralForCausalLM |
| Pluto_24B_DPO_63 | Fine Tuned Models | 128.8 | 75.63 | 73.98 | 88.17 | 64.49 | 79.36 | 81.69 | 66.11 | MixtralForCausalLM |
| Phoenix_DPO_60B | Fine Tuned Models | 608.1 | 75.48 | 71.16 | 85.46 | 77.66 | 63.84 | 84.93 | 69.83 | MixtralForCausalLM |
| Helion-4x34B | Fine Tuned Models | 1136.6 | 75.48 | 69.71 | 85.28 | 77.33 | 63.91 | 84.37 | 72.25 | MixtralForCausalLM |
| UNA-34BeagleSimpleMath-32K-v1 | Fine Tuned Models | 343.9 | 75.45 | 74.15 | 85.98 | 76.52 | 73.74 | 83.27 | 59.06 | LlamaForCausalLM |
| Senku-70B-Full | Fine Tuned Models | 689.8 | 75.44 | 71.5 | 87.88 | 75.2 | 61.96 | 84.77 | 71.34 | LlamaForCausalLM |
| UNA-34Beagles-32K-bf16-v1 | Fine Tuned Models | 343.9 | 75.41 | 73.55 | 85.93 | 76.45 | 73.55 | 82.95 | 60.05 | LlamaForCausalLM |
| Cosmosis-3x34B | Fine Tuned Models | 872.4 | 75.39 | 69.71 | 85.18 | 77.25 | 63.82 | 84.14 | 72.25 | MixtralForCausalLM |
| WestLake-7B-v2-laser-truthy-dpo | Fine Tuned Models | 72.4 | 75.37 | 73.89 | 88.85 | 64.84 | 69.81 | 86.66 | 68.16 | MistralForCausalLM |
| Senku-70B-Full | Fine Tuned Models | 689.8 | 75.36 | 71.33 | 87.86 | 75.14 | 61.95 | 84.53 | 71.34 | LlamaForCausalLM |
| WestSeverus-7B-DPO-v2 | Fine Tuned Models | 72.4 | 75.29 | 71.42 | 88.27 | 64.79 | 72.37 | 83.27 | 71.65 | MistralForCausalLM |
| test3_sft_16bit | Fine Tuned Models | 72.4 | 75.28 | 73.55 | 88.87 | 64.63 | 69.77 | 84.45 | 70.43 | MistralForCausalLM |
| Faraday-7B | Fine Tuned Models | 72.4 | 75.25 | 72.27 | 88.9 | 64.69 | 73.07 | 85.32 | 67.25 | MistralForCausalLM |
| Astralis-4x34B | Fine Tuned Models | 1136.6 | 75.24 | 69.71 | 85.17 | 77.24 | 63.55 | 84.14 | 71.65 | MixtralForCausalLM |
| Faraday-7B | Fine Tuned Models | 72.4 | 75.22 | 72.44 | 88.91 | 64.68 | 73.03 | 85.56 | 66.72 | MistralForCausalLM |
| WestSeverus-7B-DPO | Fine Tuned Models | 72.4 | 75.17 | 70.73 | 88.01 | 64.93 | 70.53 | 83.5 | 73.31 | MistralForCausalLM |
| Sectumsempra-7B-DPO | Fine Tuned Models | 72.4 | 75.14 | 71.5 | 88.7 | 64.9 | 72.49 | 83.19 | 70.05 | MistralForCausalLM |
| MiquMaid-v1-70B | Fine Tuned Models | 700 | 75.12 | 71.67 | 87.96 | 74.9 | 61.79 | 85.08 | 69.29 | LlamaForCausalLM |
| Bagel-Hermes-2x34B | Fine Tuned Models | 608.1 | 75.1 | 69.8 | 85.26 | 77.24 | 64.82 | 84.77 | 68.69 | MixtralForCausalLM |
| WestLakeX-7B-EvoMerge-Variant2 | Fine Tuned Models | 72.4 | 75.04 | 72.53 | 88.52 | 64.77 | 70.35 | 85.79 | 68.31 | MistralForCausalLM |
| test3_sft_16bit_dpo2 | Chat Models | 72.4 | 74.98 | 73.63 | 89.03 | 64.63 | 70.71 | 84.37 | 67.48 | MistralForCausalLM |
| MetaMath-Bagel-DPO-34B | Chat Models | 343.9 | 74.8 | 68.17 | 84.23 | 76.54 | 65.44 | 82.24 | 72.18 | LlamaForCausalLM |
| WestLake-7B-v2-laser | Fine Tuned Models | 72.4 | 74.78 | 73.29 | 88.66 | 64.72 | 67.04 | 86.74 | 68.23 | MistralForCausalLM |
| raccoon-small | Chat Models | 191.9 | 74.78 | 74.4 | 88.73 | 64.55 | 76.74 | 87.37 | 56.86 | MixtralForCausalLM |
| 60B_MoE_Coder_v3 | Fine Tuned Models | 608.1 | 74.75 | 71.16 | 85.44 | 75.37 | 67.01 | 82.56 | 66.94 | MixtralForCausalLM |
| nontoxic-bagel-34b-v0.2 | Fine Tuned Models | 343.9 | 74.69 | 72.44 | 85.64 | 76.41 | 72.7 | 82.48 | 58.45 | LlamaForCausalLM |
| bagel-dpo-34b-v0.2 | Chat Models | 343.9 | 74.69 | 71.93 | 85.25 | 76.58 | 70.05 | 83.35 | 60.96 | LlamaForCausalLM |
| WestLake-7B-v2 | Fine Tuned Models | 72.4 | 74.68 | 73.04 | 88.65 | 64.71 | 67.06 | 86.98 | 67.63 | MistralForCausalLM |
| MoMo-72B-LoRA-V1.4 | Fine Tuned Models | 722.9 | 74.67 | 69.2 | 85.07 | 77.12 | 62.66 | 83.74 | 70.2 | LlamaForCausalLM |
| llamaRAGdrama | Chat Models | 72.4 | 74.65 | 72.01 | 88.83 | 64.5 | 70.24 | 86.66 | 65.66 | MistralForCausalLM |
| Mixtral-7Bx2-truthy | Chat Models | 128.8 | 74.64 | 72.18 | 87.88 | 65.2 | 74.68 | 80.66 | 67.25 | MixtralForCausalLM |
| MoMo-72B-LoRA-V1.4 | Fine Tuned Models | 722.9 | 74.64 | 69.11 | 85.0 | 77.26 | 62.71 | 83.74 | 69.98 | LlamaForCausalLM |
| bagel-dpo-34b-v0.2 | Fine Tuned Models | 343.9 | 74.5 | 72.01 | 85.24 | 76.58 | 70.16 | 83.03 | 59.97 | LlamaForCausalLM |
| Westlake-7B | Fine Tuned Models | 72.4 | 74.48 | 73.21 | 88.49 | 64.64 | 67.36 | 86.03 | 67.17 | MistralForCausalLM |
| MM-OV-bagel-DPO-34b-c1000-250 | Chat Models | 343.9 | 74.47 | 68.17 | 83.97 | 76.33 | 63.67 | 82.4 | 72.25 | LlamaForCausalLM |
| WestLakeX-7B-EvoMerge | Fine Tuned Models | 72.4 | 74.37 | 71.42 | 88.08 | 64.84 | 67.5 | 84.77 | 69.6 | MistralForCausalLM |
| Truthful_DPO_MOE_19B | Chat Models | 191.9 | 74.3 | 71.08 | 88.46 | 66.13 | 72.29 | 83.35 | 64.52 | MixtralForCausalLM |
| CarbonVillain-en-13B-v1 | Chat Models | 107.3 | 74.28 | 71.25 | 88.46 | 66.42 | 71.98 | 83.27 | 64.29 | Unknown |
| Patronum-7B | Fine Tuned Models | 72.4 | 74.27 | 71.67 | 88.33 | 64.84 | 70.41 | 81.85 | 68.54 | MistralForCausalLM |
| SOLAR-10B-OrcaDPO-Jawade | Chat Models | 107.3 | 74.27 | 71.16 | 88.27 | 66.12 | 71.57 | 83.66 | 64.82 | LlamaForCausalLM |
| LaserPipe-7B-SLERP | Merged Models or MoE Models | 72.4 | 74.22 | 71.08 | 87.89 | 64.86 | 65.38 | 83.35 | 72.78 | MistralForCausalLM |
| SauerkrautLM-SOLAR-Instruct | Fine Tuned Models | 107.3 | 74.21 | 70.82 | 88.63 | 66.2 | 71.95 | 83.5 | 64.14 | LlamaForCausalLM |
| SOLAR-10.7B-Instruct-v1.0 | Chat Models | 107.3 | 74.2 | 71.08 | 88.16 | 66.21 | 71.43 | 83.58 | 64.75 | LlamaForCausalLM |
| UNA-SOLAR-10.7B-Instruct-v1.0 | Fine Tuned Models | 107.3 | 74.2 | 70.56 | 88.18 | 66.08 | 72.05 | 83.66 | 64.67 | LlamaForCausalLM |
| SOLAR-10B-Nector-DPO-Jawade | Chat Models | 107.3 | 74.19 | 71.33 | 88.62 | 66.22 | 70.92 | 83.43 | 64.59 | LlamaForCausalLM |
| SOLAR-Instruct-ko-Adapter-Attach | Chat Models | 107.3 | 74.11 | 71.08 | 88.2 | 66.09 | 71.51 | 83.5 | 64.29 | LlamaForCausalLM |
| SOLAR-10.7b-Instruct-truthy-dpo | Fine Tuned Models | 107.3 | 74.11 | 72.1 | 88.44 | 65.45 | 76.75 | 82.72 | 59.21 | LlamaForCausalLM |
| 19B_MATH_DPO | Fine Tuned Models | 191.9 | 74.1 | 71.08 | 88.43 | 66.25 | 72.11 | 82.95 | 63.76 | MixtralForCausalLM |
| LaserPipe-7B-SLERP | Merged Models or MoE Models | 72.4 | 74.08 | 70.82 | 87.88 | 64.77 | 65.34 | 83.27 | 72.4 | MistralForCausalLM |
| BrokenKeyboard | Chat Models | 107.3 | 74.08 | 71.25 | 88.34 | 66.04 | 71.36 | 83.19 | 64.29 | LlamaForCausalLM |
| UNA-SOLAR-10.7B-Instruct-v1.0 | Chat Models | 107.3 | 74.07 | 70.73 | 88.32 | 66.1 | 72.52 | 83.35 | 63.38 | LlamaForCausalLM |
| UNA-POLAR-10.7B-InstructMath-v2 | Fine Tuned Models | 107.3 | 74.07 | 70.73 | 88.2 | 66.03 | 71.73 | 82.95 | 64.75 | LlamaForCausalLM |
| LMCocktail-10.7B-v1 | Fine Tuned Models | 107.3 | 74.06 | 70.65 | 88.13 | 66.21 | 71.03 | 83.35 | 64.97 | Unknown |
| UNAversal-2x7B-v1 | Fine Tuned Models | 128.8 | 74.05 | 73.38 | 87.87 | 63.49 | 69.93 | 82.08 | 67.55 | MixtralForCausalLM |
| Merge_Sakura_Solar | Merged Models or MoE Models | 107.3 | 74.03 | 70.73 | 88.51 | 66.03 | 72.21 | 82.72 | 63.99 | LlamaForCausalLM |
| Laser-WestLake-2x7b | Merged Models or MoE Models | 128.8 | 74.0 | 72.27 | 88.44 | 64.71 | 69.25 | 85.79 | 63.53 | MixtralForCausalLM |
| MixTAO-7Bx2-MoE-DPO | Fine Tuned Models | 128.8 | 73.94 | 70.9 | 87.12 | 64.72 | 69.34 | 81.22 | 70.36 | MixtralForCausalLM |
| meow | Fine Tuned Models | 107.3 | 73.94 | 70.48 | 88.08 | 66.25 | 70.49 | 83.43 | 64.9 | Unknown |
| UNA-TheBeagle-7b-v1 | Chat Models | 72.4 | 73.87 | 73.04 | 88.0 | 63.48 | 69.85 | 82.16 | 66.72 | MistralForCausalLM |
| CCK_Gony_v3 | Fine Tuned Models | 467 | 73.83 | 71.33 | 88.71 | 71.07 | 73.33 | 81.22 | 57.32 | MixtralForCausalLM |
| UNAversal-8x7B-v1beta | Fine Tuned Models | 467 | 73.78 | 69.8 | 86.9 | 70.39 | 71.97 | 82.0 | 61.64 | MixtralForCausalLM |
| tulu-2-dpo-70b | Chat Models | 689.8 | 73.77 | 72.1 | 88.99 | 69.84 | 65.78 | 83.27 | 62.62 | LlamaForCausalLM |
| Aurora-Nights-70B-v1.0 | Fine Tuned Models | 689.8 | 73.77 | 71.33 | 88.33 | 70.47 | 62.81 | 83.35 | 66.34 | LlamaForCausalLM |
| Lumosia-v2-MoE-4x10.7 | Merged Models or MoE Models | 361 | 73.75 | 70.39 | 87.87 | 66.45 | 68.48 | 84.21 | 65.13 | MixtralForCausalLM |
| SuperBruphin-3x7B | Fine Tuned Models | 185.2 | 73.75 | 71.16 | 87.74 | 64.58 | 66.85 | 81.53 | 70.66 | MixtralForCausalLM |
| Nous-Hermes-2-Yi-34B | Fine Tuned Models | 343.9 | 73.74 | 66.89 | 85.49 | 76.7 | 60.37 | 82.95 | 70.05 | LlamaForCausalLM |
| kellemar-DPO-Orca-Distilled-7B-SLERP | Chat Models | 72.4 | 73.71 | 70.48 | 87.56 | 65.33 | 64.97 | 81.93 | 72.02 | MistralForCausalLM |
| MM-Orc-Vic-bagel-34b-c1000 | Fine Tuned Models | 343.9 | 73.68 | 67.32 | 83.52 | 76.09 | 60.57 | 82.32 | 72.25 | LlamaForCausalLM |
⚠️数据仅供参考,以官方来源为准。模型名称旁的链接可跳转到 DataLearner 模型详情页。