加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
数据来源: HuggingFace
| 模型名称 | 模型类型 | 参数大小(亿) | 平均分 | ARC分数 | HellaSwag分数 | MMLU分数 | TruthfulQA分数 | Winogrande分数 | GSM8K分数 | 模型架构 |
|---|---|---|---|---|---|---|---|---|---|---|
| openchat_v3.2_super | Fine Tuned Models | 0 | 54.99 | 59.81 | 82.5 | 55.9 | 42.3 | 75.93 | 13.5 | LlamaForCausalLM |
| chinese-alpaca-2-13b | Fine Tuned Models | 129.7 | 54.99 | 58.7 | 79.74 |
数据仅供参考,以官方来源为准。模型名称旁的链接可跳转到 DataLearner 模型详情页。
| 55.1 |
| 50.22 |
| 75.69 |
| 10.46 |
| Unknown |
| Asimov-7B-v1 | Fine Tuned Models | 70 | 54.98 | 59.04 | 80.04 | 56.35 | 51.15 | 73.95 | 9.33 | MistralForCausalLM |
| vicuna-13b-v1.5-16k | Fine Tuned Models | 130 | 54.97 | 56.74 | 80.37 | 55.28 | 51.96 | 72.38 | 13.12 | LlamaForCausalLM |
| Hippolyta-7B-bf16 | Fine Tuned Models | 72.4 | 54.96 | 60.58 | 79.98 | 57.71 | 55.74 | 73.95 | 1.82 | MistralForCausalLM |
| Mistral-7B-Instruct-v0.1 | Chat Models | 72.4 | 54.96 | 54.52 | 75.63 | 55.38 | 56.28 | 73.72 | 14.25 | MistralForCausalLM |
| synapsellm-7b-mistral-v0.4-preview3 | Fine Tuned Models | 72.4 | 54.94 | 51.28 | 74.83 | 52.93 | 52.35 | 73.48 | 24.79 | MistralForCausalLM |
| Sydney_Overthinker_13b_HF | Chat Models | 130.2 | 54.94 | 58.96 | 80.85 | 51.28 | 45.7 | 73.95 | 18.88 | LlamaForCausalLM |
| wizardLM-13B-1.0-fp16 | Fine Tuned Models | 128.5 | 54.93 | 57.25 | 80.88 | 52.9 | 50.55 | 74.11 | 13.87 | Unknown |
| Yi-6B-200K-AEZAKMI-v2 | Fine Tuned Models | 60.6 | 54.93 | 52.99 | 71.2 | 63.0 | 46.79 | 70.48 | 25.09 | LlamaForCausalLM |
| Mistral-7B-AEZAKMI-v1 | Chat Models | 72.4 | 54.92 | 58.87 | 82.01 | 58.72 | 53.54 | 75.69 | 0.68 | MistralForCausalLM |
| CodeUp-Llama-2-13b-chat-hf | Fine Tuned Models | 130 | 54.91 | 59.04 | 81.93 | 54.63 | 44.12 | 74.51 | 15.24 | LlamaForCausalLM |
| Platypus2-13B | Fine Tuned Models | 130.2 | 54.89 | 61.26 | 82.56 | 56.7 | 44.86 | 76.87 | 7.05 | LlamaForCausalLM |
| LewdEngine | Fine Tuned Models | 0 | 54.88 | 60.49 | 83.08 | 54.84 | 43.63 | 74.9 | 12.36 | LlamaForCausalLM |
| Redmond-Puffin-13B | Fine Tuned Models | 130 | 54.86 | 60.41 | 83.2 | 55.36 | 42.12 | 76.64 | 11.45 | LlamaForCausalLM |
| OpenOrcaPlatypus2-Platypus2-13B-QLora-0.80-epoch | Chat Models | 130.2 | 54.86 | 59.81 | 82.69 | 56.96 | 52.92 | 74.43 | 2.35 | Unknown |
| Code-13B | Fine Tuned Models | 130 | 54.81 | 57.34 | 83.28 | 53.17 | 42.46 | 73.56 | 19.03 | LlamaForCausalLM |
| Samantha-1.11-CodeLlama-34b | Fine Tuned Models | 334.8 | 54.8 | 56.57 | 75.47 | 53.51 | 50.46 | 73.48 | 19.33 | Unknown |
| Ensemble5-Platypus2-13B-QLora-0.80-epoch | Chat Models | 130.2 | 54.76 | 59.73 | 82.66 | 56.94 | 52.92 | 74.43 | 1.9 | Unknown |
| WizardLM-13B-V1.2 | Fine Tuned Models | 130 | 54.76 | 59.04 | 82.21 | 54.64 | 47.27 | 71.9 | 13.5 | LlamaForCausalLM |
| Redmond-Puffin-13B | Fine Tuned Models | 130 | 54.74 | 60.49 | 83.21 | 54.95 | 42.08 | 76.48 | 11.22 | LlamaForCausalLM |
| MythoMix-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 54.74 | 60.32 | 83.72 | 55.74 | 52.18 | 75.53 | 0.91 | Unknown |
| Medusa-13b | Fine Tuned Models | 130.2 | 54.72 | 58.19 | 81.35 | 57.39 | 51.24 | 73.32 | 6.82 | LlamaForCausalLM |
| ELYZA-japanese-Llama-2-13b-instruct | Fine Tuned Models | 130 | 54.72 | 58.36 | 82.2 | 55.65 | 42.4 | 75.22 | 14.48 | LlamaForCausalLM |
| speechless-codellama-34b-v1.9 | Fine Tuned Models | 340 | 54.64 | 54.27 | 75.2 | 56.12 | 43.92 | 73.56 | 24.79 | LlamaForCausalLM |
| llama-2-13B-instructed | Chat Models | 130 | 54.63 | 59.39 | 83.88 | 55.57 | 46.89 | 74.03 | 8.04 | LlamaForCausalLM |
| UltraLM-13B-fp16 | Fine Tuned Models | 130 | 54.62 | 57.59 | 80.2 | 51.85 | 51.56 | 75.85 | 10.69 | LlamaForCausalLM |
| Chat-Stheno-L2-13B | Fine Tuned Models | 130 | 54.61 | 58.45 | 80.96 | 54.8 | 43.31 | 75.37 | 14.78 | LlamaForCausalLM |
| Llama-2-13b-hf-ds_eli5_1024_r_64_alpha_16 | Chat Models | 130 | 54.61 | 60.41 | 82.58 | 55.86 | 43.61 | 76.72 | 8.49 | Unknown |
| Nous-Hermes-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 54.6 | 59.9 | 83.29 | 56.69 | 51.08 | 75.22 | 1.44 | Unknown |
| Samantha-Nebula-7B | Chat Models | 72.4 | 54.58 | 57.0 | 82.25 | 54.21 | 49.58 | 73.09 | 11.37 | MistralForCausalLM |
| Aquila2-34B | Fine Tuned Models | 340 | 54.57 | 52.47 | 81.9 | 76.03 | 40.85 | 75.53 | 0.61 | LlamaForCausalLM |
| xxxI-Ixxx | Fine Tuned Models | 72.4 | 54.56 | 54.18 | 72.55 | 52.02 | 54.42 | 70.24 | 23.96 | LlamaForCausalLM |
| Orca-2-7b | Fine Tuned Models | 70 | 54.55 | 54.1 | 76.19 | 56.37 | 52.45 | 73.48 | 14.71 | LlamaForCausalLM |
| vicuna-class-shishya-all-hal-13b-ep3 | Fine Tuned Models | 130 | 54.51 | 48.63 | 80.28 | 56.4 | 42.75 | 73.16 | 25.85 | LlamaForCausalLM |
| gpt4-alpaca-lora-13b-decapoda-1024 | Fine Tuned Models | 130 | 54.51 | 59.39 | 81.87 | 47.75 | 52.59 | 77.35 | 8.11 | Unknown |
| Aquila2-34B | Pretrained Models | 340 | 54.5 | 52.65 | 81.99 | 76.02 | 40.8 | 75.06 | 0.45 | LlamaForCausalLM |
| Camelidae-8x7B | Chat Models | 70 | 54.47 | 55.63 | 79.18 | 50.1 | 42.86 | 76.24 | 22.82 | LlamaForCausalLM |
| Limarp-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 54.46 | 60.49 | 82.76 | 56.52 | 44.14 | 76.8 | 6.07 | Unknown |
| zephyr-7b-beta-128k | Fine Tuned Models | 72.4 | 54.45 | 58.28 | 81.0 | 53.57 | 46.1 | 74.74 | 13.04 | MistralForCausalLM |
| llama-2-13b-FINETUNE1_17w-r16 | Fine Tuned Models | 130 | 54.37 | 57.25 | 82.27 | 56.16 | 39.75 | 77.43 | 13.34 | LlamaForCausalLM |
| llama-2-13b-huangyt_Fintune_1_17w-q_k_v_o_proj | Fine Tuned Models | 128.5 | 54.35 | 59.73 | 81.06 | 54.53 | 38.64 | 78.14 | 14.03 | Unknown |
| TarsDolly | Fine Tuned Models | 72.4 | 54.35 | 59.3 | 81.85 | 56.26 | 42.29 | 76.24 | 10.16 | MistralForCausalLM |
| XwinCoder-34B | Fine Tuned Models | 340 | 54.35 | 51.02 | 74.02 | 49.53 | 43.82 | 68.35 | 39.35 | LlamaForCausalLM |
| orca_mini_v3_7B-GPTQ | Unkown Model Types | 90.5 | 54.35 | 54.52 | 78.53 | 51.85 | 51.2 | 74.66 | 15.31 | LlamaForCausalLM |
| vigogne-13b-instruct | Fine Tuned Models | 130 | 54.34 | 57.94 | 81.32 | 47.62 | 50.23 | 77.11 | 11.83 | LlamaForCausalLM |
| llama-op-v4 | Fine Tuned Models | 128.5 | 54.34 | 61.52 | 79.21 | 57.01 | 42.72 | 75.93 | 9.63 | Unknown |
| Alpagasus-2-13b-QLoRA-merged | Fine Tuned Models | 130 | 54.31 | 61.09 | 82.46 | 55.27 | 38.53 | 77.35 | 11.14 | LlamaForCausalLM |
| WizardLM-13B-V1.1-GPTQ | Unkown Model Types | 162.2 | 54.28 | 58.53 | 80.66 | 49.59 | 54.35 | 74.43 | 8.11 | LlamaForCausalLM |
| SwahiliInstruct-v0.2 | Fine Tuned Models | 72.4 | 54.25 | 55.2 | 78.22 | 50.3 | 57.08 | 73.24 | 11.45 | MistralForCausalLM |
| llama-2-13b-code-alpaca | Fine Tuned Models | 128.5 | 54.25 | 60.84 | 82.14 | 55.93 | 38.27 | 76.4 | 11.9 | Unknown |
| EverythingLM-13b-V3-peft | Chat Models | 128.5 | 54.24 | 58.36 | 81.03 | 54.7 | 52.98 | 72.85 | 5.53 | Unknown |
| Llama-2-13b-hf_Open-Platypus | Fine Tuned Models | 128.5 | 54.22 | 58.87 | 82.14 | 54.98 | 42.84 | 77.11 | 9.4 | Unknown |
| llama-2-13b-hf-platypus | Chat Models | 130.2 | 54.22 | 58.87 | 82.14 | 54.98 | 42.84 | 77.11 | 9.4 | LlamaForCausalLM |
| Alpagasus-2-13b-QLoRA-merged | Chat Models | 130 | 54.2 | 60.84 | 82.43 | 55.55 | 38.65 | 76.87 | 10.84 | LlamaForCausalLM |
| PuddleJumper-13b-V2 | Chat Models | 130 | 54.19 | 57.0 | 81.06 | 58.3 | 52.66 | 72.45 | 3.64 | LlamaForCausalLM |
| llama-2-13b-FINETUNE1_17w-r4 | Fine Tuned Models | 130 | 54.18 | 56.74 | 82.27 | 56.18 | 39.65 | 77.03 | 13.19 | LlamaForCausalLM |
| Wizard-Vicuna-13B-Uncensored | Fine Tuned Models | 128.5 | 54.14 | 58.96 | 81.95 | 47.92 | 51.69 | 75.69 | 8.64 | Unknown |
| Wizard-Vicuna-13B-Uncensored-HF | Fine Tuned Models | 130 | 54.14 | 58.96 | 81.95 | 47.92 | 51.69 | 75.69 | 8.64 | LlamaForCausalLM |
| Llama-2-13b-hf-eli5-wiki-1024_r_64_alpha_16 | Chat Models | 130 | 54.14 | 59.98 | 82.43 | 55.41 | 39.9 | 76.56 | 10.54 | Unknown |
| speechless-orca-platypus-coig-lite-4k-0.5e-13b | Fine Tuned Models | 130 | 54.13 | 58.02 | 80.15 | 57.26 | 48.04 | 75.45 | 5.84 | LlamaForCausalLM |
| chinese-alpaca-2-13b-16k | Chat Models | 130 | 54.12 | 55.03 | 77.41 | 51.28 | 46.5 | 73.4 | 21.08 | LlamaForCausalLM |
| llama-2-13b-huangyt_Fintune_1_17w-gate_up_down_proj | Fine Tuned Models | 128.5 | 54.12 | 57.17 | 82.26 | 55.89 | 39.93 | 76.56 | 12.89 | Unknown |
| Guanaco-13B-Uncensored | Fine Tuned Models | 130 | 54.1 | 59.56 | 82.7 | 53.65 | 43.26 | 76.32 | 9.1 | LlamaForCausalLM |
| Yi-6B | Pretrained Models | 60.6 | 54.08 | 55.55 | 76.57 | 64.11 | 41.96 | 74.19 | 12.13 | LlamaForCausalLM |
| StableBeluga1-Delta | Fine Tuned Models | 652.9 | 54.08 | 68.17 | 85.88 | 64.83 | 55.81 | 49.8 | 0.0 | LlamaForCausalLM |
| ReMM-L2-13B-PIPPA | Fine Tuned Models | 130 | 54.06 | 59.73 | 83.12 | 54.1 | 49.94 | 74.51 | 2.96 | LlamaForCausalLM |
| ReMM-L2-13B | Fine Tuned Models | 130.2 | 54.06 | 59.73 | 83.1 | 54.11 | 49.94 | 74.51 | 2.96 | Unknown |
| airoboros-13B-HF | Fine Tuned Models | 130 | 54.05 | 58.28 | 81.05 | 50.03 | 51.57 | 76.24 | 7.13 | LlamaForCausalLM |
| Nous-Hermes-13b | Fine Tuned Models | 130 | 54.04 | 56.57 | 82.11 | 50.44 | 51.5 | 75.3 | 8.34 | LlamaForCausalLM |
| mnsim-dpo-peftmerged-2-eos | Chat Models | 131.6 | 54.04 | 55.63 | 77.82 | 51.25 | 46.37 | 76.24 | 16.91 | LlamaForCausalLM |
| airoboros-13b | Fine Tuned Models | 130 | 54.02 | 58.28 | 81.05 | 50.03 | 51.57 | 76.24 | 6.97 | LlamaForCausalLM |
| Yi-6B | Pretrained Models | 60.6 | 54.02 | 55.55 | 76.42 | 63.85 | 41.86 | 73.8 | 12.66 | LlamaForCausalLM |
| MythicalDestroyerV2-Platypus2-13B-QLora-0.80-epoch | Chat Models | 130.2 | 54.01 | 57.34 | 81.24 | 55.64 | 55.98 | 73.88 | 0.0 | Unknown |
| based-30b | Fine Tuned Models | 323.2 | 54.0 | 63.91 | 85.67 | 58.28 | 35.7 | 80.11 | 0.3 | Unknown |
| speechless-orca-platypus-coig-lite-4k-0.6e-13b | Fine Tuned Models | 130 | 53.99 | 58.79 | 79.93 | 56.77 | 48.29 | 75.93 | 4.25 | LlamaForCausalLM |
| webMistral-7B | Fine Tuned Models | 72.4 | 53.97 | 59.04 | 80.89 | 59.0 | 39.71 | 76.32 | 8.87 | MistralForCausalLM |
| WizardMath-13B-V1.0 | Chat Models | 130 | 53.97 | 60.07 | 82.01 | 54.8 | 42.7 | 71.9 | 12.36 | LlamaForCausalLM |
| minotaur-13b | Fine Tuned Models | 130 | 53.97 | 56.4 | 79.13 | 49.61 | 49.62 | 76.56 | 12.51 | LlamaForCausalLM |
| openchat_v2_w | Fine Tuned Models | 0 | 53.96 | 57.34 | 81.23 | 50.17 | 50.7 | 75.93 | 8.42 | LlamaForCausalLM |
| openchat_v2 | Fine Tuned Models | 0 | 53.96 | 57.17 | 81.14 | 50.58 | 49.54 | 76.24 | 9.1 | LlamaForCausalLM |
| Platypus-Nebula-v2-7B | Chat Models | 72.4 | 53.95 | 55.38 | 83.02 | 56.07 | 46.94 | 72.22 | 10.08 | MistralForCausalLM |
| llama-2-13b-huangyt_FINETUNE2_3w-gate_up_down_proj | Fine Tuned Models | 128.5 | 53.95 | 57.42 | 82.42 | 55.57 | 39.19 | 77.03 | 12.05 | Unknown |
| MedMerge-6-7b-alpha-dpo | Fine Tuned Models | 70 | 53.94 | 54.27 | 75.6 | 52.65 | 43.94 | 71.03 | 26.16 | ? |
| KoSoLAR-10.7B-v0.2_1.4_dedup | Fine Tuned Models | 108 | 53.93 | 60.07 | 82.18 | 61.3 | 45.38 | 74.66 | 0.0 | LlamaForCausalLM |
| Pwen-7B-Chat-20_30 | Fine Tuned Models | 70 | 53.93 | 51.45 | 73.99 | 62.08 | 47.01 | 68.43 | 20.62 | Unknown |
| Ferret-7B | Chat Models | 70 | 53.93 | 62.29 | 81.31 | 60.27 | 40.01 | 77.66 | 2.05 | Unknown |
| llama-2-13b-chat-platypus | Chat Models | 130.2 | 53.92 | 53.84 | 80.67 | 54.44 | 46.23 | 76.01 | 12.36 | LlamaForCausalLM |
| llama-2-13b-huangyt_Fintune_1_17w | Fine Tuned Models | 128.5 | 53.91 | 59.47 | 81.0 | 54.31 | 38.17 | 77.27 | 13.27 | Unknown |
| MythoBoros-13b | Fine Tuned Models | 130 | 53.9 | 58.19 | 81.75 | 50.13 | 48.93 | 75.77 | 8.64 | LlamaForCausalLM |
| airoboros-13b-gpt4-1.4 | Fine Tuned Models | 130 | 53.87 | 59.64 | 83.22 | 47.56 | 48.82 | 76.24 | 7.73 | LlamaForCausalLM |
| Ferret_7B | Chat Models | 70 | 53.87 | 62.29 | 81.33 | 60.09 | 39.94 | 77.51 | 2.05 | MistralForCausalLM |
| Ferret-7B | Chat Models | 70 | 53.87 | 62.29 | 81.33 | 60.09 | 39.94 | 77.51 | 2.05 | Unknown |
| llama2-13b-ft-openllm-leaderboard-v1 | Unkown Model Types | 130.2 | 53.86 | 59.64 | 83.14 | 60.93 | 40.72 | 77.35 | 1.36 | LlamaForCausalLM |
| MythoLogic-13b | Fine Tuned Models | 130 | 53.85 | 58.45 | 81.56 | 49.36 | 49.47 | 75.61 | 8.64 | LlamaForCausalLM |
| Libra-19B | Chat Models | 190 | 53.83 | 60.58 | 82.04 | 55.57 | 48.41 | 76.32 | 0.08 | LlamaForCausalLM |
| llama-2-13b-Open_Platypus_and_ccp_2.6w-3_epoch | Fine Tuned Models | 130 | 53.8 | 58.62 | 82.56 | 55.84 | 42.09 | 76.64 | 7.05 | LlamaForCausalLM |
| test-help-steer-filtered-orig | Fine Tuned Models | 72.4 | 53.77 | 57.59 | 80.42 | 57.24 | 41.1 | 76.64 | 9.63 | MistralForCausalLM |
| GiftedConvo13bLoraNoEconsE4 | Chat Models | 130 | 53.74 | 59.9 | 84.11 | 54.67 | 41.94 | 74.03 | 7.81 | Unknown |
| llama-2-13b-huangyt_FINETUNE2_3w-q_k_v_o_proj | Fine Tuned Models | 128.5 | 53.74 | 58.53 | 82.47 | 53.9 | 37.92 | 76.8 | 12.81 | Unknown |