加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
Data source: HuggingFace
| Model | Type | Parameters (B) | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | Architecture |
|---|---|---|---|---|---|---|---|---|---|---|
| Stheno-1.2-L2-13B | Fine Tuned Models | 130 | 56.15 | 60.75 | 83.67 | 56.27 | 50.32 | 74.98 | 10.92 | LlamaForCausalLM |
| Orca-2-13b-SFT-v6 | Chat Models | 130.2 | 56.15 | 60.41 | 80.46 |
Data is for reference only. Official sources are authoritative. Click model names to view DataLearner model profiles.
| 59.51 |
| 54.01 |
| 77.43 |
| 5.08 |
| LlamaForCausalLM |
| ELYZA-japanese-Llama-2-13b | Fine Tuned Models | 130 | 56.14 | 57.0 | 80.89 | 54.38 | 40.43 | 76.87 | 27.29 | LlamaForCausalLM |
| SpeechlessV1-Nova-13B | Chat Models | 130.2 | 56.14 | 61.77 | 82.68 | 57.75 | 51.44 | 77.43 | 5.76 | Unknown |
| Instruct_Yi-6B_Dolly_CodeAlpaca | Fine Tuned Models | 60.6 | 56.11 | 53.16 | 75.3 | 63.06 | 41.42 | 75.37 | 28.35 | LlamaForCausalLM |
| NewHope_HF_not_official | Fine Tuned Models | 0 | 56.11 | 61.09 | 84.03 | 55.73 | 44.96 | 74.98 | 15.85 | LlamaForCausalLM |
| chronos-hermes-13b-v2 | Fine Tuned Models | 130 | 56.1 | 60.32 | 83.21 | 55.05 | 50.91 | 75.37 | 11.75 | LlamaForCausalLM |
| Nebula-7B | Chat Models | 72.4 | 56.1 | 59.3 | 83.46 | 57.0 | 45.56 | 76.4 | 14.86 | Unknown |
| prometheus-13b-v1.0 | Fine Tuned Models | 130 | 56.09 | 53.24 | 80.75 | 51.49 | 45.66 | 73.72 | 31.69 | LlamaForCausalLM |
| qCammel-13 | Fine Tuned Models | 0 | 56.05 | 60.84 | 83.66 | 56.73 | 47.54 | 76.16 | 11.37 | LlamaForCausalLM |
| ReMM-SLERP-L2-13B | Fine Tuned Models | 130 | 56.03 | 60.92 | 83.56 | 55.33 | 51.97 | 75.22 | 9.17 | LlamaForCausalLM |
| carl-33b | Fine Tuned Models | 330 | 56.03 | 64.59 | 85.27 | 58.38 | 45.32 | 76.24 | 6.37 | LlamaForCausalLM |
| synapsellm-7b-mistral-v0.5-preview | Fine Tuned Models | 72.4 | 56.03 | 52.73 | 76.51 | 54.67 | 55.16 | 74.35 | 22.74 | MistralForCausalLM |
| neural-chat-7b-v3-1-Nebula-v2-7B | Fine Tuned Models | 72.4 | 56.01 | 61.77 | 80.21 | 59.07 | 58.56 | 71.82 | 4.62 | MistralForCausalLM |
| MythoMax-L2-13b | Fine Tuned Models | 130 | 56.0 | 60.92 | 83.56 | 55.33 | 51.97 | 75.22 | 9.02 | LlamaForCausalLM |
| huginnv1.2 | Chat Models | 128.5 | 55.98 | 62.37 | 84.28 | 57.02 | 47.81 | 75.22 | 9.17 | Unknown |
| Nous-Hermes-Llama2-13b | Fine Tuned Models | 130 | 55.97 | 61.52 | 83.29 | 55.11 | 50.38 | 75.45 | 10.08 | LlamaForCausalLM |
| Samantha-1.11-13b | Fine Tuned Models | 128.5 | 55.97 | 60.84 | 82.99 | 55.96 | 47.72 | 76.01 | 12.28 | Unknown |
| LongQLoRA-Vicuna-13b-8k | Fine Tuned Models | 130 | 55.96 | 56.4 | 81.05 | 53.68 | 47.07 | 74.51 | 23.05 | LlamaForCausalLM |
| Walter-SOLAR-11B | Chat Models | 107.3 | 55.95 | 60.41 | 84.86 | 64.99 | 44.88 | 79.56 | 0.99 | LlamaForCausalLM |
| Nous-Hermes-13B-Code | Fine Tuned Models | 130 | 55.93 | 61.18 | 83.21 | 55.13 | 50.56 | 75.14 | 10.39 | LlamaForCausalLM |
| Chat-AYB-Platypus2-13B | Chat Models | 130.2 | 55.93 | 60.49 | 84.03 | 57.83 | 54.52 | 75.77 | 2.96 | Unknown |
| synapsellm-7b-mistral-v0.4-preview2 | Fine Tuned Models | 72.4 | 55.93 | 52.99 | 74.54 | 54.6 | 53.79 | 73.95 | 25.7 | MistralForCausalLM |
| synapsellm-7b-mistral-v0.5-preview2 | Fine Tuned Models | 72.4 | 55.93 | 52.22 | 75.54 | 51.64 | 55.47 | 73.09 | 27.6 | MistralForCausalLM |
| AppleSauce-L2-13b | Fine Tuned Models | 130.2 | 55.91 | 61.01 | 83.61 | 57.07 | 47.81 | 75.93 | 10.01 | LlamaForCausalLM |
| Synthia-13B-v1.2 | Fine Tuned Models | 130 | 55.9 | 61.26 | 82.93 | 56.47 | 47.27 | 76.48 | 10.99 | LlamaForCausalLM |
| openbuddy-llama2-34b-v11.1-bf16 | Fine Tuned Models | 335.3 | 55.88 | 50.0 | 71.19 | 55.71 | 53.01 | 70.8 | 34.57 | Unknown |
| vicuna-class-tutor-13b-ep3 | Fine Tuned Models | 130 | 55.88 | 57.34 | 81.51 | 57.02 | 52.99 | 74.35 | 12.05 | LlamaForCausalLM |
| Synatra-V0.1-7B-Instruct | Chat Models | 70 | 55.86 | 55.29 | 76.63 | 55.29 | 55.76 | 72.77 | 19.41 | MistralForCausalLM |
| Synatra-V0.1-7B | Fine Tuned Models | 71.1 | 55.86 | 55.29 | 76.63 | 55.29 | 55.76 | 72.77 | 19.41 | Unknown |
| Newton-7B | Fine Tuned Models | 72.4 | 55.85 | 63.99 | 81.72 | 62.78 | 44.36 | 78.85 | 3.41 | MistralForCausalLM |
| Mistral-7B-Instruct-v0.2-DARE | Fine Tuned Models | 72.4 | 55.84 | 61.95 | 75.62 | 49.99 | 54.36 | 74.98 | 18.12 | Unknown |
| Metamath-reproduce-7b | Fine Tuned Models | 70 | 55.81 | 47.18 | 73.65 | 42.94 | 41.58 | 71.35 | 58.15 | LlamaForCausalLM |
| llama-2-13b-OpenOrca_5w | Fine Tuned Models | 130 | 55.8 | 61.01 | 82.82 | 56.09 | 44.87 | 77.74 | 12.28 | LlamaForCausalLM |
| Nous-Hermes-Llama2-13b | Fine Tuned Models | 130 | 55.75 | 61.26 | 83.26 | 55.04 | 50.41 | 75.37 | 9.17 | LlamaForCausalLM |
| Stable-Platypus2-13B | Fine Tuned Models | 130.2 | 55.75 | 62.71 | 82.29 | 58.3 | 52.52 | 76.87 | 1.82 | LlamaForCausalLM |
| CollectiveCognition-v1.1-Nebula-7B | Fine Tuned Models | 72.4 | 55.72 | 58.11 | 82.39 | 57.03 | 53.53 | 73.72 | 9.55 | Unknown |
| openchat_v3.1 | Fine Tuned Models | 0 | 55.71 | 60.15 | 82.84 | 56.84 | 44.38 | 76.24 | 13.8 | LlamaForCausalLM |
| Stheno-1.1-L2-13B | Fine Tuned Models | 130 | 55.71 | 60.75 | 83.64 | 56.39 | 50.3 | 75.22 | 7.96 | LlamaForCausalLM |
| openchat_v3.2 | Fine Tuned Models | 0 | 55.68 | 59.64 | 82.68 | 56.68 | 44.49 | 76.95 | 13.65 | LlamaForCausalLM |
| ELYZA-japanese-Llama-2-13b-fast | Fine Tuned Models | 130 | 55.67 | 55.89 | 80.73 | 54.4 | 40.31 | 77.19 | 25.47 | LlamaForCausalLM |
| speechless-hermes-coig-lite-13b | Fine Tuned Models | 130.2 | 55.65 | 59.47 | 82.28 | 55.18 | 47.6 | 78.61 | 10.77 | LlamaForCausalLM |
| U-Amethyst-20B | Fine Tuned Models | 199.9 | 55.65 | 62.2 | 83.11 | 55.88 | 53.2 | 74.19 | 5.31 | LlamaForCausalLM |
| Uncensored-Frank-13B | Fine Tuned Models | 130 | 55.64 | 61.6 | 82.62 | 54.55 | 48.34 | 74.74 | 11.98 | LlamaForCausalLM |
| Nova-13B-50-step | Fine Tuned Models | 130.2 | 55.61 | 61.6 | 82.31 | 57.27 | 51.53 | 76.56 | 4.4 | Unknown |
| ANIMA-Phi-Neptune-Mistral-7B-v4 | Fine Tuned Models | 71.1 | 55.61 | 55.46 | 77.63 | 53.12 | 59.01 | 73.48 | 14.94 | Unknown |
| sqlcoder-34b-alpha | Fine Tuned Models | 340 | 55.59 | 54.18 | 75.93 | 54.42 | 40.63 | 73.48 | 34.87 | LlamaForCausalLM |
| Stable-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 55.56 | 62.29 | 82.46 | 57.09 | 51.41 | 76.56 | 3.56 | Unknown |
| ANIMA-Phi-Neptune-Mistral-7B | Fine Tuned Models | 70 | 55.54 | 55.97 | 76.22 | 52.89 | 59.76 | 73.48 | 14.94 | MistralForCausalLM |
| internlm-20b-chat | Fine Tuned Models | 200 | 55.53 | 55.38 | 78.58 | 58.53 | 43.22 | 78.77 | 18.73 | Unknown |
| llama-2-13b-dolphin_5w | Fine Tuned Models | 130 | 55.53 | 60.67 | 82.69 | 56.23 | 44.41 | 77.35 | 11.83 | LlamaForCausalLM |
| speechless-hermes-coig-lite-13b | Fine Tuned Models | 130.2 | 55.51 | 59.56 | 82.26 | 55.3 | 47.56 | 78.53 | 9.86 | LlamaForCausalLM |
| shisa-gamma-7b-v1 | Fine Tuned Models | 72.4 | 55.5 | 53.16 | 77.3 | 55.23 | 50.73 | 73.88 | 22.74 | MistralForCausalLM |
| Stheno-Inverted-1.2-L2-13B | Fine Tuned Models | 130 | 55.5 | 59.39 | 83.01 | 55.77 | 51.22 | 74.66 | 8.95 | LlamaForCausalLM |
| UndiMix-v1-13b | Fine Tuned Models | 130.2 | 55.5 | 59.47 | 82.45 | 55.83 | 49.78 | 75.45 | 10.01 | LlamaForCausalLM |
| chronolima-airo-grad-l2-13B | Fine Tuned Models | 130 | 55.5 | 59.56 | 83.47 | 55.8 | 44.58 | 75.61 | 13.95 | LlamaForCausalLM |
| openchat_v3.2 | Fine Tuned Models | 0 | 55.49 | 59.47 | 82.6 | 56.82 | 44.51 | 76.09 | 13.42 | LlamaForCausalLM |
| Zhongjing-LLaMA-base | Fine Tuned Models | 0 | 55.47 | 55.12 | 79.72 | 48.23 | 48.88 | 74.82 | 26.08 | LlamaForCausalLM |
| vicuna-13b-v1.5 | Fine Tuned Models | 130 | 55.41 | 57.08 | 81.24 | 56.67 | 51.51 | 74.66 | 11.3 | LlamaForCausalLM |
| model_007_13b_v2 | Fine Tuned Models | 128.5 | 55.41 | 61.95 | 82.48 | 57.32 | 53.5 | 75.85 | 1.36 | Unknown |
| llama2_13b_instructed_version2 | Fine Tuned Models | 130 | 55.41 | 60.07 | 84.05 | 55.61 | 46.12 | 75.61 | 10.99 | LlamaForCausalLM |
| Synthia-13B | Fine Tuned Models | 130 | 55.41 | 59.98 | 81.86 | 56.11 | 47.41 | 76.09 | 10.99 | LlamaForCausalLM |
| nash-vicuna-13b-v1dot5-ep2-w-rag-w-simple | Fine Tuned Models | 130 | 55.4 | 59.13 | 80.64 | 56.12 | 51.29 | 74.66 | 10.54 | LlamaForCausalLM |
| speechless-orca-platypus-coig-lite-2k-0.6e-13b | Fine Tuned Models | 130 | 55.4 | 59.9 | 80.76 | 58.34 | 47.97 | 77.9 | 7.51 | LlamaForCausalLM |
| yuren-13b-chatml | Fine Tuned Models | 130 | 55.39 | 53.07 | 78.03 | 56.34 | 42.32 | 74.43 | 28.13 | LlamaForCausalLM |
| tora-13b-v1.0 | Fine Tuned Models | 130 | 55.37 | 58.96 | 82.31 | 54.59 | 40.22 | 75.37 | 20.77 | LlamaForCausalLM |
| minotaur-llama2-13b-qlora | Chat Models | 130 | 55.37 | 60.07 | 82.42 | 55.87 | 45.57 | 76.24 | 12.05 | Unknown |
| Tinybra_13B | Fine Tuned Models | 130.2 | 55.36 | 55.72 | 80.99 | 54.37 | 49.14 | 73.8 | 18.12 | LlamaForCausalLM |
| Luban-Platypus2-13B-QLora-0.80-epoch | Chat Models | 130.2 | 55.34 | 60.24 | 82.22 | 58.03 | 55.26 | 75.37 | 0.91 | Unknown |
| SthenoWriter-L2-13B | Fine Tuned Models | 130.2 | 55.33 | 62.29 | 83.28 | 56.14 | 44.72 | 74.35 | 11.22 | LlamaForCausalLM |
| 2x-LoRA-Assemble-Platypus2-13B | Chat Models | 130.2 | 55.33 | 60.58 | 82.56 | 58.25 | 54.77 | 74.9 | 0.91 | Unknown |
| mistral-se-inst-ppo | Fine Tuned Models | 72.4 | 55.3 | 56.31 | 79.49 | 60.91 | 51.34 | 78.14 | 5.61 | Unknown |
| Xwin-LM-13B-V0.1 | Fine Tuned Models | 130 | 55.29 | 62.54 | 82.8 | 56.53 | 45.96 | 74.27 | 9.63 | LlamaForCausalLM |
| CodeLlama-34b-hf | Pretrained Models | 337.4 | 55.28 | 54.18 | 75.82 | 54.92 | 39.11 | 73.32 | 34.34 | LlamaForCausalLM |
| llama-2-13b-OpenOrca_20w | Fine Tuned Models | 130 | 55.28 | 59.9 | 82.51 | 56.3 | 43.14 | 77.19 | 12.66 | LlamaForCausalLM |
| chronos-13b-v2 | Fine Tuned Models | 130 | 55.25 | 58.7 | 82.52 | 53.39 | 50.55 | 75.06 | 11.3 | LlamaForCausalLM |
| SOLAR-Platypus-10.7B-v2 | Chat Models | 107.3 | 55.25 | 59.39 | 83.57 | 59.93 | 43.15 | 81.45 | 4.02 | LlamaForCausalLM |
| CreativityEngine | Fine Tuned Models | 0 | 55.25 | 59.3 | 82.42 | 53.55 | 52.46 | 74.19 | 9.55 | LlamaForCausalLM |
| OpenHermes-13B | Fine Tuned Models | 130 | 55.24 | 59.81 | 82.24 | 56.35 | 46.01 | 75.45 | 11.6 | LlamaForCausalLM |
| llama2-13b-Chinese-chat | Fine Tuned Models | 130 | 55.22 | 60.58 | 82.19 | 55.45 | 45.11 | 76.64 | 11.37 | Unknown |
| OrcaMini-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 55.22 | 60.84 | 82.56 | 56.42 | 53.32 | 75.93 | 2.27 | Unknown |
| airoboros-l2-13b-3.0 | Fine Tuned Models | 130.2 | 55.21 | 59.81 | 83.71 | 54.86 | 47.79 | 76.16 | 8.95 | LlamaForCausalLM |
| Mythical-Destroyer-V2-L2-13B | Fine Tuned Models | 130.2 | 55.2 | 59.3 | 82.66 | 57.39 | 57.09 | 74.74 | 0.0 | LlamaForCausalLM |
| minotaur-13b-fixed | Fine Tuned Models | 130 | 55.19 | 59.04 | 81.66 | 50.1 | 50.36 | 76.87 | 13.12 | LlamaForCausalLM |
| Dionysus-Mistral-n1-v1 | Fine Tuned Models | 72.4 | 55.18 | 60.24 | 81.6 | 59.32 | 47.94 | 71.35 | 10.61 | Unknown |
| zephyr_7b_norobots | Chat Models | 70 | 55.16 | 56.48 | 79.64 | 55.52 | 44.6 | 74.11 | 20.62 | Unknown |
| airoboros-c34b-2.2.1 | Chat Models | 340 | 55.15 | 54.69 | 76.84 | 55.43 | 51.36 | 72.53 | 20.02 | LlamaForCausalLM |
| Llama-2-13B-Instruct-v0.2 | Chat Models | 130 | 55.14 | 60.58 | 81.96 | 55.46 | 45.71 | 77.82 | 9.33 | ? |
| WizardLM-1.0-Uncensored-Llama2-13b | Fine Tuned Models | 128.5 | 55.14 | 55.72 | 80.34 | 55.4 | 51.44 | 74.66 | 13.27 | Unknown |
| athene-noctua-13b | Fine Tuned Models | 130.2 | 55.13 | 57.17 | 81.52 | 55.91 | 47.49 | 73.4 | 15.31 | LlamaForCausalLM |
| 13B-Legerdemain-L2 | Fine Tuned Models | 130 | 55.13 | 61.26 | 83.26 | 56.0 | 41.99 | 75.22 | 13.04 | LlamaForCausalLM |
| pygmalion-2-13b | Fine Tuned Models | 130.2 | 55.12 | 60.32 | 82.37 | 56.02 | 42.22 | 78.06 | 11.75 | LlamaForCausalLM |
| PuddleJumper-13b | Fine Tuned Models | 130 | 55.11 | 58.7 | 81.18 | 58.25 | 56.44 | 72.77 | 3.34 | LlamaForCausalLM |
| WizardLM-1.0-Uncensored-Llama2-13b | Fine Tuned Models | 128.5 | 55.1 | 55.8 | 80.41 | 55.59 | 51.42 | 74.11 | 13.27 | Unknown |
| llama2-13b-orca-8k-3319 | Fine Tuned Models | 130 | 55.09 | 60.75 | 81.91 | 57.06 | 42.64 | 77.19 | 10.99 | LlamaForCausalLM |
| Llama2-Chinese-13b-Chat | Fine Tuned Models | 130 | 55.07 | 55.97 | 82.05 | 54.74 | 48.9 | 76.16 | 12.59 | LlamaForCausalLM |
| llama-2-13b-dolphin_20w | Fine Tuned Models | 130 | 55.06 | 59.56 | 82.55 | 55.89 | 42.67 | 77.27 | 12.43 | LlamaForCausalLM |
| Python-Code-33B | Fine Tuned Models | 330 | 55.06 | 56.31 | 81.01 | 54.22 | 44.39 | 75.22 | 19.18 | LlamaForCausalLM |
| shisa-7b-v1 | Fine Tuned Models | 79.6 | 55.01 | 56.14 | 78.63 | 23.12 | 52.49 | 78.06 | 41.62 | MistralForCausalLM |
| dulia-13b-8k-alpha | Fine Tuned Models | 130.2 | 55.0 | 60.67 | 82.0 | 56.87 | 42.59 | 77.19 | 10.69 | LlamaForCausalLM |