加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
Data source: HuggingFace
| Model | Type | Parameters (B) | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | Architecture |
|---|---|---|---|---|---|---|---|---|---|---|
| Mistralic-7B-1 | Fine Tuned Models | 71.1 | 57.4 | 60.84 | 82.29 | 60.8 | 52.38 | 77.03 | 11.07 | Unknown |
| digital-socrates-13b | Fine Tuned Models | 130 | 57.34 | 58.36 | 80.14 |
Data is for reference only. Official sources are authoritative. Click model names to view DataLearner model profiles.
| 57.01 |
| 44.47 |
| 74.59 |
| 29.49 |
| LlamaForCausalLM |
| VicUnlocked-30B-LoRA-HF | Unkown Model Types | 300 | 57.33 | 59.73 | 84.02 | 57.81 | 48.54 | 79.48 | 14.4 | LlamaForCausalLM |
| airoboros-33b-gpt4 | Unkown Model Types | 330 | 57.32 | 63.74 | 84.87 | 58.54 | 47.06 | 77.03 | 12.66 | LlamaForCausalLM |
| BrainDerp2 | Fine Tuned Models | 130.2 | 57.32 | 60.92 | 81.94 | 58.9 | 57.19 | 75.93 | 9.02 | LlamaForCausalLM |
| OpenOrca-Platypus2-13B-QLoRA-0.80-epoch | Chat Models | 130.2 | 57.31 | 62.37 | 82.99 | 59.38 | 52.2 | 75.77 | 11.14 | Unknown |
| LLaMA_2_13B_SFT_v0 | Fine Tuned Models | 130 | 57.31 | 62.03 | 83.8 | 58.39 | 49.92 | 77.27 | 12.43 | LlamaForCausalLM |
| CodeLlama-34b-Instruct-hf | Fine Tuned Models | 337.4 | 57.29 | 54.27 | 76.92 | 55.54 | 44.44 | 74.59 | 37.98 | LlamaForCausalLM |
| OpenOrca-Platypus2-13B | Fine Tuned Models | 130 | 57.28 | 62.8 | 83.15 | 59.39 | 53.08 | 76.24 | 9.02 | LlamaForCausalLM |
| 2x-LoRA-Assemble-Nova-13B | Chat Models | 130.2 | 57.26 | 62.63 | 83.24 | 58.64 | 51.88 | 76.95 | 10.24 | Unknown |
| MLewd-ReMM-L2-Chat-20B-Inverted | Fine Tuned Models | 199.9 | 57.25 | 61.69 | 85.32 | 58.0 | 53.77 | 75.61 | 9.1 | LlamaForCausalLM |
| Giraffe-13b-32k-v3 | Chat Models | 130.2 | 57.24 | 59.04 | 79.59 | 55.01 | 46.68 | 76.95 | 26.16 | LlamaForCausalLM |
| orca_mini_v3_13b | Fine Tuned Models | 128.5 | 57.24 | 63.14 | 82.35 | 56.52 | 51.81 | 76.48 | 13.12 | Unknown |
| orca_mini_v3_13b | Chat Models | 130 | 57.24 | 63.14 | 82.35 | 56.52 | 51.81 | 76.48 | 13.12 | LlamaForCausalLM |
| MLewd-Chat-v2-13B | Fine Tuned Models | 130 | 57.23 | 61.86 | 83.81 | 57.0 | 54.51 | 75.77 | 10.46 | LlamaForCausalLM |
| Athena-v4 | Fine Tuned Models | 130.2 | 57.23 | 62.54 | 84.19 | 57.33 | 50.87 | 76.48 | 11.98 | LlamaForCausalLM |
| airoboros-33b-2.1 | Chat Models | 330 | 57.16 | 63.65 | 84.97 | 57.37 | 52.17 | 78.22 | 6.6 | LlamaForCausalLM |
| airoboros-33b-gpt4-m2.0 | Fine Tuned Models | 330 | 57.16 | 64.68 | 84.95 | 57.77 | 47.44 | 77.74 | 10.39 | LlamaForCausalLM |
| BrainDerp3 | Fine Tuned Models | 130.2 | 57.13 | 60.92 | 82.1 | 58.91 | 57.18 | 75.61 | 8.04 | LlamaForCausalLM |
| BrainDerp | Fine Tuned Models | 130.2 | 57.11 | 60.75 | 82.1 | 58.81 | 56.9 | 75.85 | 8.26 | LlamaForCausalLM |
| SOLID-SFT-WoDPO-MixQV2-Zephyr-7b-beta | Fine Tuned Models | 72.4 | 57.1 | 59.73 | 81.72 | 60.47 | 53.77 | 74.66 | 12.28 | MistralForCausalLM |
| ReMM-v2.2-L2-13B | Fine Tuned Models | 130.2 | 57.1 | 61.26 | 84.16 | 56.22 | 51.35 | 75.61 | 14.03 | LlamaForCausalLM |
| Athena-v3 | Fine Tuned Models | 0 | 57.09 | 61.69 | 84.34 | 57.87 | 51.26 | 75.77 | 11.6 | LlamaForCausalLM |
| vicuna-33b-coder | Fine Tuned Models | 330 | 57.07 | 60.41 | 83.27 | 57.17 | 51.79 | 76.87 | 12.89 | LlamaForCausalLM |
| Emerhyst-20B | Fine Tuned Models | 199.9 | 57.07 | 61.69 | 84.98 | 56.98 | 54.16 | 76.09 | 8.49 | LlamaForCausalLM |
| Llama-2-13b-orca-v1 | Fine Tuned Models | 130 | 57.05 | 62.03 | 82.27 | 57.71 | 49.61 | 76.87 | 13.8 | LlamaForCausalLM |
| airoboros-33b-gpt4-m2.0 | Fine Tuned Models | 330 | 57.03 | 63.4 | 85.19 | 57.46 | 48.15 | 78.37 | 9.63 | LlamaForCausalLM |
| zephyr-beta-Nebula-v2-7B | Fine Tuned Models | 72.4 | 57.03 | 56.57 | 82.53 | 56.4 | 58.68 | 70.48 | 17.51 | MistralForCausalLM |
| airoboros-33b-gpt4-2.0 | Fine Tuned Models | 330 | 57.02 | 63.91 | 85.67 | 57.95 | 45.54 | 77.98 | 11.07 | LlamaForCausalLM |
| wendigo-14b-alpha2 | Fine Tuned Models | 142.2 | 57.02 | 56.66 | 77.19 | 58.0 | 53.71 | 73.64 | 22.9 | Unknown |
| synapsellm-7b-mistral-v0.3-preview | Fine Tuned Models | 72.4 | 57.01 | 53.84 | 74.86 | 54.81 | 55.03 | 74.59 | 28.96 | MistralForCausalLM |
| wendigo-14b-alpha1 | Fine Tuned Models | 142.2 | 57.01 | 56.48 | 77.2 | 57.83 | 53.76 | 73.01 | 23.81 | Unknown |
| airoboros-33b-gpt4-2.0 | Fine Tuned Models | 330 | 57.01 | 63.82 | 85.65 | 58.44 | 45.57 | 77.9 | 10.69 | LlamaForCausalLM |
| YuLan-Chat-2-13b-fp16 | Chat Models | 130 | 57.01 | 59.04 | 80.66 | 56.72 | 52.18 | 79.64 | 13.8 | LlamaForCausalLM |
| ReMM-v2-L2-13B | Fine Tuned Models | 130.2 | 56.99 | 61.95 | 84.0 | 56.14 | 50.81 | 75.85 | 13.19 | LlamaForCausalLM |
| AISquare-Instruct-llama2-koen-13b-v0.9.24 | Chat Models | 131.6 | 56.98 | 55.63 | 81.35 | 51.76 | 53.0 | 76.95 | 23.2 | LlamaForCausalLM |
| OpenOrca-Platypus2-13B-GPTQ | Fine Tuned Models | 162.4 | 56.98 | 62.54 | 82.67 | 58.56 | 51.93 | 76.8 | 9.4 | LlamaForCausalLM |
| GenAI-Nova-13B | Chat Models | 130.2 | 56.98 | 62.29 | 83.27 | 59.47 | 51.79 | 77.35 | 7.73 | Unknown |
| airoboros-33b-gpt4-m2.0 | Fine Tuned Models | 330 | 56.97 | 63.14 | 85.19 | 57.28 | 48.07 | 78.45 | 9.7 | LlamaForCausalLM |
| llama-30b | Unkown Model Types | 325.3 | 56.96 | 61.43 | 84.73 | 58.45 | 42.27 | 80.03 | 14.86 | LlamaForCausalLM |
| llama-30B-hf-openassitant | Fine Tuned Models | 300 | 56.94 | 61.26 | 84.73 | 58.47 | 42.27 | 80.03 | 14.86 | LlamaForCausalLM |
| CantoneseLLM-6B-preview202402 | Pretrained Models | 60.6 | 56.93 | 55.63 | 75.8 | 63.07 | 42.26 | 74.11 | 30.71 | LlamaForCausalLM |
| UndiMix-v4-13B | Fine Tuned Models | 130.2 | 56.93 | 61.95 | 83.88 | 56.9 | 48.96 | 76.16 | 13.72 | LlamaForCausalLM |
| LosslessMegaCoder-llama2-13b-mini | Chat Models | 130 | 56.92 | 60.58 | 81.26 | 57.92 | 48.89 | 76.95 | 15.92 | LlamaForCausalLM |
| Llama-2-13b-orca-v1 | Fine Tuned Models | 130 | 56.91 | 62.2 | 82.32 | 57.67 | 49.6 | 76.8 | 12.89 | LlamaForCausalLM |
| Emerald-13B | Fine Tuned Models | 130 | 56.89 | 62.29 | 83.69 | 55.7 | 50.94 | 75.93 | 12.81 | LlamaForCausalLM |
| ghost-7b-v0.9.0 | Chat Models | 72.4 | 56.89 | 53.07 | 77.93 | 55.09 | 47.79 | 73.72 | 33.74 | MistralForCausalLM |
| ReMM-Mistral-13B | Fine Tuned Models | 128.5 | 56.89 | 62.2 | 83.82 | 55.43 | 53.32 | 74.51 | 12.05 | Unknown |
| Instruct_Yi-6B_Dolly15K | Fine Tuned Models | 60.6 | 56.85 | 54.86 | 75.87 | 63.37 | 42.84 | 74.9 | 29.26 | LlamaForCausalLM |
| OpenOrcaxOpenChat-Preview2-13B-GPTQ | Fine Tuned Models | 162.4 | 56.84 | 61.26 | 82.14 | 57.85 | 50.22 | 77.11 | 12.43 | LlamaForCausalLM |
| Barcenas-Orca-2-7b | Fine Tuned Models | 67.4 | 56.81 | 55.2 | 77.08 | 56.02 | 43.72 | 75.53 | 33.28 | LlamaForCausalLM |
| SOLID_SFT-WoDPO-WoMixQ | Fine Tuned Models | 72.4 | 56.8 | 59.64 | 81.69 | 60.1 | 55.25 | 74.66 | 9.48 | MistralForCausalLM |
| TIGERScore-13B | Fine Tuned Models | 130.2 | 56.79 | 59.04 | 82.79 | 55.07 | 40.38 | 74.74 | 28.73 | LlamaForCausalLM |
| Orca-2-13b-SFT_v5 | Fine Tuned Models | 130.2 | 56.77 | 59.22 | 80.09 | 60.19 | 51.84 | 80.9 | 8.42 | LlamaForCausalLM |
| Yi-6B-200K | Pretrained Models | 60.6 | 56.76 | 53.75 | 75.57 | 64.65 | 41.56 | 73.64 | 31.39 | LlamaForCausalLM |
| Orca-Nova-13B | Chat Models | 130.2 | 56.72 | 62.37 | 82.47 | 57.44 | 45.97 | 77.58 | 14.48 | Unknown |
| ReMM-v2.1-L2-13B | Fine Tuned Models | 130.2 | 56.71 | 61.43 | 83.92 | 55.95 | 50.3 | 75.93 | 12.74 | LlamaForCausalLM |
| OpenOrcaxOpenChat-Preview2-13B | Fine Tuned Models | 130 | 56.7 | 62.71 | 81.99 | 57.51 | 47.45 | 76.8 | 13.72 | LlamaForCausalLM |
| Yi-6B-200K | Pretrained Models | 60.6 | 56.69 | 53.58 | 75.58 | 64.65 | 41.74 | 74.27 | 30.33 | LlamaForCausalLM |
| test3_sft_4bit | Fine Tuned Models | 72.4 | 56.66 | 61.52 | 83.89 | 64.79 | 47.83 | 81.93 | 0.0 | MistralForCausalLM |
| storytime-13b | Chat Models | 130.2 | 56.64 | 62.03 | 83.96 | 57.48 | 52.5 | 75.53 | 8.34 | LlamaForCausalLM |
| Amethyst-13B | Fine Tuned Models | 130 | 56.62 | 62.63 | 83.17 | 55.91 | 52.43 | 74.74 | 10.84 | LlamaForCausalLM |
| Amethyst-13B-Mistral | Fine Tuned Models | 128.5 | 56.62 | 62.63 | 83.17 | 55.91 | 52.43 | 74.74 | 10.84 | Unknown |
| BELLE-Llama2-13B-chat-0.4M | Fine Tuned Models | 130 | 56.62 | 60.67 | 82.31 | 55.94 | 50.85 | 75.53 | 14.4 | LlamaForCausalLM |
| Clover3-17B | Fine Tuned Models | 168.4 | 56.61 | 59.9 | 81.18 | 60.47 | 40.72 | 78.61 | 18.8 | MistralForCausalLM |
| chronos-33b | Fine Tuned Models | 330 | 56.59 | 62.2 | 83.48 | 55.87 | 46.67 | 78.3 | 13.04 | LlamaForCausalLM |
| LlongOrca-13B-16k | Fine Tuned Models | 130 | 56.59 | 62.46 | 82.75 | 55.54 | 50.11 | 76.4 | 12.28 | LlamaForCausalLM |
| llama2-13b-megacode2-oasst | Fine Tuned Models | 130 | 56.59 | 60.67 | 81.93 | 57.38 | 47.85 | 76.16 | 15.54 | LlamaForCausalLM |
| Novocode7b-v2 | Fine Tuned Models | 72.4 | 56.57 | 61.01 | 84.12 | 64.05 | 42.21 | 79.87 | 8.19 | MistralForCausalLM |
| NyakuraV2.1-m7 | Fine Tuned Models | 0 | 56.57 | 58.62 | 81.89 | 58.46 | 45.01 | 72.77 | 22.67 | MistralForCausalLM |
| OpenRP-13B | Fine Tuned Models | 130 | 56.57 | 62.12 | 82.6 | 57.5 | 48.29 | 76.01 | 12.89 | LlamaForCausalLM |
| MM-ReMM-L2-20B | Fine Tuned Models | 199.9 | 56.55 | 60.84 | 85.18 | 56.45 | 53.33 | 75.77 | 7.73 | LlamaForCausalLM |
| BerrySauce-L2-13b | Fine Tuned Models | 130.2 | 56.55 | 62.29 | 83.78 | 57.1 | 48.3 | 76.09 | 11.75 | LlamaForCausalLM |
| MLewdBoros-L2-13B | Fine Tuned Models | 130.2 | 56.51 | 62.54 | 83.9 | 56.57 | 48.14 | 76.95 | 10.99 | LlamaForCausalLM |
| EnsembleV5-Nova-13B | Chat Models | 130.2 | 56.49 | 62.71 | 82.55 | 56.79 | 49.86 | 76.24 | 10.77 | Unknown |
| EnsembleV5-Nova-13B | Chat Models | 130.2 | 56.49 | 62.71 | 82.55 | 56.79 | 49.86 | 76.24 | 10.77 | Unknown |
| mythalion-13b | Fine Tuned Models | 130.2 | 56.48 | 61.26 | 83.81 | 56.53 | 46.56 | 77.43 | 13.27 | LlamaForCausalLM |
| SOLAR_KO_1.3_deup | Chat Models | 108.5 | 56.47 | 55.97 | 79.97 | 55.88 | 47.55 | 76.87 | 22.59 | LlamaForCausalLM |
| speechless-code-mistral-7b-v2.0 | Fine Tuned Models | 70 | 56.47 | 52.47 | 75.61 | 51.31 | 52.05 | 71.43 | 35.94 | MistralForCausalLM |
| WizardLM-30B-Uncensored | Fine Tuned Models | 323.2 | 56.46 | 60.24 | 82.93 | 56.8 | 51.57 | 74.35 | 12.89 | Unknown |
| SciPhi-Self-RAG-Mistral-7B-32k | Fine Tuned Models | 70 | 56.46 | 57.34 | 80.44 | 60.81 | 45.63 | 74.82 | 19.71 | MistralForCausalLM |
| Pygmalion-2-13b-SuperCOT | Fine Tuned Models | 130.2 | 56.46 | 63.23 | 83.68 | 54.9 | 53.14 | 77.51 | 6.29 | LlamaForCausalLM |
| Stheno-Inverted-L2-13B | Fine Tuned Models | 130 | 56.44 | 59.3 | 82.9 | 56.45 | 52.04 | 74.74 | 13.19 | LlamaForCausalLM |
| Nova-13B | Chat Models | 130.2 | 56.44 | 62.71 | 82.57 | 57.98 | 51.34 | 77.27 | 6.75 | Unknown |
| Stheno-L2-13B | Fine Tuned Models | 130 | 56.43 | 61.01 | 83.95 | 56.33 | 50.18 | 75.14 | 11.98 | LlamaForCausalLM |
| Mythical-Destroyer-L2-13B | Fine Tuned Models | 130.2 | 56.39 | 58.7 | 82.0 | 57.66 | 56.35 | 74.66 | 8.95 | LlamaForCausalLM |
| llama2_7b_merge_orcafamily | Fine Tuned Models | 67.4 | 56.38 | 56.91 | 81.17 | 51.49 | 49.68 | 75.93 | 23.12 | LlamaForCausalLM |
| Dans-AdventurousWinds-Mk2-7b | Fine Tuned Models | 72.4 | 56.38 | 58.19 | 83.48 | 61.8 | 43.56 | 76.32 | 14.94 | MistralForCausalLM |
| MLewd-v2.4-13B | Fine Tuned Models | 130 | 56.37 | 61.69 | 83.83 | 55.1 | 53.34 | 74.51 | 9.78 | LlamaForCausalLM |
| speechless-code-mistral-7b-v2.0 | Fine Tuned Models | 70 | 56.37 | 52.3 | 75.61 | 51.28 | 52.05 | 71.35 | 35.63 | MistralForCausalLM |
| airoboros-l2-13b-2.2.1 | Chat Models | 130 | 56.36 | 60.92 | 83.77 | 56.47 | 49.42 | 76.01 | 11.6 | LlamaForCausalLM |
| openchat_v3.1 | Fine Tuned Models | 0 | 56.36 | 59.81 | 82.8 | 56.76 | 44.45 | 76.24 | 18.12 | LlamaForCausalLM |
| internlm2-math-20b-llama | Fine Tuned Models | 198.6 | 56.35 | 59.98 | 81.64 | 65.07 | 52.9 | 76.4 | 2.12 | LlamaForCausalLM |
| LLAMA-13B-test-finetuning | Fine Tuned Models | 130.2 | 56.34 | 58.02 | 82.36 | 54.27 | 44.14 | 76.72 | 22.52 | Unknown |
| MythoMix-L2-13b | Fine Tuned Models | 130 | 56.31 | 61.09 | 83.86 | 55.42 | 52.08 | 75.45 | 9.93 | LlamaForCausalLM |
| mistral-7b_open_platypus | Chat Models | 70 | 56.29 | 55.8 | 82.13 | 59.76 | 48.87 | 78.61 | 12.59 | MistralForCausalLM |
| Uncensored-Jordan-13B | Fine Tuned Models | 130 | 56.27 | 57.42 | 82.7 | 55.75 | 50.51 | 76.16 | 15.09 | LlamaForCausalLM |
| speechless-code-mistral-orca-7b-v1.0 | Fine Tuned Models | 70 | 56.24 | 59.64 | 82.25 | 61.33 | 48.45 | 77.51 | 8.26 | MistralForCausalLM |
| MythoLogic-L2-13b | Fine Tuned Models | 130 | 56.19 | 61.01 | 83.93 | 55.7 | 48.64 | 76.09 | 11.75 | LlamaForCausalLM |
| Synatra-11B-Testbench | Chat Models | 110 | 56.17 | 57.34 | 78.66 | 55.56 | 51.97 | 75.77 | 17.74 | Unknown |