加载中...
加载中...
Open LLM Leaderboard是追踪大模型评测结果的排行榜,通过追踪大语言模型和ChatBot在不同评测任务上的表现来对模型进行排名和评估。
Data source: HuggingFace
| Model | Type | Parameters (B) | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K | Architecture |
|---|---|---|---|---|---|---|---|---|---|---|
| Qwen-72B | Pretrained Models | 722.9 | 73.6 | 65.19 | 85.94 | 77.37 | 60.19 | 82.48 | 70.43 | QWenLMHeadModel |
| Yi-34B-Llama | Pretrained Models | 343.9 | 70.95 |
Data is for reference only. Official sources are authoritative. Click model names to view DataLearner model profiles.
| 64.59 |
| 85.63 |
| 76.31 |
| 55.6 |
| 82.79 |
| 60.8 |
| LlamaForCausalLM |
| Yi-34B-200K | Pretrained Models | 343.9 | 70.81 | 65.36 | 85.58 | 76.06 | 53.64 | 82.56 | 61.64 | LlamaForCausalLM |
| internlm2-20b-llama | Pretrained Models | 198.6 | 70.66 | 64.59 | 83.12 | 67.27 | 54.13 | 84.21 | 70.66 | LlamaForCausalLM |
| internlm2-20b-llama | Pretrained Models | 198.6 | 70.61 | 64.68 | 83.16 | 67.17 | 54.17 | 84.29 | 70.2 | L;l;a;m;a;F;o;r;C;a;u;s;a;l;L;M |
| internlm2-20b | Pretrained Models | 200 | 69.75 | 62.97 | 83.21 | 67.58 | 51.27 | 85.56 | 67.93 | Unknown |
| Qwen-72B-Llama | Pretrained Models | 722.9 | 69.53 | 64.85 | 83.27 | 73.66 | 57.6 | 81.53 | 56.25 | LlamaForCausalLM |
| Yi-34B | Pretrained Models | 343.9 | 69.42 | 64.59 | 85.69 | 76.35 | 56.23 | 83.03 | 50.64 | LlamaForCausalLM |
| deepseek-llm-67b-base | Pretrained Models | 670 | 69.38 | 65.44 | 87.1 | 71.78 | 51.08 | 84.14 | 56.71 | LlamaForCausalLM |
| Mixtral-8x7B-v0.1 | Pretrained Models | 467 | 68.42 | 66.04 | 86.49 | 71.82 | 46.78 | 81.93 | 57.47 | MixtralForCausalLM |
| falcon-180B | Pretrained Models | 1795.2 | 67.85 | 69.45 | 88.86 | 70.5 | 45.47 | 86.9 | 45.94 | FalconForCausalLM |
| internlm2-7b-llama | Pretrained Models | 77.4 | 66.94 | 60.49 | 80.99 | 63.16 | 54.25 | 79.87 | 62.85 | L;l;a;m;a;F;o;r;C;a;u;s;a;l;L;M |
| Chinese-Mixtral-8x7B | Pretrained Models | 469.1 | 66.69 | 63.57 | 85.98 | 70.95 | 45.86 | 82.08 | 51.71 | MixtralForCausalLM |
| internlm2-7b | Pretrained Models | 70 | 66.68 | 58.02 | 81.24 | 65.24 | 48.73 | 83.82 | 63.0 | Unknown |
| SOLAR-10.7B-v1.0 | Pretrained Models | 107.3 | 66.04 | 61.95 | 84.6 | 65.48 | 45.04 | 83.66 | 55.5 | LlamaForCausalLM |
| falcon-180B | Pretrained Models | 1795.2 | 65.46 | 69.2 | 88.89 | 69.59 | 45.16 | 86.74 | 33.21 | FalconForCausalLM |
| internlm-20b-llama | Pretrained Models | 200 | 65.09 | 61.35 | 82.08 | 61.59 | 57.71 | 76.72 | 51.1 | LlamaForCausalLM |
| KoSOLAR-10.7B-v0.3 | Pretrained Models | 108 | 64.76 | 62.8 | 83.73 | 64.51 | 44.57 | 82.48 | 50.49 | LlamaForCausalLM |
| KoSOLAR-10.7B-v0.2 | Pretrained Models | 107 | 64.2 | 61.35 | 82.63 | 64.85 | 47.94 | 80.74 | 47.69 | LlamaForCausalLM |
| tigerbot-70b-base | Pretrained Models | 689.5 | 63.71 | 62.46 | 83.61 | 65.49 | 52.76 | 80.19 | 37.76 | Unknown |
| Qwen-14B-Llamafied | Pretrained Models | 140 | 63.09 | 55.2 | 82.31 | 66.11 | 45.6 | 76.56 | 52.77 | LlamaForCausalLM |
| llama-65b | Pretrained Models | 652.9 | 62.79 | 63.48 | 86.09 | 63.93 | 43.43 | 82.56 | 37.23 | LlamaForCausalLM |
| internlm2-base-20b-llama | Pretrained Models | 198.6 | 62.69 | 63.05 | 82.11 | 63.97 | 43.97 | 78.22 | 44.81 | LlamaForCausalLM |
| internlm2-base-20b-llama | Pretrained Models | 198.6 | 62.69 | 62.97 | 82.15 | 63.78 | 44.11 | 78.22 | 44.88 | LlamaForCausalLM |
| DeciLM-7B | Pretrained Models | 70.4 | 61.55 | 59.39 | 82.51 | 59.76 | 40.33 | 79.95 | 47.38 | DeciLMForCausalLM |
| phi-2 | Pretrained Models | 27.8 | 61.33 | 61.09 | 75.11 | 58.11 | 44.47 | 74.35 | 54.81 | PhiForCausalLM |
| Mistral-7B-v0.1 | Pretrained Models | 72.4 | 60.97 | 59.98 | 83.31 | 64.16 | 42.15 | 78.37 | 37.83 | MistralForCausalLM |
| Mistral-7B-v0.1 | Pretrained Models | 72.4 | 60.97 | 59.98 | 83.31 | 64.16 | 42.15 | 78.37 | 37.83 | MistralForCausalLM |
| mistral-sft-v3 | Pretrained Models | 72.4 | 60.93 | 61.35 | 82.23 | 63.4 | 48.49 | 77.66 | 32.45 | MistralForCausalLM |
| Nanbeige-16B-Base-Llama | Pretrained Models | 158.3 | 60.7 | 56.48 | 78.97 | 63.34 | 42.6 | 75.77 | 47.01 | LlamaForCausalLM |
| CodeLlama-70b-Instruct-hf | Pretrained Models | 689.8 | 59.98 | 55.03 | 77.24 | 56.4 | 50.44 | 74.51 | 46.25 | LlamaForCausalLM |
| internlm-20b | Pretrained Models | 200 | 59.55 | 60.49 | 82.13 | 61.85 | 52.61 | 76.72 | 23.5 | InternLMForCausalLM |
| Qwen-7B | Pretrained Models | 77.2 | 59.19 | 51.37 | 78.47 | 59.84 | 47.79 | 72.69 | 44.96 | QWenLMHeadModel |
| CodeLlama-70b-hf | Pretrained Models | 689.8 | 58.93 | 56.74 | 78.21 | 59.67 | 39.79 | 75.22 | 43.97 | LlamaForCausalLM |
| CodeLlama-70b-hf | Pretrained Models | 689.8 | 58.93 | 56.74 | 78.21 | 59.67 | 39.79 | 75.22 | 43.97 | LlamaForCausalLM |
| CodeLlama-70b-hf | Pretrained Models | 689.8 | 58.93 | 56.74 | 78.21 | 59.67 | 39.79 | 75.22 | 43.97 | LlamaForCausalLM |
| chinese-mixtral | Pretrained Models | 467 | 58.57 | 67.49 | 85.25 | 70.31 | 46.75 | 81.61 | 0.0 | MixtralForCausalLM |
| falcon-40b | Pretrained Models | 400 | 58.07 | 61.86 | 85.28 | 56.89 | 41.65 | 81.29 | 21.46 | FalconForCausalLM |
| typhoon-7b | Pretrained Models | 70 | 58.05 | 58.53 | 81.55 | 59.54 | 40.52 | 76.56 | 31.61 | MistralForCausalLM |
| CodeLlama-70b-Python-hf | Pretrained Models | 689.8 | 58.0 | 55.12 | 78.48 | 56.17 | 41.78 | 73.01 | 43.44 | LlamaForCausalLM |
| CantoneseLLM-6B-preview202402 | Pretrained Models | 60.6 | 56.93 | 55.63 | 75.8 | 63.07 | 42.26 | 74.11 | 30.71 | LlamaForCausalLM |
| Yi-6B-200K | Pretrained Models | 60.6 | 56.76 | 53.75 | 75.57 | 64.65 | 41.56 | 73.64 | 31.39 | LlamaForCausalLM |
| Yi-6B-200K | Pretrained Models | 60.6 | 56.69 | 53.58 | 75.58 | 64.65 | 41.74 | 74.27 | 30.33 | LlamaForCausalLM |
| CodeLlama-34b-hf | Pretrained Models | 337.4 | 55.28 | 54.18 | 75.82 | 54.92 | 39.11 | 73.32 | 34.34 | LlamaForCausalLM |
| Aquila2-34B | Pretrained Models | 340 | 54.5 | 52.65 | 81.99 | 76.02 | 40.8 | 75.06 | 0.45 | LlamaForCausalLM |
| Yi-6B | Pretrained Models | 60.6 | 54.08 | 55.55 | 76.57 | 64.11 | 41.96 | 74.19 | 12.13 | LlamaForCausalLM |
| Yi-6B | Pretrained Models | 60.6 | 54.02 | 55.55 | 76.42 | 63.85 | 41.86 | 73.8 | 12.66 | LlamaForCausalLM |
| internlm2-base-7b-llama | Pretrained Models | 77.4 | 53.62 | 54.35 | 79.47 | 54.05 | 43.23 | 71.43 | 19.18 | LlamaForCausalLM |
| tigerbot-13b-base | Pretrained Models | 130 | 53.42 | 53.84 | 77.05 | 53.57 | 44.06 | 74.98 | 17.06 | Unknown |
| shisa-base-7b-v1 | Pretrained Models | 79.6 | 51.64 | 52.3 | 77.63 | 23.12 | 42.4 | 78.53 | 35.86 | MistralForCausalLM |
| llama-13b | Pretrained Models | 130.2 | 51.33 | 56.14 | 80.92 | 47.61 | 39.48 | 76.24 | 7.58 | LlamaForCausalLM |
| FuseLLM-7B | Pretrained Models | 70 | 51.07 | 53.24 | 78.72 | 47.93 | 38.17 | 74.03 | 14.33 | LlamaForCausalLM |
| deepseek-moe-16b-base | Pretrained Models | 163.8 | 51.07 | 53.24 | 79.77 | 46.31 | 36.08 | 73.72 | 17.29 | Unknown |
| Llama-2-7B-GPTQ | Pretrained Models | 90.5 | 48.48 | 52.05 | 77.59 | 43.99 | 39.32 | 72.93 | 5.0 | LlamaForCausalLM |
| yayi2-30b-llama | Pretrained Models | 304 | 48.46 | 35.67 | 53.37 | 70.6 | 49.08 | 63.14 | 18.88 | LlamaForCausalLM |
| openllama-7b-icl | Pretrained Models | 70 | 47.93 | 47.95 | 77.04 | 44.37 | 37.06 | 70.17 | 10.99 | LlamaForCausalLM |
| phi-1_5 | Pretrained Models | 0 | 47.69 | 52.9 | 63.79 | 43.89 | 40.89 | 72.22 | 12.43 | PhiForCausalLM |
| phi-1_5 | Pretrained Models | 0 | 47.69 | 52.9 | 63.79 | 43.89 | 40.89 | 72.22 | 12.43 | PhiForCausalLM |
| open_llama_13b | Pretrained Models | 130 | 47.26 | 51.19 | 75.23 | 43.75 | 38.08 | 72.06 | 3.26 | LlamaForCausalLM |
| openllama-7b-base | Pretrained Models | 70 | 47.09 | 46.16 | 76.4 | 42.82 | 36.65 | 70.88 | 9.63 | LlamaForCausalLM |
| stablelm-3b-4e1t | Pretrained Models | 28 | 46.58 | 46.59 | 75.94 | 45.23 | 37.2 | 71.19 | 3.34 | StableLMEpochForCausalLM |
| stablelm-base-alpha-7b-v2 | Pretrained Models | 68.9 | 46.18 | 47.35 | 77.08 | 45.1 | 36.46 | 68.51 | 2.58 | StableLMAlphaForCausalLM |
| bloom | Pretrained Models | 1762.5 | 46.07 | 50.43 | 76.41 | 30.85 | 39.76 | 72.06 | 6.9 | BloomForCausalLM |
| llama-base-7b | Pretrained Models | 66.1 | 45.62 | 50.94 | 77.8 | 35.67 | 34.34 | 71.43 | 3.56 | Unknown |
| stablelm-2-1_6b | Pretrained Models | 16.4 | 45.25 | 43.34 | 70.45 | 38.95 | 36.78 | 64.56 | 17.44 | Unknown |
| Qwen-1_8B-Llamafied | Pretrained Models | 18.4 | 44.75 | 37.71 | 58.87 | 46.37 | 39.41 | 61.72 | 24.41 | LlamaForCausalLM |
| ThetaWave-14B-v0.1 | Pretrained Models | 142.2 | 44.54 | 42.83 | 47.09 | 61.45 | 50.41 | 65.43 | 0.0 | MistralForCausalLM |
| open_llama_7b_v2 | Pretrained Models | 70 | 44.26 | 43.69 | 72.2 | 41.29 | 35.54 | 69.38 | 3.49 | LlamaForCausalLM |
| falcon-7b | Pretrained Models | 70 | 44.17 | 47.87 | 78.13 | 27.79 | 34.26 | 72.38 | 4.62 | FalconForCausalLM |
| gpt-sw3-40b | Pretrained Models | 399.3 | 43.42 | 43.0 | 72.37 | 34.97 | 37.52 | 67.96 | 4.7 | GPT2LMHeadModel |
| CodeLlama-13b-hf | Pretrained Models | 130.2 | 43.35 | 40.87 | 63.35 | 32.81 | 43.79 | 67.17 | 12.13 | LlamaForCausalLM |
| quan-1.8b-base | Pretrained Models | 18 | 43.35 | 36.95 | 58.46 | 45.44 | 41.6 | 57.93 | 19.71 | LlamaForCausalLM |
| opt-66b | Pretrained Models | 660 | 42.78 | 46.33 | 76.25 | 26.99 | 35.43 | 70.01 | 1.67 | OPTForCausalLM |
| codegen-16B-nl | Pretrained Models | 160 | 42.59 | 46.76 | 71.87 | 32.35 | 33.95 | 67.96 | 2.65 | CodeGenForCausalLM |
| open_llama_7b | Pretrained Models | 70 | 42.31 | 47.01 | 71.98 | 30.49 | 34.85 | 67.96 | 1.59 | LlamaForCausalLM |
| palmyra-large | Pretrained Models | 0 | 42.09 | 44.97 | 71.85 | 28.54 | 35.93 | 67.88 | 3.41 | GPT2LMHeadModel |
| opt-30b | Pretrained Models | 300 | 42.0 | 43.26 | 74.07 | 26.66 | 35.16 | 70.64 | 2.2 | OPTForCausalLM |
| gpt-neox-20b | Pretrained Models | 207.4 | 41.69 | 45.73 | 73.45 | 25.0 | 31.61 | 68.9 | 5.46 | GPTNeoXForCausalLM |
| RedPajama-INCITE-7B-Base | Pretrained Models | 70 | 41.49 | 46.25 | 71.63 | 27.68 | 33.03 | 67.32 | 3.03 | GPTNeoXForCausalLM |
| MiniMA-3B | Pretrained Models | 30.2 | 41.44 | 43.43 | 68.06 | 28.69 | 39.76 | 65.98 | 2.73 | LlamaForCausalLM |
| nucleus-22B-token-500B | Pretrained Models | 218.3 | 41.33 | 40.7 | 69.39 | 30.11 | 39.16 | 67.64 | 0.99 | LlamaForCausalLM |
| RedPajama-INCITE-Base-7B-v0.1 | Pretrained Models | 66.5 | 41.25 | 46.25 | 71.63 | 27.68 | 33.03 | 67.32 | 1.59 | Unknown |
| Amber | Pretrained Models | 0 | 40.97 | 40.96 | 73.79 | 26.84 | 33.56 | 67.88 | 2.81 | LlamaForCausalLM |
| gpt-sw3-20b | Pretrained Models | 209.2 | 40.71 | 41.81 | 68.75 | 28.47 | 37.1 | 67.17 | 0.99 | GPT2LMHeadModel |
| ThetaWave-28B-v0.1 | Pretrained Models | 281.8 | 40.4 | 36.6 | 35.54 | 54.5 | 49.86 | 65.9 | 0.0 | MistralForCausalLM |
| open_llama_3b_v2 | Pretrained Models | 30 | 40.28 | 40.27 | 71.6 | 27.12 | 34.78 | 67.01 | 0.91 | LlamaForCausalLM |
| gpt-j-6b | Pretrained Models | 60 | 40.1 | 41.38 | 67.54 | 26.78 | 35.96 | 65.98 | 2.96 | GPTJForCausalLM |
| opt-13b | Pretrained Models | 130 | 40.06 | 39.93 | 71.2 | 24.9 | 34.1 | 68.51 | 1.74 | OPTForCausalLM |
| codegen-6B-nl | Pretrained Models | 60 | 40.0 | 42.32 | 68.59 | 25.93 | 34.47 | 66.46 | 2.2 | CodeGenForCausalLM |
| CodeLlama-7b-hf | Pretrained Models | 67.4 | 39.81 | 39.93 | 60.8 | 31.12 | 37.82 | 64.01 | 5.16 | LlamaForCausalLM |
| pythia-12b-deduped | Pretrained Models | 120 | 39.7 | 41.38 | 70.26 | 25.63 | 33.0 | 66.46 | 1.44 | GPTNeoXForCausalLM |
| gpt-sw3-6.7b-v2 | Pretrained Models | 71.1 | 39.49 | 39.42 | 66.39 | 30.09 | 35.6 | 64.25 | 1.21 | GPT2LMHeadModel |
| pythia-6.9b-deduped | Pretrained Models | 69 | 39.3 | 41.3 | 67.05 | 26.48 | 35.19 | 64.09 | 1.67 | GPTNeoXForCausalLM |
| bloom-7b1 | Pretrained Models | 70.7 | 39.18 | 41.13 | 62.0 | 26.25 | 38.9 | 65.43 | 1.36 | BloomForCausalLM |
| h2o-danube-1.8b-base | Pretrained Models | 18.3 | 39.12 | 39.42 | 69.58 | 25.94 | 33.86 | 64.48 | 1.44 | MistralForCausalLM |
| opt-6.7b | Pretrained Models | 67 | 39.08 | 39.16 | 68.66 | 24.57 | 35.12 | 65.98 | 0.99 | OPTForCausalLM |
| pythia-12b | Pretrained Models | 120 | 38.82 | 39.59 | 68.82 | 26.76 | 31.85 | 64.17 | 1.74 | GPTNeoXForCausalLM |
| weblab-10b | Pretrained Models | 100 | 38.59 | 39.51 | 65.76 | 26.29 | 36.02 | 62.51 | 1.44 | GPTNeoXForCausalLM |
| RedPajama-INCITE-Base-3B-v1 | Pretrained Models | 30 | 38.54 | 40.19 | 64.77 | 27.03 | 33.23 | 64.72 | 1.29 | GPTNeoXForCausalLM |
| open_llama_3b | Pretrained Models | 30 | 38.26 | 39.85 | 62.65 | 26.94 | 34.97 | 64.72 | 0.45 | LlamaForCausalLM |