CMMLU: Measuring massive multitask language understanding in Chinese
☆806Dec 6, 2024Updated last year
Alternatives and similar repositories for CMMLU
Users that are interested in CMMLU are comparing it to the libraries listed below
Sorting:
- Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]☆1,818Jul 27, 2025Updated 7 months ago
- OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, …☆6,705Feb 27, 2026Updated last week
- GAOKAO-Bench is an evaluation framework that utilizes GAOKAO questions as a dataset to evaluate large language models.☆725Jan 7, 2025Updated last year
- Measuring Massive Multitask Language Understanding | ICLR 2021☆1,558May 28, 2023Updated 2 years ago
- SuperCLUE: 中文通用大模型综合性基准 | A Benchmark for Foundation Models in Chinese☆3,274Feb 6, 2026Updated last month
- ☆772Jun 13, 2024Updated last year
- A series of large language models developed by Baichuan Intelligent Technology☆4,117Nov 8, 2024Updated last year
- 面向中文大模型价值观的评估与对齐研究☆554Jul 20, 2023Updated 2 years ago
- BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)☆8,283Oct 16, 2024Updated last year
- A 13B large language model developed by Baichuan Intelligent Technology☆2,947Sep 6, 2023Updated 2 years ago
- A large-scale 7B pretraining language model developed by BaiChuan-Inc.☆5,681Jul 18, 2024Updated last year
- Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评估和提升大模型的安全性。☆1,132Feb 27, 2024Updated 2 years ago
- Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、…☆6,638Oct 24, 2024Updated last year
- FlagEval is an evaluation toolkit for AI large foundation models.☆338Apr 24, 2025Updated 10 months ago
- Chinese Generation Evaluation☆13Aug 14, 2023Updated 2 years ago
- 万卷1.0多模态语料☆571Oct 20, 2023Updated 2 years ago
- TigerBot: A multi-language multi-task LLM☆2,263Dec 28, 2024Updated last year
- 中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)☆18,969Jul 15, 2025Updated 7 months ago
- ☆99Dec 5, 2023Updated 2 years ago
- The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.☆20,566Jan 30, 2026Updated last month
- Retrieval and Retrieval-augmented LLMs☆11,352Dec 15, 2025Updated 2 months ago
- Chinese-LLaMA 1&2、Chinese-Falcon 基础模型;ChatFlow中文对话模型;中文OpenLLaMA模型;NLP预训练/指令微调数据集☆3,055Apr 14, 2024Updated last year
- Official release of InternLM series (InternLM, InternLM2, InternLM2.5, InternLM3).☆7,159Oct 30, 2025Updated 4 months ago
- 轩辕:度小满中文金融对话大模型☆1,301Jan 7, 2025Updated last year
- 大模型多维度中文对齐评测基准 (ACL 2024)☆421Oct 25, 2025Updated 4 months ago
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,953Aug 9, 2025Updated 7 months ago
- 中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard☆4,232Feb 6, 2026Updated last month
- ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型☆15,652Jun 27, 2024Updated last year
- LongBench v2 and LongBench (ACL 25'&24')☆1,101Jan 15, 2025Updated last year
- Official github repo for ACLUE, an evaluation benchmark focused on ancient Chinese language comprehension☆33Mar 20, 2024Updated last year
- Yuan 2.0 Large Language Model☆689Jul 11, 2024Updated last year
- pCLUE: 1000000+多任务提示学习数据集☆506Oct 4, 2022Updated 3 years ago
- Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)☆2,794Mar 13, 2024Updated last year
- Fengshenbang-LM(封神榜大模型)是IDEA研究院认知计算与自然语言研究中心主导的大模型开源体系,成为中文AIGC和认知智能的基础设施。☆4,149Aug 13, 2024Updated last year
- alpaca中文指令微调数据集☆397Mar 26, 2023Updated 2 years ago
- Best practice for training LLaMA models in Megatron-LM☆663Jan 2, 2024Updated 2 years ago
- GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)☆7,669Jul 25, 2023Updated 2 years ago
- MEASURING MASSIVE MULTITASK CHINESE UNDERSTANDING☆89Mar 24, 2024Updated last year
- A framework for few-shot evaluation of language models.☆11,540Mar 2, 2026Updated last week