Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs. 一个由工具、基准/数据、演示、排行榜和大模型等组成的精选列表,主要面向基础大模型评测,旨在探求生成式AI的技术边界.
☆637Nov 24, 2025Updated 5 months ago
Alternatives and similar repositories for Awesome-LLM-Eval
Users that are interested in Awesome-LLM-Eval are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.☆801May 8, 2024Updated 2 years ago
- The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".☆1,602Apr 17, 2026Updated 3 weeks ago
- 大模型多维度中文对齐评测基准 (ACL 2024)☆427Oct 25, 2025Updated 6 months ago
- OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, …☆6,959Apr 20, 2026Updated 2 weeks ago
- A framework for few-shot evaluation of language models.☆12,411Apr 30, 2026Updated last week
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- FlagEval is an evaluation toolkit for AI large foundation models.☆336Apr 24, 2025Updated last year
- Awesome LLM Benchmarks to evaluate the LLMs across text, code, image, audio, video and more.☆162Jan 3, 2024Updated 2 years ago
- [ACL 2024 Demo] Official GitHub repo for UltraEval: An open source framework for evaluating foundation models.☆258Oct 30, 2024Updated last year
- Summarize existing representative LLMs text datasets.☆1,461Mar 11, 2026Updated last month
- Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评估和提升大模型的安全性。☆1,153Feb 27, 2024Updated 2 years ago
- The official GitHub page for the survey paper "A Survey of Large Language Models".☆12,153Mar 11, 2025Updated last year
- [AAAI 2024] LLMEval Phase II dataset — professional domain evaluation across 12 academic disciplines☆71Apr 15, 2026Updated 3 weeks ago
- 面向中文大模型价值观的评估与对齐研究☆556Jul 20, 2023Updated 2 years ago
- Data and code for paper "M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models"☆104Jun 15, 2023Updated 2 years ago
- Deploy to Railway using AI coding agents - Free Credits Offer • AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,771Aug 4, 2024Updated last year
- Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]☆1,845Jul 27, 2025Updated 9 months ago
- From Chain-of-Thought prompting to OpenAI o1 and DeepSeek-R1 🍓☆3,605Apr 20, 2026Updated 2 weeks ago
- aigc evals☆10Dec 2, 2023Updated 2 years ago
- A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)☆3,399Feb 8, 2026Updated 3 months ago
- Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)☆70,969Updated this week
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,982Aug 9, 2025Updated 9 months ago
- ☆2,899Feb 20, 2025Updated last year
- 总结Prompt&LLM论文,开源数据&模型,AIGC应用☆3,408Updated this week
- Proton VPN Special Offer - Get 70% off • AdSpecial partner offer. Trusted by over 100 million users worldwide. Tested, Approved and Recommended by Experts.
- BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)☆8,284Oct 16, 2024Updated last year
- Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large …☆1,082Sep 27, 2025Updated 7 months ago
- A unified evaluation framework for large language models☆2,803Feb 20, 2026Updated 2 months ago
- 整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。☆22,563Apr 23, 2026Updated 2 weeks ago
- CFBench: A Comprehensive Constraints-Following Benchmark for LLMs☆52Aug 26, 2024Updated last year
- CMMLU: Measuring massive multitask language understanding in Chinese☆812Dec 6, 2024Updated last year
- Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷☆6,368Apr 28, 2026Updated last week
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & VLM & TIS & vLLM & Ray & Asy…☆9,441Updated this week
- ⚡LLM Zoo is a project that provides data, models, and evaluation benchmark for large language models.⚡☆2,948Nov 26, 2023Updated 2 years ago
- Wordpress hosting with auto-scaling - Free Trial Offer • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- Aligning Large Language Models with Human: A Survey☆741Sep 11, 2023Updated 2 years ago
- Retrieval and Retrieval-augmented LLMs☆11,642Apr 22, 2026Updated 2 weeks ago
- LLM Inference benchmark☆437Jul 23, 2024Updated last year
- LongBench v2 and LongBench (ACL 25'&24')☆1,164Jan 15, 2025Updated last year
- GPT-Fathom is an open-source and reproducible LLM evaluation suite, benchmarking 10+ leading open-source and closed-source LLMs as well a…☆344Apr 10, 2024Updated 2 years ago
- 汽车行业中文大模型测评基准,基于多轮开放式问题的细粒度评测☆37Dec 26, 2023Updated 2 years ago
- [ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.☆5,628May 21, 2025Updated 11 months ago