大模型多维度中文对齐评测基准 (ACL 2024)
☆420Oct 25, 2025Updated 4 months ago
Alternatives and similar repositories for AlignBench
Users that are interested in AlignBench are comparing it to the libraries listed below
Sorting:
- ☆148Jul 1, 2024Updated last year
- Benchmarking Complex Instruction-Following with Multiple Constraints Composition (NeurIPS 2024 Datasets and Benchmarks Track)☆102Feb 20, 2025Updated last year
- OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, …☆6,688Updated this week
- [ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset☆111May 22, 2025Updated 9 months ago
- Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评 估和提升大模型的安全性。☆1,129Feb 27, 2024Updated 2 years ago
- Dataset and evaluation script for "Evaluating Hallucinations in Chinese Large Language Models"☆136Jun 5, 2024Updated last year
- Generative Judge for Evaluating Alignment☆250Jan 18, 2024Updated 2 years ago
- [ACL 2024] FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models☆119Jun 12, 2025Updated 8 months ago
- FlagEval is an evaluation toolkit for AI large foundation models.☆338Apr 24, 2025Updated 10 months ago
- Awesome-LLM-Eval: a curated list of tools, datasets/benchmark, demos, leaderboard, papers, docs and models, mainly for Evaluation on LLMs…☆612Nov 24, 2025Updated 3 months ago
- Collection of papers for scalable automated alignment.☆93Oct 22, 2024Updated last year
- ☆285May 27, 2025Updated 9 months ago
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large L…☆53Jun 24, 2024Updated last year
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)☆9,037Feb 21, 2026Updated last week
- InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning☆285Aug 20, 2023Updated 2 years ago
- A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)☆3,187Feb 8, 2026Updated 2 weeks ago
- 面向中文大模型价值观的评估与对齐研究☆553Jul 20, 2023Updated 2 years ago
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Following☆137Jul 8, 2024Updated last year
- ☆84Apr 18, 2024Updated last year
- LongBench v2 and LongBench (ACL 25'&24')☆1,095Jan 15, 2025Updated last year
- O1 Replication Journey☆1,999Jan 14, 2025Updated last year
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆416Jun 25, 2025Updated 8 months ago
- Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback☆1,585Nov 24, 2025Updated 3 months ago
- ☆324Jul 25, 2024Updated last year
- ☆979Feb 7, 2025Updated last year
- Scalable toolkit for efficient model alignment☆851Oct 6, 2025Updated 4 months ago
- CMMLU: Measuring massive multitask language understanding in Chinese☆804Dec 6, 2024Updated last year
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆260Dec 16, 2024Updated last year
- Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]☆1,815Jul 27, 2025Updated 7 months ago
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆588Dec 9, 2024Updated last year
- ☆30Dec 27, 2024Updated last year
- BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)☆8,281Oct 16, 2024Updated last year
- ☆99Dec 5, 2023Updated 2 years ago
- A series of technical report on Slow Thinking with LLM☆760Aug 13, 2025Updated 6 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆139Jun 12, 2024Updated last year
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,003Jun 21, 2025Updated 8 months ago
- ☆59Aug 22, 2024Updated last year
- OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models☆1,833Jan 17, 2025Updated last year
- [ACL'24 Outstanding] Data and code for L-Eval, a comprehensive long context language models evaluation benchmark☆393Jul 9, 2024Updated last year