AI-EDU-LAB / E-EVALLinks
Official github repo for E-Eval, a Chinese K12 education evaluation benchmark for LLMs.
☆27Updated last year
Alternatives and similar repositories for E-EVAL
Users that are interested in E-EVAL are comparing it to the libraries listed below
Sorting:
- The implementation of paper "LLM Critics Help Catch Bugs in Mathematics: Towards a Better Mathematical Verifier with Natural Language Fee…☆38Updated 11 months ago
- Benchmarking Complex Instruction-Following with Multiple Constraints Composition (NeurIPS 2024 Datasets and Benchmarks Track)☆87Updated 4 months ago
- Llama-3-SynE: A Significantly Enhanced Version of Llama-3 with Advanced Scientific Reasoning and Chinese Language Capabilities | 继续预训练提升 …☆33Updated 3 weeks ago
- ☆82Updated last year
- [ICML'2024] Can AI Assistants Know What They Don't Know?☆81Updated last year
- 基于DPO算法微调语言大模型,简单好上手。☆39Updated 11 months ago
- ☆48Updated last year
- ☆30Updated 4 months ago
- Clustering and Ranking: Diversity-preserved Instruction Selection through Expert-aligned Quality Estimation☆81Updated 7 months ago
- ☆15Updated 7 months ago
- Collection of papers for scalable automated alignment.☆91Updated 8 months ago
- [ACL 2024] FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models☆104Updated 2 weeks ago
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆54Updated 6 months ago
- this is an implementation for the paper Improve Mathematical Reasoning in Language Models by Automated Process Supervision from google de…☆33Updated 2 months ago
- A Bilingual Role Evaluation Benchmark for Large Language Models☆41Updated last year
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆78Updated 5 months ago
- ☆101Updated 8 months ago
- Official completion of “Training on the Benchmark Is Not All You Need”.☆34Updated 5 months ago
- ☆44Updated 4 months ago
- ☆142Updated 11 months ago
- ☆141Updated last year
- [ACL 2025, Main Conference, Oral] Intuitive Fine-Tuning: Towards Simplifying Alignment into a Single Process☆28Updated 10 months ago
- [COLING 2025] ToolEyes: Fine-Grained Evaluation for Tool Learning Capabilities of Large Language Models in Real-world Scenarios☆68Updated last month
- Dataset and evaluation script for "Evaluating Hallucinations in Chinese Large Language Models"☆130Updated last year
- A simple implementation of ReasonGenRM.☆13Updated 2 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆122Updated 7 months ago
- Flames is a highly adversarial benchmark in Chinese for LLM's harmlessness evaluation developed by Shanghai AI Lab and Fudan NLP Group.☆51Updated last year
- Personality Alignment of Language Models☆37Updated 3 months ago
- ☆54Updated 10 months ago
- Official github repo for AutoDetect, an automated weakness detection framework for LLMs.☆42Updated last year