icip-cas / awesome-auto-alignmentView external linksLinks
Collection of papers for scalable automated alignment.
โ93Oct 22, 2024Updated last year
Alternatives and similar repositories for awesome-auto-alignment
Users that are interested in awesome-auto-alignment are comparing it to the libraries listed below
Sorting:
- Paper list of "The Life Cycle of Knowledge in Big Language Models: A Survey"โ59Aug 24, 2023Updated 2 years ago
- Official Code Repository for [AutoScale๐: Scale-Aware Data Mixing for Pre-Training LLMs] Published as a conference paper at **COLM 2025*โฆโ13Aug 8, 2025Updated 6 months ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AIโ107Mar 6, 2025Updated 11 months ago
- Self-Supervised Alignment with Mutual Informationโ20May 24, 2024Updated last year
- RACE is a multi-dimensional benchmark for code generation that focuses on Readability, mAintainability, Correctness, and Efficiency.โ12Oct 12, 2024Updated last year
- Official repository for ALT (ALignment with Textual feedback).โ10Jul 25, 2024Updated last year
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineeringโ63Dec 5, 2024Updated last year
- โ23Jul 5, 2024Updated last year
- Reproducing R1 for Code with Reliable Rewardsโ12Apr 9, 2025Updated 10 months ago
- Code for paper "ProgGen: Generating Named Entity Recognition Datasets Step-by-step with Self-Reflexive Large Language Models"โ17Mar 29, 2024Updated last year
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)โ73Jun 25, 2024Updated last year
- Code and data for the paper "Can Large Language Models Understand Real-World Complex Instructions?"(AAAI2024)โ50Apr 19, 2024Updated last year
- StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedbackโ74Aug 31, 2024Updated last year
- Benchmarking Complex Instruction-Following with Multiple Constraints Composition (NeurIPS 2024 Datasets and Benchmarks Track)โ101Feb 20, 2025Updated 11 months ago
- Implementation of Not All Contexts Are Equal: Teaching LLMs Credibility-aware Generation. Paper: https://arxiv.org/abs/2404.06809โ22Oct 22, 2024Updated last year
- [๐๐๐๐๐ ๐ ๐ข๐ง๐๐ข๐ง๐ ๐ฌ ๐๐๐๐ & ๐๐๐ ๐๐๐๐ ๐๐๐๐๐ ๐๐ซ๐๐ฅ] ๐๐ฏ๐ฉ๐ข๐ฏ๐ค๐ช๐ฏ๐จ ๐๐ข๐ต๐ฉ๐ฆ๐ฎ๐ข๐ต๐ช๐ค๐ข๐ญ ๐๐ฆ๐ข๐ด๐ฐ๐ฏ๐ช๐ฏโฆโ51May 4, 2024Updated last year
- Code for paper: Long cOntext aliGnment via efficient preference Optimizationโ24Oct 10, 2025Updated 4 months ago
- โ322Jul 25, 2024Updated last year
- ๅคงๆจกๅๅค็ปดๅบฆไธญๆๅฏน้ฝ่ฏๆตๅบๅ (ACL 2024)โ421Oct 25, 2025Updated 3 months ago
- A Multi-Turn Dialogue Corpus based on Alpaca Instructionsโ177Jun 1, 2023Updated 2 years ago
- A scalable automated alignment method for large language models. Resources for "Aligning Large Language Models via Self-Steering Optimizaโฆโ20Nov 21, 2024Updated last year
- Small, simple agent task environments for training and evaluationโ19Nov 1, 2024Updated last year
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).โ56Jun 3, 2024Updated last year
- โ62Oct 29, 2024Updated last year
- โ27Oct 9, 2025Updated 4 months ago
- Aligning Large Language Models with Human: A Surveyโ742Sep 11, 2023Updated 2 years ago
- Scalable toolkit for efficient model alignmentโ852Oct 6, 2025Updated 4 months ago
- The official repository of our survey paper: "Towards a Unified View of Preference Learning for Large Language Models: A Survey"โ189Oct 28, 2024Updated last year
- [2025-TMLR] A Survey on the Honesty of Large Language Modelsโ64Dec 8, 2024Updated last year
- Code for Paper (ReMax: A Simple, Efficient and Effective Reinforcement Learning Method for Aligning Large Language Models)โ199Dec 16, 2023Updated 2 years ago
- โ321Sep 18, 2024Updated last year
- Repository containing the SPIN experiments on the DIBT 10k ranked promptsโ23Mar 12, 2024Updated last year
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]โ588Dec 9, 2024Updated last year
- [ICLR 2024] Evaluating Large Language Models at Evaluating Instruction Followingโ136Jul 8, 2024Updated last year
- [ACL 2024] Code for the paper "ALaRM: Align Language Models via Hierarchical Rewards Modeling"โ25Mar 28, 2024Updated last year
- A comprehensive benchmark for evaluating deep research agents on academic survey tasksโ49Sep 4, 2025Updated 5 months ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoningโ30Mar 5, 2024Updated last year
- Repository for tw.org siteโ14Jan 8, 2026Updated last month
- The code and data for the paper JiuZhang3.0โ49May 26, 2024Updated last year