Official repository for ORPO
☆472May 31, 2024Updated last year
Alternatives and similar repositories for orpo
Users that are interested in orpo are comparing it to the libraries listed below
Sorting:
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).☆906Sep 30, 2025Updated 5 months ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆946Feb 16, 2025Updated last year
- Robust recipes to align language models with human and AI preferences☆5,510Sep 8, 2025Updated 5 months ago
- ☆130Oct 1, 2024Updated last year
- Official code for "MAmmoTH2: Scaling Instructions from the Web" [NeurIPS 2024]☆149Oct 27, 2024Updated last year
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,235May 8, 2024Updated last year
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆589Dec 9, 2024Updated last year
- Stanford NLP Python library for Representation Finetuning (ReFT)☆1,560Jan 14, 2026Updated last month
- RewardBench: the first evaluation tool for reward models.☆697Feb 16, 2026Updated 2 weeks ago
- A recipe for online RLHF and online iterative DPO.☆540Dec 28, 2024Updated last year
- Reference implementation for DPO (Direct Preference Optimization)☆2,859Aug 11, 2024Updated last year
- ☆16Jul 23, 2024Updated last year
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifi…☆3,108Feb 23, 2026Updated last week
- Tools for merging pretrained large language models.☆6,826Updated this week
- ☆320Sep 18, 2024Updated last year
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆209May 20, 2024Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆61Aug 30, 2024Updated last year
- Code for Paper (ReMax: A Simple, Efficient and Effective Reinforcement Learning Method for Aligning Large Language Models)☆200Dec 16, 2023Updated 2 years ago
- Scalable toolkit for efficient model alignment☆849Oct 6, 2025Updated 5 months ago
- Self-Supervised Alignment with Mutual Information☆20May 24, 2024Updated last year
- AllenAI's post-training codebase☆3,605Updated this week
- Self-Alignment with Principle-Following Reward Models☆170Sep 18, 2025Updated 5 months ago
- Training LLMs with QLoRA + FSDP☆1,538Nov 9, 2024Updated last year
- Official repository for ACL 2025 paper "Model Extrapolation Expedites Alignment"☆75May 20, 2025Updated 9 months ago
- Minimalistic large language model 3D-parallelism training☆2,579Feb 19, 2026Updated 2 weeks ago
- Official implementation of Half-Quadratic Quantization (HQQ)☆915Feb 26, 2026Updated last week
- ☆325Jul 25, 2024Updated last year
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆833Mar 17, 2025Updated 11 months ago
- CodeUltraFeedback: aligning large language models to coding preferences (TOSEM 2025)☆73Jun 25, 2024Updated last year
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆191Jan 16, 2025Updated last year
- An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.☆1,953Aug 9, 2025Updated 6 months ago
- Evaluate your LLM's response with Prometheus and GPT4 💯☆1,050Apr 25, 2025Updated 10 months ago
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆512Oct 20, 2024Updated last year
- Code for Quiet-STaR☆741Aug 21, 2024Updated last year
- Arena-Hard-Auto: An automatic LLM benchmark.☆1,003Jun 21, 2025Updated 8 months ago
- Low-Rank adapter extraction for fine-tuned transformers models☆180May 2, 2024Updated last year
- Recipes to train reward model for RLHF.☆1,517Apr 24, 2025Updated 10 months ago
- YaRN: Efficient Context Window Extension of Large Language Models☆1,676Apr 17, 2024Updated last year
- A bagel, with everything.☆326Apr 11, 2024Updated last year