ContextualAI / HALOsLinks
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
☆894Updated 2 months ago
Alternatives and similar repositories for HALOs
Users that are interested in HALOs are comparing it to the libraries listed below
Sorting:
- RewardBench: the first evaluation tool for reward models.☆660Updated 5 months ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆931Updated 9 months ago
- Official repository for ORPO☆467Updated last year
- Representation Engineering: A Top-Down Approach to AI Transparency☆918Updated last year
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆793Updated 8 months ago
- Codebase for Merging Language Models (ICML 2024)☆861Updated last year
- A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.☆837Updated last year
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.