ContextualAI / HALOsLinks
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
☆857Updated last week
Alternatives and similar repositories for HALOs
Users that are interested in HALOs are comparing it to the libraries listed below
Sorting:
- RewardBench: the first evaluation tool for reward models.☆604Updated last week
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆897Updated 4 months ago
- Official repository for ORPO☆455Updated last year
- Generative Representational Instruction Tuning☆651Updated 3 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆713Updated 3 months ago
- [ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning☆617Updated last year
- [COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition☆640Updated 11 months ago
- ☆773Updated last month
- Scalable toolkit for efficient model alignment☆814Updated 3 weeks ago
- ☆520Updated 7 months ago
- Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).☆768Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.