louieworth / awesome-rlhfLinks
An index of algorithms for reinforcement learning from human feedback (rlhf))
☆93Updated last year
Alternatives and similar repositories for awesome-rlhf
Users that are interested in awesome-rlhf are comparing it to the libraries listed below
Sorting:
- Reference implementation for Token-level Direct Preference Optimization(TDPO)☆146Updated 6 months ago
- Code for Paper (ReMax: A Simple, Efficient and Effective Reinforcement Learning Method for Aligning Large Language Models)☆191Updated last year
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆85Updated last year
- Code for the ICML 2024 paper "Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment"☆72Updated 2 months ago
- AI Alignment: A Comprehensive Survey☆135Updated last year
- A curated reading list for large language model (LLM) alignment. Take a look at our new survey "Large Language Model Alignment: A Survey"…