raghavc / LLM-RLHF-Tuning-with-PPO-and-DPOLinks
Comprehensive toolkit for Reinforcement Learning from Human Feedback (RLHF) training, featuring instruction fine-tuning, reward model training, and support for PPO and DPO algorithms with various configurations for the Alpaca, LLaMA, and LLaMA2 models.
☆182Updated last year
Alternatives and similar repositories for LLM-RLHF-Tuning-with-PPO-and-DPO
Users that are interested in LLM-RLHF-Tuning-with-PPO-and-DPO are comparing it to the libraries listed below
Sorting:
- minimal GRPO implementation from scratch☆102Updated 10 months ago
- A simplified implementation for experimenting with RLVR on GSM8K, This repository provides a starting point for exploring reasoning.☆158Updated 11 months ago
- This is work done by the Oxen.ai Community, trying to reproduce the Self-Rewarding Language Model paper from MetaAI.☆132Updated last year
- Self-playing Adversarial Language Game Enhances LLM Reasoning, NeurIPS 2024☆143Updated 11 months ago
- ☆117Updated last year
- Code for STaR: Bootstrapping Reasoning With Reasoning (NeurIPS 2022)☆220Updated 2 years ago
- ☆99Updated last year
- (ICML 2024) Alphazero-like Tree-Search can guide large language model decoding and training☆284Updated last year
- Tina: Tiny Reasoning Models via LoRA☆313Updated 4 months ago
- Repository for the paper Stream of Search: Learning to Search in Language☆152Updated 11 months ago
- [NeurIPS 2024] GTBench: Uncovering the Strategic Reasoning Limitations of LLMs via Game-Theoretic Evaluations☆68Updated last year
- An implemtation of Everyting of Thoughts (XoT).☆156Updated last year
- Code and data for "Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs"☆474Updated last year
- ☆320Updated last year
- ☆123Updated last year
- 🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.☆621Updated 3 weeks ago
- Minimal hackable GRPO implementation☆319Updated 11 months ago
- [ACL 2024] LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement☆193Updated last year
- [ICLR 2026] Learning to Reason without External Rewards☆388Updated this week
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆114Updated last week
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆366Updated last year
- A simple unified framework for evaluating LLMs☆258Updated 9 months ago
- ☆160Updated last year
- ☆123Updated 11 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆124Updated 7 months ago
- nanoGRPO is a lightweight implementation of Group Relative Policy Optimization (GRPO)☆141Updated 8 months ago
- Official repository for ORPO☆469Updated last year
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆114Updated 6 months ago
- ☆130Updated last year
- ☆328Updated 7 months ago