ZhaolinGao / REFUELLinks
Regressing the Relative Future: Efficient Policy Optimization for Multi-turn RLHF
☆19Updated 8 months ago
Alternatives and similar repositories for REFUEL
Users that are interested in REFUEL are comparing it to the libraries listed below
Sorting:
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- Critique-out-Loud Reward Models☆66Updated 8 months ago
- ☆36Updated 2 weeks ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆59Updated 4 months ago
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- AdaRFT: Efficient Reinforcement Finetuning via Adaptive Curriculum Learning☆37Updated last week
- ☆53Updated last week
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆48Updated last year
- ☆17Updated 3 months ago
- ☆46Updated 4 months ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆48Updated 7 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆30Updated last year
- Revisiting Mid-training in the Era of RL Scaling☆56Updated 2 months ago
- Dialogue Action Tokens: Steering Language Models in Goal-Directed Dialogue with a Multi-Turn Planner☆24Updated 11 months ago
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆82Updated last month
- Official Repo for InSTA: Towards Internet-Scale Training For Agents☆42Updated this week
- A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models☆57Updated 3 months ago
- Code for "Reasoning to Learn from Latent Thoughts"☆104Updated 2 months ago
- Exploration of automated dataset selection approaches at large scales.☆45Updated 3 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆59Updated 6 months ago
- official implementation of ICLR'2025 paper: Rethinking Bradley-Terry Models in Preference-based Reward Modeling: Foundations, Theory, and…☆62Updated 2 months ago
- ☆25Updated last year
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learning☆99Updated last month
- ☆114Updated 5 months ago
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Scheduling☆32Updated 3 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆44Updated 2 months ago
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"☆17Updated last year
- Codebase for Instruction Following without Instruction Tuning☆34Updated 9 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆120Updated 9 months ago
- Reinforcing General Reasoning without Verifiers☆60Updated last week