RLHFlow / RAFT
This is an official implementation of the Reward rAnked Fine-Tuning Algorithm (RAFT), also known as iterative best-of-n fine-tuning or rejection sampling fine-tuning.
☆14Updated last month
Related projects ⓘ
Alternatives and complementary repositories for RAFT
- Source codes for "Preference-grounded Token-level Guidance for Language Model Fine-tuning" (NeurIPS 2023).☆14Updated last year
- This is the official implementation of ScaleBiO: Scalable Bilevel Optimization for LLM Data Reweighting☆14Updated 3 months ago
- Directional Preference Alignment☆49Updated last month
- This is an official implementation of the paper ``Building Math Agents with Multi-Turn Iterative Preference Learning'' with multi-turn DP…☆15Updated last week
- Code for ACL2024 paper - Adversarial Preference Optimization (APO).☆49Updated 5 months ago
- ☆35Updated 9 months ago
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆24Updated 8 months ago
- [ACL 2024 Findings] CriticBench: Benchmarking LLMs for Critique-Correct Reasoning☆20Updated 8 months ago
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆52Updated 2 months ago
- ☆27Updated 8 months ago
- Long Context Extension and Generalization in LLMs☆39Updated last month
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆69Updated last year
- GSM-Plus: Data, Code, and Evaluation for Enhancing Robust Mathematical Reasoning in Math Word Problems.☆46Updated 4 months ago
- [ACL 2023 Findings] What In-Context Learning “Learns” In-Context: Disentangling Task Recognition and Task Learning☆21Updated last year
- ☆85Updated 11 months ago
- Explore what LLMs are really leanring over SFT☆26Updated 7 months ago
- ☆18Updated 2 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆95Updated 2 months ago
- ☆24Updated 6 months ago
- Methods and evaluation for aligning language models temporally☆24Updated 8 months ago
- This is code for most of the experiments in the paper Understanding the Effects of RLHF on LLM Generalisation and Diversity☆38Updated 9 months ago
- ☆22Updated 2 years ago
- ☆34Updated 3 months ago
- Self-Supervised Alignment with Mutual Information☆14Updated 5 months ago
- Official implementation of Bootstrapping Language Models via DPO Implicit Rewards☆39Updated 3 months ago
- Code for paper "Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning"☆63Updated 9 months ago
- Code for Paper (Policy Optimization in RLHF: The Impact of Out-of-preference Data)☆23Updated 10 months ago
- A curated list of awesome resources dedicated to Scaling Laws for LLMs☆63Updated last year
- Teaching Models to Express Their Uncertainty in Words☆36Updated 2 years ago