swtheing / PF-PPO-RLHF
☆30Updated 4 months ago
Alternatives and similar repositories for PF-PPO-RLHF:
Users that are interested in PF-PPO-RLHF are comparing it to the libraries listed below
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆44Updated last month
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆33Updated last year
- The code and data for the paper JiuZhang3.0☆40Updated 7 months ago
- [NeurIPS 2024] Official code of $\beta$-DPO: Direct Preference Optimization with Dynamic $\beta$☆38Updated 2 months ago
- The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agen…☆23Updated 10 months ago
- Domain-specific preference (DSP) data and customized RM fine-tuning.☆24Updated 10 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated 3 weeks ago
- ☆46Updated last month
- Feeling confused about super alignment? Here is a reading list☆42Updated last year
- ☆21Updated 3 months ago
- ☆93Updated last year
- Code and models for EMNLP 2024 paper "WPO: Enhancing RLHF with Weighted Preference Optimization"☆32Updated 3 months ago
- ☆61Updated 9 months ago
- The source code of "Merging Experts into One: Improving Computational Efficiency of Mixture of Experts (EMNLP 2023)":☆34Updated 9 months ago
- ☆93Updated 3 months ago
- Implementation of ICML 23 Paper: Specializing Smaller Language Models towards Multi-Step Reasoning.☆128Updated last year
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆71Updated 7 months ago
- Code implementation of synthetic continued pretraining☆79Updated last week
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆62Updated this week
- A curated list of awesome resources dedicated to Scaling Laws for LLMs☆69Updated last year
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆88Updated 3 months ago
- Long Context Extension and Generalization in LLMs☆40Updated 3 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆27Updated 7 months ago
- ☆47Updated 9 months ago
- ☆26Updated 3 weeks ago
- [NeurIPS 2024] Fast Best-of-N Decoding via Speculative Rejection☆38Updated 2 months ago
- ☆28Updated last year
- This the implementation of LeCo☆30Updated 6 months ago
- GenRM-CoT: Data release for verification rationales☆42Updated 3 months ago
- This repo is reproduction resources for linear alignment paper, still working☆17Updated 7 months ago