sail-sg / CPO
[NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.
☆112Updated 3 weeks ago
Alternatives and similar repositories for CPO:
Users that are interested in CPO are comparing it to the libraries listed below
- ☆86Updated last week
- Repo of paper "Free Process Rewards without Process Labels"☆140Updated 3 weeks ago
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆52Updated 4 months ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆58Updated 3 months ago
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆99Updated 2 months ago
- ☆49Updated last month
- ☆54Updated 5 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆58Updated 4 months ago
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆57Updated 5 months ago
- ☆44Updated 5 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆66Updated 2 weeks ago
- ☆182Updated last month
- ☆146Updated 2 weeks ago
- This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"☆57Updated 3 weeks ago
- Code for Paper: Teaching Language Models to Critique via Reinforcement Learning☆88Updated last month
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆74Updated 2 months ago
- ☆91Updated last month
- ☆147Updated 3 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆175Updated 3 weeks ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆120Updated 7 months ago
- Curation of resources for LLM mathematical reasoning, most of which are screened by @tongyx361 to ensure high quality and accompanied wit…☆123Updated 9 months ago
- What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆63Updated last month
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆61Updated 5 months ago
- ☆65Updated last year
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆89Updated last month
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆48Updated 5 months ago
- The official repository of the Omni-MATH benchmark.☆79Updated 3 months ago
- ☆59Updated 7 months ago
- Code implementation of synthetic continued pretraining☆98Updated 3 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆114Updated last month