NineAbyss / S2R
This is the official implementation of the paper "S²R: Teaching LLMs to Self-verify and Self-correct via Reinforcement Learning"
☆62Updated 2 weeks ago
Alternatives and similar repositories for S2R:
Users that are interested in S2R are comparing it to the libraries listed below
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆118Updated last month
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆120Updated 7 months ago
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆71Updated 2 weeks ago
- AdaRFT: Efficient Reinforcement Finetuning via Adaptive Curriculum Learning☆29Updated 3 weeks ago
- ☆163Updated last month
- Implementation for the research paper "Enhancing LLM Reasoning via Critique Models with Test-Time and Training-Time Supervision".☆52Updated 5 months ago
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆90Updated 3 weeks ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆48Updated 5 months ago
- ☆44Updated 6 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆145Updated last month
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆57Updated 5 months ago
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆38Updated last year
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆94Updated last month
- This my attempt to create Self-Correcting-LLM based on the paper Training Language Models to Self-Correct via Reinforcement Learning by g…☆35Updated last month
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆101Updated 3 months ago
- ☆109Updated 3 months ago
- Code for "A Sober Look at Progress in Language Model Reasoning" paper☆41Updated 3 weeks ago
- Code for "Reasoning to Learn from Latent Thoughts"☆93Updated last month
- Reference implementation for Token-level Direct Preference Optimization(TDPO)☆138Updated 2 months ago
- ☆46Updated 2 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆195Updated last month
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆57Updated 6 months ago
- Code and models for EMNLP 2024 paper "WPO: Enhancing RLHF with Weighted Preference Optimization"☆40Updated 7 months ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆60Updated 4 months ago
- Directional Preference Alignment☆57Updated 7 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆56Updated 3 months ago
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆67Updated 2 months ago
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆170Updated 3 months ago
- Missing Premise exacerbates Overthinking: Are Reasoning Models losing Critical Thinking Skill?☆26Updated 3 weeks ago
- The official code repository for PRMBench.☆72Updated 2 months ago