convergence-ai / lm2Links
Official repo of paper LM2
☆46Updated 9 months ago
Alternatives and similar repositories for lm2
Users that are interested in lm2 are comparing it to the libraries listed below
Sorting:
- ☆124Updated 9 months ago
- Code Implementation, Evaluations, Documentation, Links and Resources for Min P paper☆45Updated 3 months ago
- [EMNLP'25 Industry] Repo for "Z1: Efficient Test-time Scaling with Code"☆67Updated 7 months ago
- ☆29Updated 3 weeks ago
- Official implementation of Regularized Policy Gradient (RPG) (https://arxiv.org/abs/2505.17508)☆54Updated last month
- ☆89Updated last year
- ☆109Updated last year
- [Preprint] RLVE: Scaling Up Reinforcement Learning for Language Models with Adaptive Verifiable Environments☆151Updated 3 weeks ago
- ☆226Updated 9 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆112Updated 4 months ago
- Esoteric Language Models☆107Updated last week
- ☆51Updated 9 months ago
- Process Reward Models That Think☆63Updated last week
- [NeurIPS 2025 Spotlight] Co-Evolving LLM Coder and Unit Tester via Reinforcement Learning☆135Updated 2 months ago
- The official implementation of Self-Exploring Language Models (SELM)☆63Updated last year
- Sotopia-RL: Reward Design for Social Intelligence☆44Updated 3 months ago
- SSRL: Self-Search Reinforcement Learning☆157Updated 3 months ago
- [EMNLP 2025] The official implementation for paper "Agentic-R1: Distilled Dual-Strategy Reasoning"☆100Updated 3 months ago
- Reinforcing General Reasoning without Verifiers☆92Updated 5 months ago
- SPIRAL: Self-Play on Zero-Sum Games Incentivizes Reasoning via Multi-Agent Multi-Turn Reinforcement Learning☆161Updated 2 months ago
- accompanying material for sleep-time compute paper☆118Updated 7 months ago
- Natural Language Reinforcement Learning☆100Updated 4 months ago
- ☆85Updated 5 months ago
- This is the official repository for Inheritune.☆115Updated 9 months ago
- Code for the paper: "Learning to Reason without External Rewards"☆380Updated 4 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆60Updated last year
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆115Updated 5 months ago
- RL Scaling and Test-Time Scaling (ICML'25)☆112Updated 10 months ago
- ☆26Updated 10 months ago
- Code for "Reasoning to Learn from Latent Thoughts"☆122Updated 8 months ago