THUDM / T1
Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling
☆95Updated 2 months ago
Alternatives and similar repositories for T1:
Users that are interested in T1 are comparing it to the libraries listed below
- Code for Paper: Teaching Language Models to Critique via Reinforcement Learning☆84Updated last month
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆104Updated last week
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆131Updated last month
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆73Updated 9 months ago
- ☆143Updated 3 months ago
- ☆128Updated last week
- Repo of paper "Free Process Rewards without Process Labels"☆138Updated 2 weeks ago
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆83Updated last week
- ☆40Updated 3 weeks ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆46Updated 4 months ago
- The official repository of the Omni-MATH benchmark.☆77Updated 3 months ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆74Updated 2 weeks ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆148Updated last week
- A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models☆44Updated last month
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆119Updated 6 months ago
- ☆49Updated 2 weeks ago
- [AAAI 2025 oral] Evaluating Mathematical Reasoning Beyond Accuracy☆55Updated 3 months ago
- Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"☆166Updated 3 weeks ago
- [NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*☆99Updated 3 months ago
- Critique-out-Loud Reward Models☆56Updated 5 months ago
- Official github repo for the paper "Compression Represents Intelligence Linearly" [COLM 2024]☆130Updated 6 months ago
- ☆103Updated 2 months ago
- We introduce ScaleQuest, a scalable, novel and cost-effective data synthesis method to unleash the reasoning capability of LLMs.☆60Updated 5 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆62Updated last month
- B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners☆75Updated 2 months ago
- ☆59Updated last week
- [EMNLP 2024] Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".☆73Updated 2 months ago
- Reformatted Alignment☆115Updated 6 months ago
- official implementation of paper "Process Reward Model with Q-value Rankings"☆51Updated last month
- Large Language Models Can Self-Improve in Long-context Reasoning☆67Updated 4 months ago