eddycmu / demystify-long-cot
☆262Updated 2 weeks ago
Alternatives and similar repositories for demystify-long-cot:
Users that are interested in demystify-long-cot are comparing it to the libraries listed below
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆162Updated 2 weeks ago
- ☆171Updated last month
- Repo of paper "Free Process Rewards without Process Labels"☆138Updated 2 weeks ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆216Updated last week
- A Survey on Efficient Reasoning for LLMs☆204Updated this week
- Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"☆170Updated 3 weeks ago
- ☆144Updated 3 months ago
- This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.☆301Updated 7 months ago
- ☆264Updated 8 months ago
- Offical Repo for "Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale"☆229Updated last month
- A series of technical report on Slow Thinking with LLM☆595Updated last week
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆161Updated last week
- 🌾 OAT: A research-friendly framework for LLM online alignment, including preference learning, reinforcement learning, etc.☆300Updated last week
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆78Updated 2 weeks ago
- ☆325Updated last month
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨☆191Updated 11 months ago
- ☆129Updated this week
- Code and example data for the paper: Rule Based Rewards for Language Model Safety☆183Updated 8 months ago
- A Comprehensive Survey on Long Context Language Modeling☆113Updated this week
- ☆574Updated 2 weeks ago
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆125Updated 3 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆131Updated last month
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆104Updated last week
- ☆125Updated 3 weeks ago
- ☆160Updated 3 weeks ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆166Updated last week
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆251Updated 6 months ago
- ReSearch: Learning to Reason with Search for LLMs via Reinforcement Learning☆395Updated this week
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆140Updated this week
- R1-searcher: Incentivizing the Search Capability in LLMs via Reinforcement Learning☆391Updated last week