sail-sg / oat-zero
A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.
☆212Updated this week
Alternatives and similar repositories for oat-zero:
Users that are interested in oat-zero are comparing it to the libraries listed below
- ☆166Updated last month
- ☆260Updated last week
- A series of technical report on Slow Thinking with LLM☆581Updated this week
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆158Updated this week
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆148Updated last week
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨☆189Updated 10 months ago
- This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.☆297Updated 7 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆138Updated last week
- ☆263Updated 8 months ago
- On Memorization of Large Language Models in Logical Reasoning☆56Updated 4 months ago
- Offical Repo for "Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale"☆229Updated last month
- ☆139Updated last week
- The related works and background techniques about Openai o1☆217Updated 2 months ago
- Reproducing R1 for Code with Reliable Rewards☆132Updated 3 weeks ago
- ☆325Updated last month
- Code for Paper (ReMax: A Simple, Efficient and Effective Reinforcement Learning Method for Aligning Large Language Models)☆178Updated last year
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆246Updated 3 months ago
- ☆124Updated 2 weeks ago
- ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models☆179Updated 5 months ago
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆250Updated 6 months ago
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)☆595Updated 2 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆165Updated this week
- ☆186Updated this week
- ☆312Updated 6 months ago
- ☆143Updated 3 months ago
- Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"☆357Updated 2 months ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆74Updated last week
- R1-searcher: Incentivizing the Search Capability in LLMs via Reinforcement Learning☆376Updated this week
- [NeurIPS 2024 Oral] Aligner: Efficient Alignment by Learning to Correct☆165Updated 2 months ago
- Curation of resources for LLM mathematical reasoning, most of which are screened by @tongyx361 to ensure high quality and accompanied wit…☆118Updated 8 months ago