FreedomIntelligence / TinyDeepSeekLinks
Reproduction of the complete process of DeepSeek-R1 on small-scale models, including Pre-training, SFT, and RL.
☆27Updated 5 months ago
Alternatives and similar repositories for TinyDeepSeek
Users that are interested in TinyDeepSeek are comparing it to the libraries listed below
Sorting:
- CoT-Valve: Length-Compressible Chain-of-Thought Tuning☆84Updated 6 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆126Updated last month
- Rethinking RL Scaling for Vision Language Models: A Transparent, From-Scratch Framework and Comprehensive Evaluation Scheme☆139Updated 4 months ago
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆128Updated 4 months ago
- ☆117Updated 2 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆174Updated 2 months ago
- ☆280Updated 3 months ago
- Inference Code for Paper "Harder Tasks Need More Experts: Dynamic Routing in MoE Models"☆62Updated last year
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models☆149Updated 3 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆248Updated 3 months ago
- [arXiv 2025] Efficient Reasoning Models: A Survey☆259Updated last week
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆81Updated 2 months ago
- ☆100Updated 4 months ago
- ☆65Updated 9 months ago
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆128Updated 10 months ago
- Official Repository of "Learning to Reason under Off-Policy Guidance"☆288Updated last month
- Chain of Thoughts (CoT) is so hot! so long! We need short reasoning process!☆69Updated 5 months ago
- A light-weight tool for evaluating LLMs in rule-based ways.☆69Updated 2 months ago
- Efficient Mixture of Experts for LLM Paper List☆118Updated this week
- ☆198Updated 4 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆63Updated last week
- qwen-nsa☆74Updated 4 months ago
- ☆207Updated 6 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆79Updated 2 months ago
- Extrapolating RLVR to General Domains without Verifiers☆146Updated 3 weeks ago
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆87Updated 5 months ago
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory☆173Updated last month
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆190Updated 5 months ago
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆298Updated this week
- ☆67Updated 2 months ago