BytedTsinghua-SIA / DAPO
An Open-source RL System from ByteDance Seed and Tsinghua AIR
☆767Updated last week
Alternatives and similar repositories for DAPO:
Users that are interested in DAPO are comparing it to the libraries listed below
- Large Reasoning Models☆799Updated 3 months ago
- Official Repo for Open-Reasoner-Zero☆1,667Updated 3 weeks ago
- ☆559Updated last week
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)☆597Updated 2 months ago
- A fork to add multimodal model training to open-r1☆1,108Updated last month
- A series of technical report on Slow Thinking with LLM☆595Updated this week
- ☆910Updated 2 months ago
- O1 Replication Journey☆1,977Updated 2 months ago
- OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models☆1,732Updated 2 months ago
- Explore the Multimodal “Aha Moment” on 2B Model☆524Updated last week
- ☆485Updated last week
- Scalable RL solution for advanced reasoning of language models☆1,419Updated last week
- ☆518Updated this week
- Recipes to train reward model for RLHF.☆1,257Updated last month
- Understanding R1-Zero-Like Training: A Critical Perspective☆568Updated this week
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆660Updated last week
- Muon is Scalable for LLM Training☆974Updated last month
- LIMO: Less is More for Reasoning☆864Updated last month
- EasyR1: An Efficient, Scalable, Multi-Modality RL Training Framework based on veRL☆1,681Updated this week
- RAGEN leverages reinforcement learning to train LLM reasoning agents in interactive, stochastic environments.☆1,210Updated this week
- 🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"☆590Updated last week
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆851Updated last month
- Extend OpenRLHF to support LMM RL training for reproduction of DeepSeek-R1 on multimodal tasks.☆643Updated this week
- MM-EUREKA: Exploring Visual Aha Moment with Rule-based Large-scale Reinforcement Learning☆425Updated last week
- Official codebase for "Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling".☆225Updated last month
- AN O1 REPLICATION FOR CODING☆329Updated 3 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆693Updated last week
- R1-searcher: Incentivizing the Search Capability in LLMs via Reinforcement Learning☆376Updated this week
- ☆260Updated last week
- ☆504Updated 2 months ago