pengr / DataManLinks
Our code for ICLR'25 paper "DataMan: Data Manager for Pre-training Large Language Models".
☆114Updated 5 months ago
Alternatives and similar repositories for DataMan
Users that are interested in DataMan are comparing it to the libraries listed below
Sorting:
- ☆215Updated 11 months ago
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆155Updated last year
- [ACL 2024] The official codebase for the paper "Self-Distillation Bridges Distribution Gap in Language Model Fine-tuning".☆145Updated last year
- Model merging is a highly efficient approach for long-to-short reasoning.☆98Updated 3 months ago
- ☆111Updated 7 months ago
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆143Updated 2 months ago
- [ICML 2024] Selecting High-Quality Data for Training Language Models☆201Updated 2 months ago
- ☆175Updated last year
- [R]einforcement [L]earning from [M]odel-rewarded [T]hinking - code for the paper "Language Models That Think, Chat Better"☆124Updated 3 months ago
- Pre-trained, Scalable, High-performance Reward Models via Policy Discriminative Learning.☆164Updated 4 months ago
- [ICLR 2025] 🧬 RegMix: Data Mixture as Regression for Language Model Pre-training (Spotlight)☆185Updated 11 months ago
- Scaling Agentic Reinforcement Learning with a Multi-Turn, Multi-Task Framework☆205Updated 3 weeks ago
- a-m-team's exploration in large language modeling☆195Updated 8 months ago
- ☆333Updated 8 months ago
- [ICLR 2026] PSFT is a trust-region–inspired fine-tuning objective that views SFT as a policy gradient method with constant advantages, co…☆34Updated 5 months ago
- ☆182Updated 9 months ago
- ☆196Updated last year
- Extrapolating RLVR to General Domains without Verifiers☆196Updated 5 months ago
- A Comprehensive Survey on Long Context Language Modeling☆226Updated 2 months ago
- Towards a Unified View of Large Language Model Post-Training☆200Updated 5 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆86Updated this week
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models (NeurIPS 2025)☆172Updated 3 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆249Updated 9 months ago
- ☆306Updated 7 months ago
- The related works and background techniques about Openai o1☆220Updated last year
- A research repo for experiments about Reinforcement Finetuning☆54Updated 10 months ago
- ☆178Updated 2 months ago
- [NeurIPS'24] Weak-to-Strong Search: Align Large Language Models via Searching over Small Language Models☆65Updated last year
- MiroRL is an MCP-first reinforcement learning framework for deep research agent.☆229Updated 5 months ago
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆520Updated this week