microsoft / RedStone
The RedStone repository includes code for preparing extensive datasets used in training large language models.
☆133Updated 2 weeks ago
Alternatives and similar repositories for RedStone
Users that are interested in RedStone are comparing it to the libraries listed below
Sorting:
- ☆151Updated last week
- [ICML 2025] Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale☆244Updated 3 weeks ago
- An Open Math Pre-trainng Dataset with 370B Tokens.☆80Updated last month
- ☆279Updated 9 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆235Updated 3 weeks ago
- ☆132Updated 2 weeks ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆174Updated last month
- Scaling Deep Research via Reinforcement Learning in Real-world Environments.☆345Updated last month
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignment☆75Updated 10 months ago
- ☆133Updated 3 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆132Updated 11 months ago
- Reformatted Alignment☆115Updated 7 months ago
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆249Updated 4 months ago
- ☆143Updated 10 months ago
- A Comprehensive Survey on Long Context Language Modeling☆139Updated last month
- ☆162Updated last month
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆189Updated 2 months ago
- ☆194Updated 2 months ago
- [ACL 2024 Findings] MathBench: A Comprehensive Multi-Level Difficulty Mathematics Evaluation Dataset☆98Updated 10 months ago
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆164Updated last week
- ☆315Updated 7 months ago
- code for Scaling Laws of RoPE-based Extrapolation☆73Updated last year
- Reproducing R1 for Code with Reliable Rewards☆188Updated last week
- ☆29Updated 8 months ago
- A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning☆180Updated last week
- Mixture-of-Experts (MoE) Language Model☆186Updated 8 months ago
- [ICLR 2025] The official implementation of paper "ToolGen: Unified Tool Retrieval and Calling via Generation"☆141Updated last month
- On Memorization of Large Language Models in Logical Reasoning☆64Updated last month
- ☆94Updated 4 months ago
- ☆94Updated 5 months ago