RUCAIBox / awesome-llm-pretraining
Awesome LLM pre-training resources, including data, frameworks, and methods.
☆36Updated this week
Alternatives and similar repositories for awesome-llm-pretraining:
Users that are interested in awesome-llm-pretraining are comparing it to the libraries listed below
- A Comprehensive Survey on Long Context Language Modeling☆131Updated last month
- The official repository of the Omni-MATH benchmark.☆80Updated 4 months ago
- ☆146Updated last month
- ☆187Updated 2 months ago
- Reformatted Alignment☆115Updated 7 months ago
- [NeurIPS'24] Official code for *🎯DART-Math: Difficulty-Aware Rejection Tuning for Mathematical Problem-Solving*☆104Updated 4 months ago
- Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"☆175Updated last month
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆74Updated last month
- ☆125Updated 3 weeks ago
- [NeurIPS 2024] Fast Best-of-N Decoding via Speculative Rejection☆42Updated 5 months ago
- On Memorization of Large Language Models in Logical Reasoning☆65Updated 3 weeks ago
- ☆63Updated 5 months ago
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆50Updated this week
- Code for Paper: Teaching Language Models to Critique via Reinforcement Learning☆94Updated last week
- Reproducing R1 for Code with Reliable Rewards☆179Updated this week
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".☆64Updated last week
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆133Updated last month
- ☆179Updated 2 weeks ago
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimization☆34Updated 2 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆174Updated 2 weeks ago
- ☆37Updated 2 weeks ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆173Updated last month
- Async pipelined version of Verl☆60Updated 2 weeks ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆132Updated 10 months ago
- ☆76Updated last week
- An Open Math Pre-trainng Dataset with 370B Tokens.☆72Updated 3 weeks ago
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆95Updated last month
- LongRecipe: Recipe for Efficient Long Context Generalization in Large Language Models☆76Updated 6 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆229Updated last week
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆101Updated 3 months ago