crazycth / WizardLearnerLinks
Pretrain、decay、SFT a CodeLLM from scratch 🧙♂️
☆38Updated last year
Alternatives and similar repositories for WizardLearner
Users that are interested in WizardLearner are comparing it to the libraries listed below
Sorting:
- ☆65Updated 9 months ago
- RLHF experiments on a single A100 40G GPU. Support PPO, GRPO, REINFORCE, RAFT, RLOO, ReMax, DeepSeek R1-Zero reproducing.☆72Updated 7 months ago
- ☆114Updated last year
- ☆74Updated last month
- This is a repo for showcasing using MCTS with LLMs to solve gsm8k problems☆89Updated 6 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆137Updated last year
- ☆201Updated 5 months ago
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆341Updated this week
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆253Updated 7 months ago
- ☆114Updated 10 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆214Updated last month
- ☆125Updated last year
- xVerify: Efficient Answer Verifier for Reasoning Model Evaluations☆130Updated 5 months ago
- [ACL 2024 Demo] Official GitHub repo for UltraEval: An open source framework for evaluating foundation models.☆248Updated 10 months ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆58Updated last year
- A Comprehensive Survey on Long Context Language Modeling☆187Updated 2 months ago
- ☆33Updated 6 months ago
- The related works and background techniques about Openai o1☆222Updated 8 months ago
- ☆33Updated 3 months ago
- ☆165Updated 4 months ago
- MiroRL is an MCP-first reinforcement learning framework for deep research agent.☆154Updated 3 weeks ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆190Updated 6 months ago
- 代码大模型 预训练&微调&DPO 数据处理 业界处理pipeline sota☆44Updated last year
- Adapt an LLM model to a Mixture-of-Experts model using Parameter Efficient finetuning (LoRA), injecting the LoRAs in the FFN.☆57Updated 11 months ago
- Counting-Stars (★)☆83Updated 3 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆244Updated 5 months ago
- [EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs☆256Updated 9 months ago
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆149Updated 8 months ago
- ☆159Updated 8 months ago
- Token level visualization tools for large language models☆88Updated 8 months ago