JinjieNi / MegaDLMsLinks
GPU-optimized framework for training diffusion language models at any scale. The backend of Quokka, Super Data Learners, and OpenMoE 2 training.
☆308Updated last month
Alternatives and similar repositories for MegaDLMs
Users that are interested in MegaDLMs are comparing it to the libraries listed below
Sorting:
- TraceRL & TraDo-8B: Revolutionizing Reinforcement Learning Framework for Diffusion Large Language Models☆380Updated 3 weeks ago
- The most open diffusion language model for code generation — releasing pretraining, evaluation, inference, and checkpoints.☆496Updated last month
- The official github repo for "Diffusion Language Models are Super Data Learners".☆215Updated 2 months ago
- Easy and Efficient dLLM Fine-Tuning☆190Updated 3 weeks ago
- [ICLR2025] DiffuGPT and DiffuLLaMA: Scaling Diffusion Language Models via Adaptation from Autoregressive Models☆357Updated 7 months ago
- QeRL enables RL for 32B LLMs on a single H100 GPU.☆469Updated last month
- Discrete Diffusion Forcing (D2F): dLLMs Can Do Faster-Than-AR Inference☆224Updated 3 months ago
- Esoteric Language Models☆108Updated last month
- Official JAX implementation of End-to-End Test-Time Training for Long Context☆214Updated last week
- ☆365Updated 2 months ago
- dInfer: An Efficient Inference Framework for Diffusion Language Models☆378Updated last week
- Official Implementation for the paper "d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning"☆392Updated 2 weeks ago
- Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"☆764Updated last month
- Official PyTorch implementation of the paper "dLLM-Cache: Accelerating Diffusion Large Language Models with Adaptive Caching" (dLLM-Cache…☆191Updated last month
- ☆109Updated 3 months ago
- The official GitHub repo for the survey paper "A Survey on Diffusion Language Models".☆628Updated 2 weeks ago
- PaCoRe: Learning to Scale Test-Time Compute with Parallel Coordinated Reasoning☆249Updated 3 weeks ago
- Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models. TMLR 2025.☆134Updated 3 months ago
- ☆118Updated last week
- Geometric-Mean Policy Optimization☆94Updated last month
- LLaDA2.0 is the diffusion language model series developed by InclusionAI team, Ant Group.☆207Updated 2 weeks ago
- 📖 This is a repository for organizing papers, codes, and other resources related to Latent Reasoning.☆328Updated 2 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆228Updated 2 months ago
- [NeurIPS 2025] Reinforcement Learning for Reasoning in Large Language Models with One Training Example☆390Updated last month
- [NeurIPS 2025] The official repo of SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond☆188Updated 6 months ago
- The official repo of One RL to See Them All: Visual Triple Unified Reinforcement Learning☆329Updated 7 months ago
- implementations and experimentation on mHC by deepseek - https://arxiv.org/abs/2512.24880☆102Updated this week
- Defeating the Training-Inference Mismatch via FP16☆170Updated last month
- Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation (NeurIPS 2025)☆529Updated 3 months ago
- repo for paper https://arxiv.org/abs/2504.13837☆310Updated 3 weeks ago