kuleshov-group / bd3lmsLinks
Block Diffusion: Interpolating Between Autoregressive and Diffusion Language Models
☆749Updated 3 weeks ago
Alternatives and similar repositories for bd3lms
Users that are interested in bd3lms are comparing it to the libraries listed below
Sorting:
- Dream 7B, a large diffusion language model☆873Updated last month
- MMaDA - Open-Sourced Multimodal Large Diffusion Language Models☆1,275Updated last month
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆567Updated 5 months ago
- Official Implementation for the paper "d1: Scaling Reasoning in Diffusion Large Language Models via Reinforcement Learning"☆255Updated last month
- [ICLR2025] DiffuGPT and DiffuLLaMA: Scaling Diffusion Language Models via Adaptation from Autoregressive Models☆253Updated 2 months ago
- Official PyTorch implementation for ICLR2025 paper "Scaling up Masked Diffusion Models on Text"☆267Updated 7 months ago
- code for "Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion"☆948Updated 4 months ago
- An official implementation of Flow-GRPO: Training Flow Matching Models via Online RL☆1,002Updated last week
- SEED-Voken: A Series of Powerful Visual Tokenizers☆922Updated last month
- [NeurIPS 2024] Simple and Effective Masked Diffusion Language Model☆466Updated 2 months ago
- [ICML 2024 Best Paper] Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution (https://arxiv.org/abs/2310.16834)☆612Updated last year
- Scaling Diffusion Transformers with Mixture of Experts☆356Updated 10 months ago
- This repo contains the code for 1D tokenizer and generator☆975Updated 4 months ago
- HART: Efficient Visual Generation with Hybrid Autoregressive Transformer☆621Updated 9 months ago
- Implementation of the sparse attention pattern proposed by the Deepseek team in their "Native Sparse Attention" paper☆700Updated last month
- Pytorch implementation of Transfusion, "Predict the Next Token and Diffuse Images with One Multi-Modal Model", from MetaAI☆1,184Updated last month
- Long-RL: Scaling RL to Long Sequences☆568Updated this week
- [ICLR'25 Oral] Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think☆1,251Updated 4 months ago
- Implementation of a single layer of the MMDiT, proposed in Stable Diffusion 3, in Pytorch☆401Updated 6 months ago
- H-Net: Hierarchical Network with Dynamic Chunking☆632Updated last week
- A Distributed Attention Towards Linear Scalability for Ultra-Long Context, Heterogeneous Data Training☆456Updated this week
- FlexTok: Resampling Images into 1D Token Sequences of Flexible Length☆234Updated 2 months ago
- [ICLR 2025] VILA-U: a Unified Foundation Model Integrating Visual Understanding and Generation☆374Updated 3 months ago
- [ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation☆820Updated 10 months ago
- Implementation of Autoregressive Diffusion in Pytorch☆399Updated 9 months ago
- Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"☆320Updated this week
- [Survey] Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey☆446Updated 6 months ago
- Muon is an optimizer for hidden layers in neural networks☆1,454Updated 3 weeks ago
- Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation☆367Updated this week
- Muon is Scalable for LLM Training☆1,240Updated this week