JinjieNi / QuokkaLinks
The official github repo for "Training Optimal Large Diffusion Language Models", the first-ever large-scale diffusion language models scaling law..
☆41Updated last week
Alternatives and similar repositories for Quokka
Users that are interested in Quokka are comparing it to the libraries listed below
Sorting:
- Official repository for paper "DeepCritic: Deliberate Critique with Large Language Models"☆41Updated 4 months ago
- Optimizing Anytime Reasoning via Budget Relative Policy Optimization☆47Updated 4 months ago
- V1: Toward Multimodal Reasoning by Designing Auxiliary Task☆36Updated 7 months ago
- ☆106Updated last month
- The official implementation for [NeurIPS2025 Oral] Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink…☆101Updated last month
- paper list, tutorial, and nano code snippet for Diffusion Large Language Models.☆127Updated 4 months ago
- ACL'2025: SoftCoT: Soft Chain-of-Thought for Efficient Reasoning with LLMs. and preprint: SoftCoT++: Test-Time Scaling with Soft Chain-of…☆61Updated 5 months ago
- A Collection of Papers on Diffusion Language Models☆137Updated 2 months ago
- Official Repository of LatentSeek☆67Updated 5 months ago
- [NeurIPS 2025] NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆95Updated last month
- Diffusion Language Models For Code Infilling Beyond Fixed-size Canvas☆88Updated 2 months ago
- The official repository for the paper "ThinkMorph: Emergent Properties in Multimodal Interleaved Chain-of-Thought Reasoning"☆90Updated last week
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆46Updated last year
- Laser: Learn to Reason Efficiently with Adaptive Length-based Reward Shaping☆57Updated 5 months ago
- ☆45Updated last month
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoning☆69Updated 4 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆80Updated 4 months ago
- repo for paper https://arxiv.org/abs/2504.13837☆217Updated 4 months ago
- The official code of "VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning" [NeurIPS25]☆164Updated 5 months ago
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆134Updated 4 months ago
- [NeurIPS 2024] Code for the paper "Diffusion of Thoughts: Chain-of-Thought Reasoning in Diffusion Language Models"☆182Updated 8 months ago
- ☆46Updated 7 months ago
- [NeurIPS 2025] Think Silently, Think Fast: Dynamic Latent Compression of LLM Reasoning Chains☆59Updated 3 months ago
- The official implementation of "LightTransfer: Your Long-Context LLM is Secretly a Hybrid Model with Effortless Adaptation"☆20Updated 6 months ago
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆104Updated last month
- ☆61Updated 3 weeks ago
- [NeurIPS'25] dKV-Cache: The Cache for Diffusion Language Models☆114Updated 5 months ago
- ☆72Updated last month
- 📖 This is a repository for organizing papers, codes, and other resources related to Latent Reasoning.☆269Updated last week
- TraceRL & TraDo-8B: Revolutionizing Reinforcement Learning Framework for Diffusion Large Language Models☆307Updated 3 weeks ago