RobinWu218 / ToSTLinks
[ICLR 2025 Spotlight] Official Implementation for ToST (Token Statistics Transformer)
☆130Updated 11 months ago
Alternatives and similar repositories for ToST
Users that are interested in ToST are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2025 Spotlight] TPA: Tensor ProducT ATTenTion Transformer (T6) (https://arxiv.org/abs/2501.06425)☆445Updated last week
- [ICLR 2025 & COLM 2025] Official PyTorch implementation of the Forgetting Transformer and Adaptive Computation Pruning☆137Updated last month
- [ICLR 2025] Official PyTorch Implementation of Gated Delta Networks: Improving Mamba2 with Delta Rule☆433Updated 4 months ago
- [ICML 2025] Fourier Position Embedding: Enhancing Attention’s Periodic Extension for Length Generalization☆106Updated 7 months ago
- The official implementation for [NeurIPS2025 Oral] Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink…☆806Updated last month
- Official implementation of "Hydra: Bidirectional State Space Models Through Generalized Matrix Mixers"☆170Updated 11 months ago
- Awesome list of papers that extend Mamba to various applications.☆138Updated 7 months ago
- [CVPR2025] Breaking the Low-Rank Dilemma of Linear Attention☆38Updated 10 months ago
- ☆268Updated 7 months ago
- implementations and experimentation on mHC by deepseek - https://arxiv.org/abs/2512.24880☆265Updated 3 weeks ago
- A Triton Kernel for incorporating Bi-Directionality in Mamba2☆76Updated last year
- Implementation of the proposed MaskBit from Bytedance AI☆83Updated last year
- [CVPR'25] MergeVQ: A Unified Framework for Visual Generation and Representation with Token Merging and Quantization☆47Updated 6 months ago
- ☆201Updated 2 years ago
- Implementation of the proposed DeepCrossAttention by Heddes et al at Google research, in Pytorch☆96Updated 11 months ago
- [NeurIPS 2025 Oral] Official Code for Exploring Diffusion Transformer Designs via Grafting☆70Updated 2 weeks ago
- Implementation of the dynamic chunking mechanism in H-net by Hwang et al. of Carnegie Mellon☆66Updated 2 weeks ago
- Implementation of TiTok, proposed by Bytedance in "An Image is Worth 32 Tokens for Reconstruction and Generation"☆183Updated last year
- ☆79Updated 11 months ago
- When it comes to optimizers, it's always better to be safe than sorry☆399Updated 4 months ago
- Discrete Diffusion Forcing (D2F): dLLMs Can Do Faster-Than-AR Inference☆238Updated last week
- A repository for DenseSSMs☆88Updated last year
- Towards training VQ-VAE models robustly!☆91Updated 6 months ago
- Code for the paper "Vamba: Understanding Hour-Long Videos with Hybrid Mamba-Transformers" [ICCV 2025]☆99Updated 5 months ago
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆339Updated 11 months ago
- Inference Speed Benchmark for Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆80Updated last year
- Explore how to get a VQ-VAE models efficiently!☆67Updated 6 months ago
- Pytorch Implementation of the sparse attention from the paper: "Generating Long Sequences with Sparse Transformers"☆93Updated last week
- LLaDA2.0 is the diffusion language model series developed by InclusionAI team, Ant Group.☆225Updated last month
- Triton implement of bi-directional (non-causal) linear attention☆63Updated 11 months ago