shawntan / stickbreaking-attention
Stick-breaking attention
☆43Updated last month
Alternatives and similar repositories for stickbreaking-attention:
Users that are interested in stickbreaking-attention are comparing it to the libraries listed below
- Code for NeurIPS 2024 Spotlight: "Scaling Laws and Compute-Optimal Training Beyond Fixed Training Durations"☆70Updated 3 months ago
- ☆30Updated 11 months ago
- ☆47Updated last year
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆25Updated 10 months ago
- Xmixers: A collection of SOTA efficient token/channel mixers☆11Updated 3 months ago
- ☆51Updated 9 months ago
- ☆49Updated 7 months ago
- ☆27Updated 3 months ago
- Here we will test various linear attention designs.☆58Updated 9 months ago
- [ICLR 2025] Code for the paper "Beyond Autoregression: Discrete Diffusion for Complex Reasoning and Planning"☆32Updated last week
- Official PyTorch Implementation of the Longhorn Deep State Space Model☆48Updated 2 months ago
- ☆80Updated 11 months ago
- Fast and memory-efficient exact attention☆58Updated this week
- [ICLR2025] DiffuGPT and DiffuLLaMA: Scaling Diffusion Language Models via Adaptation from Autoregressive Models☆89Updated 2 months ago
- Code for paper "Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning"☆65Updated last year
- ☆37Updated 10 months ago
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆35Updated 4 months ago
- ☆28Updated 3 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆77Updated 4 months ago
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆57Updated 3 weeks ago
- Using FlexAttention to compute attention with different masking patterns☆40Updated 4 months ago
- ☆71Updated 6 months ago
- Official code for the paper "Attention as a Hypernetwork"☆23Updated 7 months ago
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆48Updated last year
- Simple and efficient pytorch-native transformer training and inference (batched)☆68Updated 10 months ago
- Official implementation of Phi-Mamba. A MOHAWK-distilled model (Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Mode…☆96Updated 5 months ago
- ☆86Updated last year
- ☆33Updated last year
- Sparse Backpropagation for Mixture-of-Expert Training☆28Updated 7 months ago
- ☆99Updated 11 months ago