apple / ml-sigmoid-attention
☆278Updated last week
Alternatives and similar repositories for ml-sigmoid-attention:
Users that are interested in ml-sigmoid-attention are comparing it to the libraries listed below
- When it comes to optimizers, it's always better to be safe than sorry☆222Updated last month
- [ICLR 2025] Official PyTorch Implementation of Gated Delta Networks: Improving Mamba2 with Delta Rule☆159Updated last month
- Helpful tools and examples for working with flex-attention☆746Updated 3 weeks ago
- Normalized Transformer (nGPT)☆174Updated 5 months ago
- The official implementation of Tensor ProducT ATTenTion Transformer (T6)☆367Updated 2 weeks ago
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆409Updated 3 weeks ago
- Muon optimizer: +>30% sample efficiency with <3% wallclock overhead☆597Updated last month
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆555Updated 2 months ago
- [NeurIPS 2024] Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models☆215Updated this week
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆232Updated 2 months ago
- Some preliminary explorations of Mamba's context scaling.☆213Updated last year
- ☆181Updated 2 months ago
- Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI☆281Updated last month
- Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch☆286Updated last month
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆231Updated 3 months ago
- Implementation of Infini-Transformer in Pytorch☆110Updated 4 months ago
- Understand and test language model architectures on synthetic tasks.☆194Updated last month
- The AdEMAMix Optimizer: Better, Faster, Older.☆183Updated 7 months ago
- Implementation of the sparse attention pattern proposed by the Deepseek team in their "Native Sparse Attention" paper☆607Updated last month
- Official implementation of Phi-Mamba. A MOHAWK-distilled model (Transformers to SSMs: Distilling Quadratic Knowledge to Subquadratic Mode…☆106Updated 7 months ago
- PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"☆167Updated last month
- Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆407Updated 8 months ago
- TransMLA: Multi-Head Latent Attention Is All You Need☆243Updated this week
- 🔥 A minimal training framework for scaling FLA models☆117Updated this week
- ☆186Updated this week
- ☆434Updated last week
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆282Updated 2 months ago
- Code and weights for the paper "Cluster and Predict Latents Patches for Improved Masked Image Modeling"☆101Updated 3 weeks ago
- ☆176Updated 4 months ago
- Attempt to make multiple residual streams from Bytedance's Hyper-Connections paper accessible to the public☆82Updated 2 months ago