sustcsonglin / flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
☆1,339Updated this week
Related projects ⓘ
Alternatives and complementary repositories for flash-linear-attention
- Helpful tools and examples for working with flex-attention☆469Updated 3 weeks ago
- Annotated version of the Mamba paper☆457Updated 8 months ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆476Updated 3 weeks ago
- Puzzles for learning Triton☆1,135Updated this week
- Mirage: Automatically Generating Fast GPU Kernels without Programming in Triton/CUDA☆636Updated this week
- A simple and efficient Mamba implementation in pure PyTorch and MLX.☆1,012Updated 2 months ago
- Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆1,040Updated 4 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆483Updated 3 weeks ago
- [ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed Low-Rank Adaptation☆633Updated last month
- Ring attention implementation with flash attention☆585Updated last week
- Building blocks for foundation models.☆394Updated 10 months ago
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆537Updated 6 months ago
- Code for Adam-mini: Use Fewer Learning Rates To Gain More https://arxiv.org/abs/2406.16793☆328Updated 3 weeks ago
- Tile primitives for speedy kernels☆1,658Updated this week
- Schedule-Free Optimization in PyTorch☆1,898Updated 2 weeks ago
- Tutel MoE: An Optimized Mixture-of-Experts Implementation☆735Updated this week
- Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆335Updated last week
- Transformers with Arbitrarily Large Context☆641Updated 3 months ago
- Flash Attention in ~100 lines of CUDA (forward pass only)☆626Updated 7 months ago
- Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆366Updated 3 months ago
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆571Updated last week
- A bibliography and survey of the papers surrounding o1☆754Updated this week
- A collection of AWESOME things about mixture-of-experts☆972Updated 3 months ago
- Reading list for research topics in state-space models☆241Updated 2 weeks ago
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆293Updated 5 months ago
- depyf is a tool to help you understand and adapt to PyTorch compiler torch.compile.☆500Updated 2 weeks ago
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆803Updated 3 months ago
- GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection☆1,435Updated 3 weeks ago
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,149Updated last month