srush / annotated-mamba
Annotated version of the Mamba paper
☆469Updated 10 months ago
Alternatives and similar repositories for annotated-mamba:
Users that are interested in annotated-mamba are comparing it to the libraries listed below
- Helpful tools and examples for working with flex-attention☆583Updated this week
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆505Updated 2 months ago
- Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI☆270Updated 2 months ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆492Updated 2 months ago
- For optimization algorithm research and development.☆484Updated this week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆534Updated this week
- Implementation of https://srush.github.io/annotated-s4☆477Updated last year
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆297Updated 7 months ago
- Implementation of Diffusion Transformer (DiT) in JAX☆261Updated 7 months ago
- Puzzles for exploring transformers☆331Updated last year
- Some preliminary explorations of Mamba's context scaling.☆206Updated 11 months ago
- Reading list for research topics in state-space models☆253Updated 3 weeks ago
- ☆201Updated 6 months ago
- Code repository for Black Mamba☆234Updated 11 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆253Updated last year
- Muon optimizer for neural networks: >30% extra sample efficiency, <3% wallclock overhead☆210Updated last week
- Building blocks for foundation models.☆435Updated last year
- Minimalistic 4D-parallelism distributed training framework for education purpose☆644Updated this week
- Open weights language model from Google DeepMind, based on Griffin.☆614Updated 6 months ago
- ☆285Updated last month
- Understand and test language model architectures on synthetic tasks.☆175Updated this week
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆219Updated last month
- ☆146Updated last month
- ☆296Updated 6 months ago
- What would you do with 1000 H100s...☆948Updated last year
- Repo for "Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture"☆541Updated 2 weeks ago
- Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆477Updated this week
- Accelerated First Order Parallel Associative Scan☆169Updated 4 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆102Updated last month
- Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"☆370Updated last year