srush / annotated-s4Links
Implementation of https://srush.github.io/annotated-s4
☆498Updated 2 years ago
Alternatives and similar repositories for annotated-s4
Users that are interested in annotated-s4 are comparing it to the libraries listed below
Sorting:
- Annotated version of the Mamba paper☆485Updated last year
- ☆292Updated 5 months ago
- ☆178Updated last year
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆385Updated this week
- For optimization algorithm research and development.☆521Updated this week
- ☆270Updated 11 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆595Updated this week
- Code for our NeurIPS 2022 paper☆368Updated 2 years ago
- Long Range Arena for Benchmarking Efficient Transformers☆757Updated last year
- Structured state space sequence models☆2,655Updated 11 months ago
- Language Modeling with the H3 State Space Model☆519Updated last year
- Sequence modeling with Mega.☆296Updated 2 years ago
- Accelerated First Order Parallel Associative Scan☆182Updated 10 months ago
- ☆188Updated 6 months ago
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆694Updated 6 months ago
- ☆355Updated last year
- ☆778Updated 2 weeks ago
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆251Updated 3 months ago
- Puzzles for exploring transformers☆349Updated 2 years ago
- Named tensors with first-class dimensions for PyTorch☆331Updated 2 years ago
- This is the official code release for Bayesian Flow Networks.☆281Updated 11 months ago
- Neural Networks and the Chomsky Hierarchy☆205Updated last year
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆120Updated 8 months ago
- FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores☆319Updated 5 months ago
- Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"☆379Updated last year
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆340Updated last year
- ☆292Updated 6 months ago
- ☆163Updated 2 years ago
- A curated list of papers of interesting empirical study and insight on deep learning. Continually updating...