srush / annotated-s4Links
Implementation of https://srush.github.io/annotated-s4
☆498Updated 3 weeks ago
Alternatives and similar repositories for annotated-s4
Users that are interested in annotated-s4 are comparing it to the libraries listed below
Sorting:
- ☆295Updated 6 months ago
- Annotated version of the Mamba paper☆486Updated last year
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆607Updated this week
- Long Range Arena for Benchmarking Efficient Transformers☆759Updated last year
- ☆178Updated last year
- ☆356Updated last year
- Code for our NeurIPS 2022 paper☆369Updated 2 years ago
- Language Modeling with the H3 State Space Model☆519Updated last year
- Sequence modeling with Mega.☆296Updated 2 years ago
- ☆256Updated last month
- ☆273Updated 11 months ago
- Accelerated First Order Parallel Associative Scan☆182Updated 10 months ago
- Named tensors with first-class dimensions for PyTorch☆332Updated 2 years ago
- ☆163Updated 2 years ago
- For optimization algorithm research and development.☆521Updated this week
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆386Updated last week
- Neural Networks and the Chomsky Hierarchy☆206Updated last year
- ☆229Updated 5 months ago
- A PyTorch implementation of Perceiver, Perceiver IO and Perceiver AR with PyTorch Lightning scripts for distributed training☆482Updated last year
- Betty: an automatic differentiation library for generalized meta-learning and multilevel optimization☆338Updated last year
- ☆780Updated last month
- Puzzles for exploring transformers☆354Updated 2 years ago
- Implementation of Block Recurrent Transformer - Pytorch☆220Updated 10 months ago
- JAX Synergistic Memory Inspector☆175Updated 11 months ago
- maximal update parametrization (µP)☆1,554Updated 11 months ago
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆345Updated last year
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆707Updated this week
- Structured state space sequence models☆2,678Updated 11 months ago
- ☆195Updated 7 months ago
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆120Updated 8 months ago