srush / annotated-s4Links
Implementation of https://srush.github.io/annotated-s4
☆510Updated 6 months ago
Alternatives and similar repositories for annotated-s4
Users that are interested in annotated-s4 are comparing it to the libraries listed below
Sorting:
- ☆314Updated last year
- Annotated version of the Mamba paper☆494Updated last year
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆690Updated this week
- ☆287Updated last year
- For optimization algorithm research and development.☆556Updated 3 weeks ago
- ☆366Updated last year
- Language Modeling with the H3 State Space Model☆521Updated 2 years ago
- Accelerated First Order Parallel Associative Scan☆193Updated last week
- Long Range Arena for Benchmarking Efficient Transformers☆772Updated 2 years ago
- ☆192Updated last year
- Code for our NeurIPS 2022 paper☆371Updated 3 years ago
- ☆259Updated 7 months ago
- Named tensors with first-class dimensions for PyTorch☆332Updated 2 years ago
- Neural Networks and the Chomsky Hierarchy☆211Updated last year
- ☆463Updated last year
- JAX Synergistic Memory Inspector☆183Updated last year
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆406Updated last week
- ☆792Updated this week
- Sequence modeling with Mega.☆302Updated 2 years ago
- ☆164Updated 2 years ago
- maximal update parametrization (µP)☆1,657Updated last year
- Betty: an automatic differentiation library for generalized meta-learning and multilevel optimization☆344Updated last year
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆128Updated last year
- Puzzles for exploring transformers☆382Updated 2 years ago
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆788Updated 5 months ago
- A repository for log-time feedforward networks☆224Updated last year
- ☆233Updated 11 months ago
- ☆312Updated this week
- TensorDict is a pytorch dedicated tensor container.☆1,002Updated this week
- An interpreter for RASP as described in the ICML 2021 paper "Thinking Like Transformers"☆323Updated last year