srush / annotated-s4
Implementation of https://srush.github.io/annotated-s4
☆494Updated 2 years ago
Alternatives and similar repositories for annotated-s4:
Users that are interested in annotated-s4 are comparing it to the libraries listed below
- ☆289Updated 3 months ago
- Annotated version of the Mamba paper☆483Updated last year
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆672Updated 5 months ago
- Long Range Arena for Benchmarking Efficient Transformers☆751Updated last year
- Structured state space sequence models☆2,620Updated 9 months ago
- For optimization algorithm research and development.☆509Updated this week
- ☆164Updated 2 years ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆569Updated this week
- Sequence modeling with Mega.☆294Updated 2 years ago
- Accelerated First Order Parallel Associative Scan☆182Updated 8 months ago
- ☆349Updated last year
- ☆178Updated 11 months ago
- ☆217Updated 9 months ago
- Neural Networks and the Chomsky Hierarchy☆206Updated last year
- Puzzles for exploring transformers☆344Updated 2 years ago
- Language Modeling with the H3 State Space Model☆520Updated last year
- Unofficial JAX implementations of deep learning research papers☆156Updated 2 years ago
- Implementation of Block Recurrent Transformer - Pytorch☆217Updated 8 months ago
- Transformer based on a variant of attention that is linear complexity in respect to sequence length☆763Updated last year
- ☆166Updated last year
- ☆430Updated 6 months ago
- ☆256Updated 2 years ago
- Understand and test language model architectures on synthetic tasks.☆194Updated last month
- CLU lets you write beautiful training loops in JAX.☆337Updated 3 weeks ago
- ☆776Updated last week
- An implementation of local windowed attention for language modeling☆445Updated 3 months ago
- Helpful tools and examples for working with flex-attention☆746Updated 3 weeks ago
- Simple, minimal implementation of the Mamba SSM in one pytorch file. Using logcumsumexp (Heisen sequence).☆114Updated 6 months ago
- ☆177Updated 5 months ago
- VQVAEs, GumbelSoftmaxes and friends☆560Updated 3 years ago