proger / accelerated-scanLinks
Accelerated First Order Parallel Associative Scan
☆196Updated 3 weeks ago
Alternatives and similar repositories for accelerated-scan
Users that are interested in accelerated-scan are comparing it to the libraries listed below
Sorting:
- A library for unit scaling in PyTorch☆133Updated 6 months ago
- JAX bindings for Flash Attention v2☆103Updated last month
- supporting pytorch FSDP for optimizers☆84Updated last year
- Efficient optimizers☆281Updated last month
- Understand and test language model architectures on synthetic tasks.☆251Updated 2 weeks ago
- A MAD laboratory to improve AI architecture designs 🧪☆135Updated last year
- ☆124Updated last year
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆246Updated 7 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆186Updated last week
- seqax = sequence modeling + JAX☆170Updated 6 months ago
- Griffin MQA + Hawk Linear RNN Hybrid☆88Updated last year
- ☆62Updated last year
- FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores☆340Updated last year
- ☆238Updated last year
- Experiment of using Tangent to autodiff triton☆82Updated 2 years ago
- ☆92Updated last year
- ☆150Updated 2 years ago
- Normalized Transformer (nGPT)☆198Updated last year
- ☆289Updated last year
- ☆40Updated 2 years ago
- A fusion of a linear layer and a cross entropy loss, written for pytorch in triton.☆75Updated last year
- ☆132Updated last week
- 🧱 Modula software package☆322Updated 5 months ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆92Updated last year
- Implementation of the proposed Adam-atan2 from Google Deepmind in Pytorch☆134Updated 3 months ago
- Implementation of Flash Attention in Jax☆225Updated last year
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆229Updated 7 months ago
- Some preliminary explorations of Mamba's context scaling.☆218Updated last year
- ☆83Updated 2 years ago
- Quick implementation of nGPT, learning entirely on the hypersphere, from NvidiaAI☆293Updated 7 months ago