proger / nanokitchen
Parallel Associative Scan for Language Models
☆18Updated last year
Alternatives and similar repositories for nanokitchen:
Users that are interested in nanokitchen are comparing it to the libraries listed below
- Blog post☆17Updated last year
- Efficient PScan implementation in PyTorch☆16Updated last year
- ☆32Updated last year
- ☆31Updated last year
- ☆46Updated last year
- ☆39Updated last year
- ☆30Updated 5 months ago
- Triton Implementation of HyperAttention Algorithm☆47Updated last year
- ☆32Updated 6 months ago
- Source-to-Source Debuggable Derivatives in Pure Python☆15Updated last year
- ☆12Updated last month
- ☆51Updated 11 months ago
- FlexAttention w/ FlashAttention3 Support☆26Updated 6 months ago
- ☆52Updated 6 months ago
- Awesome Triton Resources☆26Updated 3 weeks ago
- Minimal but scalable implementation of large language models in JAX☆34Updated 5 months ago
- A PyTorch wrapper of parallel exclusive scan in CUDA☆12Updated last year
- Code for the note "NF4 Isn't Information Theoretically Optimal (and that's Good)☆18Updated last year
- Code for the paper: https://arxiv.org/pdf/2309.06979.pdf☆19Updated 8 months ago
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32Updated 10 months ago
- ☆29Updated 2 years ago
- Experiments on the impact of depth in transformers and SSMs.☆25Updated 5 months ago
- Combining SOAP and MUON☆15Updated 2 months ago
- Experiment of using Tangent to autodiff triton☆78Updated last year
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆36Updated last year
- sigma-MoE layer☆18Updated last year
- ☆23Updated 6 months ago
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆26Updated last year
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆83Updated last year
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated last year