glassroom / heinsen_sequence
Code implementing "Efficient Parallelization of a Ubiquitious Sequential Computation" (Heinsen, 2023)
☆89Updated last month
Alternatives and similar repositories for heinsen_sequence:
Users that are interested in heinsen_sequence are comparing it to the libraries listed below
- seqax = sequence modeling + JAX☆136Updated 6 months ago
- ☆53Updated last year
- ☆201Updated 6 months ago
- Accelerated First Order Parallel Associative Scan☆169Updated 5 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆102Updated last month
- ☆37Updated last year
- ☆50Updated 3 months ago
- Understand and test language model architectures on synthetic tasks.☆175Updated this week
- Experiment of using Tangent to autodiff triton☆74Updated 11 months ago
- Pytorch implementation of preconditioned stochastic gradient descent (Kron and affine preconditioner, low-rank approximation precondition…☆148Updated last month
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆82Updated 11 months ago
- A simple library for scaling up JAX programs☆129Updated 2 months ago
- supporting pytorch FSDP for optimizers☆75Updated last month
- Parallel Associative Scan for Language Models☆18Updated last year
- Named Tensors for Legible Deep Learning in JAX☆159Updated last week
- Implementation of PSGD optimizer in JAX☆26Updated 2 weeks ago
- A library for unit scaling in PyTorch☆118Updated last month
- LoRA for arbitrary JAX models and functions☆135Updated 10 months ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆87Updated 7 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆111Updated last month
- ☆146Updated last month
- Solve puzzles. Learn CUDA.☆61Updated last year
- 🧱 Modula software package☆132Updated this week
- ☆58Updated 2 years ago
- nanoGPT-like codebase for LLM training☆83Updated this week
- Efficient optimizers☆145Updated this week
- gzip Predicts Data-dependent Scaling Laws☆33Updated 7 months ago
- Muon optimizer for neural networks: >30% extra sample efficiency, <3% wallclock overhead☆210Updated 2 weeks ago
- Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"☆219Updated last month