zinccat / flaxattention
☆16Updated 6 months ago
Alternatives and similar repositories for flaxattention
Users that are interested in flaxattention are comparing it to the libraries listed below
Sorting:
- Experiment of using Tangent to autodiff triton☆78Updated last year
- A flexible and efficient implementation of Flash Attention 2.0 for JAX, supporting multiple backends (GPU/TPU/CPU) and platforms (Triton/…☆23Updated 2 months ago
- ☆21Updated 2 months ago
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated 7 months ago
- train with kittens!☆57Updated 6 months ago
- Make triton easier☆47Updated 11 months ago
- A set of Python scripts that makes your experience on TPU better☆53Updated 10 months ago
- ☆59Updated 3 years ago
- ☆17Updated 8 months ago
- If it quacks like a tensor...☆58Updated 6 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 10 months ago
- JAX bindings for Flash Attention v2☆88Updated 10 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆44Updated this week
- ☆79Updated 10 months ago
- Code for the paper: https://arxiv.org/pdf/2309.06979.pdf☆19Updated 9 months ago
- Jax/Flax rewrite of Karpathy's nanoGPT☆57Updated 2 years ago
- supporting pytorch FSDP for optimizers☆80Updated 5 months ago
- Accelerated First Order Parallel Associative Scan☆182Updated 8 months ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆30Updated this week
- Simple and efficient pytorch-native transformer training and inference (batched)☆75Updated last year
- A simple library for scaling up JAX programs☆134Updated 6 months ago
- Machine Learning eXperiment Utilities☆46Updated 11 months ago
- NanoGPT-speedrunning for the poor T4 enjoyers☆65Updated 3 weeks ago
- LayerNorm(SmallInit(Embedding)) in a Transformer to improve convergence☆60Updated 3 years ago
- Experimenting with how best to do multi-host dataloading☆10Updated 2 years ago
- Two implementations of ZeRO-1 optimizer sharding in JAX☆14Updated last year
- This is a port of Mistral-7B model in JAX☆32Updated 10 months ago
- Griffin MQA + Hawk Linear RNN Hybrid☆86Updated last year
- Parallel Associative Scan for Language Models☆18Updated last year
- Serialize JAX, Flax, Haiku, or Objax model params with 🤗`safetensors`☆44Updated 11 months ago