nebius / kvaxLinks
A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.
☆128Updated 3 months ago
Alternatives and similar repositories for kvax
Users that are interested in kvax are comparing it to the libraries listed below
Sorting:
- ☆132Updated 2 weeks ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆46Updated 2 weeks ago
- A simple library for scaling up JAX programs☆139Updated 8 months ago
- FlashRNN - Fast RNN Kernels with I/O Awareness☆92Updated last month
- PyTorch Single Controller☆318Updated this week
- Minimal but scalable implementation of large language models in JAX☆35Updated last week
- ☆273Updated last year
- JAX implementation of the Mistral 7b v0.2 model☆35Updated last year
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆424Updated last week
- NanoGPT-speedrunning for the poor T4 enjoyers☆68Updated 2 months ago
- seqax = sequence modeling + JAX☆165Updated last month
- JAX bindings for Flash Attention v2☆90Updated last year
- 🧱 Modula software package☆204Updated 3 months ago
- Accelerated First Order Parallel Associative Scan☆182Updated 10 months ago
- ☆79Updated last year
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆147Updated 2 weeks ago
- Distributed pretraining of large language models (LLMs) on cloud TPU slices, with Jax and Equinox.☆24Updated 9 months ago
- jax-triton contains integrations between JAX and OpenAI Triton☆405Updated 3 weeks ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- DeMo: Decoupled Momentum Optimization☆189Updated 7 months ago
- train with kittens!☆61Updated 8 months ago
- Attention Kernels for Symmetric Power Transformers☆88Updated last week
- Modular, scalable library to train ML models☆135Updated this week
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated 8 months ago
- JAX-Toolbox☆321Updated this week
- Load compute kernels from the Hub☆203Updated this week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆359Updated 2 weeks ago
- Einsum-like high-level array sharding API for JAX☆35Updated last year
- supporting pytorch FSDP for optimizers☆82Updated 7 months ago
- ☆27Updated last year