fattorib / transformer_shmapLinks
Tensor Parallelism with JAX + Shard Map
☆11Updated last year
Alternatives and similar repositories for transformer_shmap
Users that are interested in transformer_shmap are comparing it to the libraries listed below
Sorting:
- ESM2 protein language models in JAX/Flax☆17Updated 2 years ago
- Parallel Associative Scan for Language Models☆18Updated last year
- Turn jitted jax functions back into python source code☆22Updated 5 months ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- JMP is a Mixed Precision library for JAX.☆199Updated 4 months ago
- Jax/Flax rewrite of Karpathy's nanoGPT☆57Updated 2 years ago
- A port of muP to JAX/Haiku☆25Updated 2 years ago
- This is a port of Mistral-7B model in JAX☆32Updated 11 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆44Updated 10 months ago
- Einsum-like high-level array sharding API for JAX☆34Updated 10 months ago
- ☆118Updated 2 weeks ago
- A simple library for scaling up JAX programs☆137Updated 7 months ago
- Lightning-like training API for JAX with Flax☆38Updated 5 months ago
- Dive into Jax, Flax, XLA and C++☆31Updated 5 years ago
- Blog post☆17Updated last year
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆81Updated last year
- AdamW optimizer for bfloat16 models in pytorch 🔥.☆32Updated 11 months ago
- A selection of neural network models ported from torchvision for JAX & Flax.☆44Updated 4 years ago
- ☆59Updated 3 years ago
- Minimal but scalable implementation of large language models in JAX☆34Updated 7 months ago
- Personal solutions to the Triton Puzzles☆18Updated 10 months ago
- ☆31Updated last month
- ☆24Updated 6 years ago
- If it quacks like a tensor...☆58Updated 6 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆33Updated last year
- ☆32Updated 8 months ago
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated 2 years ago
- JAX implementation of the Mistral 7b v0.2 model☆34Updated 11 months ago
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated 2 years ago
- A flexible and efficient implementation of Flash Attention 2.0 for JAX, supporting multiple backends (GPU/TPU/CPU) and platforms (Triton/…☆24Updated 3 months ago