google / paxmlLinks
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆538Updated last month
Alternatives and similar repositories for paxml
Users that are interested in paxml are comparing it to the libraries listed below
Sorting:
- ☆189Updated 3 weeks ago
- jax-triton contains integrations between JAX and OpenAI Triton☆426Updated last week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆669Updated this week
- ☆332Updated last month
- ☆145Updated this week
- JAX-Toolbox☆348Updated last week
- Library for reading and processing ML training data.☆567Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆436Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆381Updated 4 months ago
- ☆538Updated last year
- ☆362Updated last year
- JAX Synergistic Memory Inspector☆179Updated last year
- seqax = sequence modeling + JAX☆167Updated 2 months ago
- ☆283Updated last year
- JAX implementation of the Llama 2 model☆216Updated last year
- Implementation of Flash Attention in Jax☆218Updated last year
- Inference code for LLaMA models in JAX☆119Updated last year
- ☆358Updated last week
- Train very large language models in Jax.☆209Updated last year
- ☆268Updated this week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆394Updated last week
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆400Updated last week
- CLU lets you write beautiful training loops in JAX.☆356Updated 3 months ago
- ☆23Updated last week
- Swarm training framework using Haiku + JAX + Ray for layer parallel transformer language models on unreliable, heterogeneous nodes☆241Updated 2 years ago
- Implementation of a Transformer, but completely in Triton☆275Updated 3 years ago
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆415Updated last week
- A library to analyze PyTorch traces.☆416Updated 2 weeks ago
- JMP is a Mixed Precision library for JAX.☆207Updated 8 months ago
- Named Tensors for Legible Deep Learning in JAX☆207Updated this week