google / paxmlLinks
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆548Updated 2 weeks ago
Alternatives and similar repositories for paxml
Users that are interested in paxml are comparing it to the libraries listed below
Sorting:
- ☆192Updated 2 weeks ago
- ☆151Updated 3 weeks ago
- jax-triton contains integrations between JAX and OpenAI Triton☆437Updated last month
- ☆344Updated 3 weeks ago
- JAX-Toolbox☆381Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆478Updated this week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆694Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆403Updated 3 weeks ago
- Library for reading and processing ML training data.☆670Updated last week
- ☆555Updated last year
- seqax = sequence modeling + JAX☆170Updated 6 months ago
- ☆367Updated last year
- JAX Synergistic Memory Inspector☆184Updated last year
- JAX implementation of the Llama 2 model☆216Updated last year
- ☆289Updated last year
- Minimal yet performant LLM examples in pure JAX☆233Updated 2 weeks ago
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆333Updated 3 weeks ago
- CLU lets you write beautiful training loops in JAX.☆366Updated last week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆411Updated this week
- Implementation of Flash Attention in Jax☆225Updated last year
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆472Updated 2 weeks ago
- ☆300Updated this week
- Train very large language models in Jax.☆210Updated 2 years ago
- ☆374Updated 2 weeks ago
- JMP is a Mixed Precision library for JAX.☆211Updated last year
- Inference code for LLaMA models in JAX☆120Updated last year
- Pipeline Parallelism for PyTorch☆785Updated last year
- Everything you want to know about Google Cloud TPU☆556Updated last year
- xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerat…☆162Updated last week
- Implementation of a Transformer, but completely in Triton☆279Updated 3 years ago