google / paxml
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆475Updated last week
Alternatives and similar repositories for paxml:
Users that are interested in paxml are comparing it to the libraries listed below
- ☆181Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆371Updated last week
- ☆130Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆328Updated this week
- ☆278Updated last week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆536Updated this week
- ☆273Updated 6 months ago
- JAX-Toolbox☆279Updated this week
- ☆203Updated 6 months ago
- ☆336Updated 9 months ago
- Library for reading and processing ML training data.☆371Updated this week
- seqax = sequence modeling + JAX☆136Updated 6 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆268Updated this week
- JAX implementation of the Llama 2 model☆213Updated 11 months ago
- Inference code for LLaMA models in JAX☆114Updated 8 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆511Updated this week
- CLU lets you write beautiful training loops in JAX.☆329Updated this week
- Implementation of a Transformer, but completely in Triton☆253Updated 2 years ago
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆341Updated last week
- Task-based datasets, preprocessing, and evaluation for sequence models.☆568Updated this week
- Train very large language models in Jax.☆198Updated last year
- Pipeline Parallelism for PyTorch☆739Updated 5 months ago
- This repository contains the experimental PyTorch native float8 training UX☆219Updated 5 months ago
- For optimization algorithm research and development.☆486Updated last week
- JAX Synergistic Memory Inspector☆165Updated 6 months ago
- ☆339Updated 2 weeks ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆216Updated this week
- Large Context Attention☆677Updated this week
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆153Updated last month
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆351Updated this week