google / paxmlLinks
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆510Updated last week
Alternatives and similar repositories for paxml
Users that are interested in paxml are comparing it to the libraries listed below
Sorting:
- ☆186Updated 3 weeks ago
- ☆141Updated 3 weeks ago
- ☆318Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆403Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆393Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆349Updated 2 weeks ago
- JAX-Toolbox☆314Updated this week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆596Updated this week
- Library for reading and processing ML training data.☆463Updated this week
- ☆270Updated 11 months ago
- ☆355Updated last year
- ☆504Updated 11 months ago
- ☆350Updated last month
- seqax = sequence modeling + JAX☆162Updated last week
- CLU lets you write beautiful training loops in JAX.☆346Updated this week
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆556Updated last week
- JAX implementation of the Llama 2 model☆219Updated last year
- JAX Synergistic Memory Inspector☆174Updated 11 months ago
- For optimization algorithm research and development.☆521Updated this week
- Implementation of a Transformer, but completely in Triton☆268Updated 3 years ago
- Train very large language models in Jax.☆205Updated last year
- ☆224Updated this week
- Inference code for LLaMA models in JAX☆118Updated last year
- Task-based datasets, preprocessing, and evaluation for sequence models.☆582Updated last month
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆368Updated last week
- Implementation of Flash Attention in Jax☆213Updated last year
- JMP is a Mixed Precision library for JAX.☆203Updated 4 months ago
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆385Updated this week
- PyTorch per step fault tolerance (actively under development)☆329Updated this week
- This repository contains the experimental PyTorch native float8 training UX☆224Updated 10 months ago