google / paxml
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆456Updated last week
Related projects ⓘ
Alternatives and complementary repositories for paxml
- ☆177Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆340Updated last week
- JAX-Toolbox☆241Updated this week
- ☆223Updated 3 months ago
- ☆262Updated this week
- ☆121Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆296Updated this week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆516Updated this week
- ☆333Updated 6 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆228Updated this week
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆479Updated 2 weeks ago
- Pipeline Parallelism for PyTorch☆725Updated 2 months ago
- seqax = sequence modeling + JAX☆132Updated 3 months ago
- Implementation of a Transformer, but completely in Triton☆248Updated 2 years ago
- A library to analyze PyTorch traces.☆297Updated this week
- This repository contains the experimental PyTorch native float8 training UX☆211Updated 3 months ago
- ☆296Updated this week
- Backward compatible ML compute opset inspired by HLO/MHLO☆408Updated this week
- ☆197Updated 3 months ago
- Implementation of Flash Attention in Jax☆194Updated 8 months ago
- Inference code for LLaMA models in JAX☆112Updated 5 months ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆146Updated this week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆332Updated 2 weeks ago
- JAX implementation of the Llama 2 model☆210Updated 9 months ago
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆330Updated last week
- 🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.☆163Updated this week
- CLU lets you write beautiful training loops in JAX.☆321Updated 2 months ago
- An open-source efficient deep learning framework/compiler, written in python.☆649Updated this week
- ☆147Updated 4 months ago
- ☆140Updated this week