google / paxmlLinks
Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
☆499Updated 2 weeks ago
Alternatives and similar repositories for paxml
Users that are interested in paxml are comparing it to the libraries listed below
Sorting:
- ☆186Updated this week
- ☆310Updated 2 weeks ago
- ☆138Updated 2 weeks ago
- JAX-Toolbox☆308Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆395Updated this week
- Orbax provides common checkpointing and persistence utilities for JAX users☆384Updated this week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆586Updated last week
- Library for reading and processing ML training data.☆447Updated last week
- ☆487Updated 10 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆335Updated this week
- ☆354Updated last year
- seqax = sequence modeling + JAX☆155Updated last month
- ☆348Updated 3 weeks ago
- ☆267Updated 10 months ago
- CLU lets you write beautiful training loops in JAX.☆343Updated last month
- Implementation of Flash Attention in Jax☆212Updated last year
- JAX implementation of the Llama 2 model☆217Updated last year
- JAX Synergistic Memory Inspector☆173Updated 10 months ago
- Inference code for LLaMA models in JAX☆117Updated last year
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆546Updated this week
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆268Updated last week
- Train very large language models in Jax.☆204Updated last year
- JMP is a Mixed Precision library for JAX.☆199Updated 4 months ago
- A simple library for scaling up JAX programs☆137Updated 7 months ago
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆366Updated 2 weeks ago
- Implementation of a Transformer, but completely in Triton☆266Updated 3 years ago
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,046Updated last year
- ☆228Updated 3 months ago
- xpk (Accelerated Processing Kit, pronounced x-p-k,) is a software tool to help Cloud developers to orchestrate training jobs on accelerat…☆123Updated this week
- Open weights language model from Google DeepMind, based on Griffin.☆639Updated 2 weeks ago