jax-ml / scaling-bookLinks
Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs
☆424Updated this week
Alternatives and similar repositories for scaling-book
Users that are interested in scaling-book are comparing it to the libraries listed below
Sorting:
- ☆273Updated 11 months ago
- ☆132Updated last week
- ☆511Updated last year
- PyTorch Single Controller☆296Updated this week
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆134Updated last year
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆607Updated this week
- ☆440Updated 8 months ago
- seqax = sequence modeling + JAX☆163Updated 3 weeks ago
- jax-triton contains integrations between JAX and OpenAI Triton☆405Updated 2 weeks ago
- Implementation of Diffusion Transformer (DiT) in JAX☆279Updated last year
- JAX-Toolbox☆320Updated this week
- A simple library for scaling up JAX programs☆139Updated 8 months ago
- 🧱 Modula software package☆202Updated 3 months ago
- Puzzles for exploring transformers☆354Updated 2 years ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆561Updated 3 weeks ago
- ☆198Updated 4 months ago
- Simple MPI implementation for prototyping or learning☆259Updated 2 weeks ago
- A repository to unravel the language of GPUs, making their kernel conversations easy to understand☆188Updated last month
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆280Updated this week
- For optimization algorithm research and development.☆521Updated this week
- ☆259Updated this week
- ☆160Updated last year
- Named Tensors for Legible Deep Learning in JAX☆185Updated this week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆359Updated last week
- Building blocks for foundation models.☆515Updated last year
- Orbax provides common checkpointing and persistence utilities for JAX users☆404Updated this week
- ☆320Updated 2 weeks ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆290Updated 10 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆468Updated this week
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆386Updated last week