ayaka14732 / tpu-starter
Everything you want to know about Google Cloud TPU
☆506Updated 6 months ago
Alternatives and similar repositories for tpu-starter:
Users that are interested in tpu-starter are comparing it to the libraries listed below
- For optimization algorithm research and development.☆486Updated last week
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆536Updated this week
- JAX Synergistic Memory Inspector☆165Updated 6 months ago
- Puzzles for exploring transformers☆331Updated last year
- Helpful tools and examples for working with flex-attention☆603Updated this week
- JAX implementation of the Llama 2 model☆213Updated 11 months ago
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆511Updated this week
- What would you do with 1000 H100s...☆970Updated last year
- ☆413Updated 3 months ago
- Annotated version of the Mamba paper☆470Updated 11 months ago
- ☆203Updated 6 months ago
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆351Updated this week
- ☆336Updated 9 months ago
- ☆296Updated 7 months ago
- jax-triton contains integrations between JAX and OpenAI Triton☆371Updated last week
- Named tensors with first-class dimensions for PyTorch☆321Updated last year
- Inference code for LLaMA models in JAX☆114Updated 8 months ago
- ☆278Updated last week
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,019Updated 9 months ago
- Language Modeling with the H3 State Space Model☆515Updated last year
- Implementation of Flash Attention in Jax☆204Updated 10 months ago
- JAX-Toolbox☆279Updated this week
- Puzzles for learning Triton☆1,337Updated 2 months ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆670Updated this week
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆475Updated last week
- Implementation of a Transformer, but completely in Triton☆253Updated 2 years ago
- maximal update parametrization (µP)☆1,438Updated 6 months ago
- Effortless plugin and play Optimizer to cut model training costs by 50%. New optimizer that is 2x faster than Adam on LLMs.☆378Updated 7 months ago
- CLU lets you write beautiful training loops in JAX.☆329Updated this week
- The official implementation of “Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training”☆945Updated last year