ayaka14732 / tpu-starterLinks
Everything you want to know about Google Cloud TPU
☆527Updated 10 months ago
Alternatives and similar repositories for tpu-starter
Users that are interested in tpu-starter are comparing it to the libraries listed below
Sorting:
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆586Updated last week
- For optimization algorithm research and development.☆518Updated this week
- ☆354Updated last year
- Annotated version of the Mamba paper☆484Updated last year
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆287Updated 9 months ago
- jax-triton contains integrations between JAX and OpenAI Triton☆395Updated this week
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆546Updated this week
- ☆269Updated 10 months ago
- ☆310Updated 2 weeks ago
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆382Updated last week
- JAX Synergistic Memory Inspector☆173Updated 10 months ago
- JAX implementation of the Llama 2 model☆217Updated last year
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆499Updated 2 weeks ago
- Implementation of Flash Attention in Jax☆212Updated last year
- Puzzles for exploring transformers☆348Updated 2 years ago
- ☆303Updated 11 months ago
- ☆228Updated 3 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆256Updated last year
- Named tensors with first-class dimensions for PyTorch☆331Updated last year
- Implementation of a Transformer, but completely in Triton☆266Updated 3 years ago
- Inference code for LLaMA models in JAX☆118Updated last year
- ☆186Updated this week
- Transform datasets at scale. Optimize datasets for fast AI model training.☆485Updated this week
- CLU lets you write beautiful training loops in JAX.☆343Updated last month
- Language Modeling with the H3 State Space Model☆519Updated last year
- Task-based datasets, preprocessing, and evaluation for sequence models.☆574Updated 3 weeks ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆514Updated 3 weeks ago
- ☆166Updated last year
- ☆433Updated 7 months ago
- Helpful tools and examples for working with flex-attention☆811Updated last week