ml-gde / jaxgardenLinks
A collection of reusable, high-performance, well-documented, thorough-tested layers and models in Jax
☆21Updated 7 months ago
Alternatives and similar repositories for jaxgarden
Users that are interested in jaxgarden are comparing it to the libraries listed below
Sorting:
- JAX Implementation of Black Forest Labs' Flux.1 family of models☆40Updated 2 months ago
- ☆52Updated last year
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆298Updated last year
- JAX implementation of the Llama 2 model☆215Updated last year
- Implementation of Flash Attention in Jax☆224Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆86Updated 2 years ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆190Updated 3 years ago
- Various transformers for FSDP research☆38Updated 3 years ago
- ☆192Updated last week
- Minimal sharded dataset loaders, decoders, and utils for multi-modal document, image, and text datasets.☆160Updated last year
- This is a port of Mistral-7B model in JAX☆33Updated last year
- JAX Synergistic Memory Inspector☆184Updated last year
- ☆69Updated last week
- HomebrewNLP in JAX flavour for maintable TPU-Training☆51Updated 2 years ago
- Slide decks, coding exercises, and quick references for learning the JAX AI Stack☆218Updated this week
- Fast, Modern, and Low Precision PyTorch Optimizers☆120Updated 3 weeks ago
- ☆314Updated last year
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.☆22Updated 3 years ago
- This repository shows various ways of deploying a vision model (TensorFlow) from 🤗 Transformers.☆30Updated 3 years ago
- Google TPU optimizations for transformers models☆132Updated last month
- Presents comprehensive benchmarks of XLA-compatible pre-trained models in Keras.☆37Updated 2 years ago
- A set of Python scripts that makes your experience on TPU better☆55Updated 4 months ago
- ☆63Updated 3 years ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆113Updated 2 months ago
- ☆92Updated last year
- some common Huggingface transformers in maximal update parametrization (µP)☆87Updated 3 years ago
- Train vision models using JAX and 🤗 transformers☆100Updated last month
- Just some miscellaneous utility functions / decorators / modules related to Pytorch and Accelerate to help speed up implementation of new…☆126Updated last year
- ☆125Updated last year
- Amos optimizer with JEstimator lib.☆82Updated last year