ml-gde / jaxgardenLinks
A collection of reusable, high-performance, well-documented, thorough-tested layers and models in Jax
β21Updated 6 months ago
Alternatives and similar repositories for jaxgarden
Users that are interested in jaxgarden are comparing it to the libraries listed below
Sorting:
- Slide decks, coding exercises, and quick references for learning the JAX AI Stackβ73Updated 3 weeks ago
- This repository shows various ways of deploying a vision model (TensorFlow) from π€ Transformers.β30Updated 3 years ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.β297Updated last year
- Implementation of Flash Attention in Jaxβ223Updated last year
- JAX Synergistic Memory Inspectorβ183Updated last year
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)β189Updated 3 years ago
- HomebrewNLP in JAX flavour for maintable TPU-Trainingβ51Updated last year
- Various transformers for FSDP researchβ38Updated 3 years ago
- β19Updated 3 years ago
- β62Updated 3 years ago
- Minimal sharded dataset loaders, decoders, and utils for multi-modal document, image, and text datasets.β160Updated last year
- JAX implementation of the Llama 2 modelβ216Updated last year
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.β22Updated 2 years ago
- β60Updated this week
- Train very large language models in Jax.β210Updated 2 years ago
- Just some miscellaneous utility functions / decorators / modules related to Pytorch and Accelerate to help speed up implementation of newβ¦β125Updated last year
- Amos optimizer with JEstimator lib.β82Updated last year
- Large scale 4D parallelism pre-training for π€ transformers in Mixture of Experts *(still work in progress)*β87Updated last year
- β190Updated 3 weeks ago
- This is a port of Mistral-7B model in JAXβ32Updated last year
- Train vision models using JAX and π€ transformersβ100Updated last month
- Unofficial JAX implementations of deep learning research papersβ159Updated 3 years ago
- JAX Implementation of Black Forest Labs' Flux.1 family of modelsβ39Updated 2 weeks ago
- β13Updated 3 years ago
- Presents comprehensive benchmarks of XLA-compatible pre-trained models in Keras.β37Updated 2 years ago
- A case study of efficient training of large language models using commodity hardware.β68Updated 3 years ago
- Implementation of numerous Vision Transformers in Google's JAX and Flax.β22Updated 3 years ago
- Automatically take good care of your preemptible TPUsβ37Updated 2 years ago
- β24Updated 3 years ago
- Fast, Modern, and Low Precision PyTorch Optimizersβ116Updated 3 months ago