ml-gde / jaxgardenLinks
A collection of reusable, high-performance, well-documented, thorough-tested layers and models in Jax
☆21Updated 3 months ago
Alternatives and similar repositories for jaxgarden
Users that are interested in jaxgarden are comparing it to the libraries listed below
Sorting:
- Slide decks, coding exercises, and quick references for learning the JAX AI Stack☆33Updated 3 weeks ago
- Implementation of Flash Attention in Jax☆216Updated last year
- JAX Synergistic Memory Inspector☆179Updated last year
- A JAX-native LLM Post-Training Library☆143Updated this week
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆291Updated last year
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆188Updated 3 years ago
- JAX implementation of the Llama 2 model☆219Updated last year
- (EasyDel Former) is a utility library designed to simplify and enhance the development in JAX☆28Updated this week
- ☆61Updated 3 years ago
- Implementation of numerous Vision Transformers in Google's JAX and Flax.☆22Updated 3 years ago
- Various transformers for FSDP research☆38Updated 2 years ago
- This is a port of Mistral-7B model in JAX☆32Updated last year
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆87Updated last year
- A set of Python scripts that makes your experience on TPU better☆54Updated last year
- This repository shows various ways of deploying a vision model (TensorFlow) from 🤗 Transformers.☆30Updated 3 years ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆107Updated 5 months ago
- Minimal sharded dataset loaders, decoders, and utils for multi-modal document, image, and text datasets.☆158Updated last year
- ☆361Updated last year
- ☆188Updated last week
- LoRA for arbitrary JAX models and functions☆142Updated last year
- Unofficial JAX implementations of deep learning research papers☆156Updated 3 years ago
- ☆50Updated 10 months ago
- Accelerate, Optimize performance with streamlined training and serving options with JAX.☆308Updated this week
- Presents comprehensive benchmarks of XLA-compatible pre-trained models in Keras.☆37Updated 2 years ago
- ☆13Updated 3 years ago
- ☆19Updated 2 years ago
- Implementation of a Transformer, but completely in Triton☆274Updated 3 years ago
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆113Updated 2 years ago
- HomebrewNLP in JAX flavour for maintable TPU-Training☆50Updated last year
- Google TPU optimizations for transformers models☆120Updated 7 months ago