ml-gde / jaxgardenLinks
A collection of reusable, high-performance, well-documented, thorough-tested layers and models in Jax
☆21Updated 5 months ago
Alternatives and similar repositories for jaxgarden
Users that are interested in jaxgarden are comparing it to the libraries listed below
Sorting:
- Slide decks, coding exercises, and quick references for learning the JAX AI Stack☆65Updated this week
- Implementation of Flash Attention in Jax☆220Updated last year
- JAX Synergistic Memory Inspector☆179Updated last year
- Presents comprehensive benchmarks of XLA-compatible pre-trained models in Keras.☆37Updated 2 years ago
- This is a port of Mistral-7B model in JAX☆32Updated last year
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆188Updated 3 years ago
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆296Updated last year
- Various transformers for FSDP research☆38Updated 2 years ago
- This repository shows various ways of deploying a vision model (TensorFlow) from 🤗 Transformers.☆30Updated 3 years ago
- JAX implementation of the Llama 2 model☆215Updated last year
- 🤗 Transformers: State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch.☆17Updated 5 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆87Updated last year
- ☆51Updated last week
- ☆19Updated 2 years ago
- JAX Implementation of Black Forest Labs' Flux.1 family of models☆39Updated 2 months ago
- ☆13Updated 3 years ago
- (EasyDel Former) is a utility library designed to simplify and enhance the development in JAX☆28Updated 2 weeks ago
- ☆62Updated 3 years ago
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆110Updated last week
- Unofficial JAX implementations of deep learning research papers☆159Updated 3 years ago
- ☆52Updated last year
- ☆363Updated last year
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.☆22Updated 2 years ago
- ☆190Updated 2 weeks ago
- Tutorials for Triton, a language for writing gpu kernels☆56Updated 2 years ago
- HomebrewNLP in JAX flavour for maintable TPU-Training☆51Updated last year
- ☆24Updated 2 years ago
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆113Updated 2 years ago
- Minimal sharded dataset loaders, decoders, and utils for multi-modal document, image, and text datasets.☆159Updated last year
- Implementation of a Transformer, but completely in Triton☆276Updated 3 years ago