AI-Hypercomputer / maxdiffusionLinks
☆290Updated this week
Alternatives and similar repositories for maxdiffusion
Users that are interested in maxdiffusion are comparing it to the libraries listed below
Sorting:
- a Jax quantization library☆79Updated last week
- ☆148Updated last month
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆542Updated last week
- Google TPU optimizations for transformers models☆131Updated last week
- JAX-Toolbox☆369Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆396Updated 6 months ago
- JAX implementation of the Llama 2 model☆216Updated last year
- Dion optimizer algorithm☆409Updated this week
- ☆340Updated 2 weeks ago
- ☆190Updated last week
- JAX Implementation of Black Forest Labs' Flux.1 family of models☆39Updated last month
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆459Updated 3 weeks ago
- Load compute kernels from the Hub☆352Updated last week
- ☆91Updated last year
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆79Updated last week
- Implementation of Flash Attention in Jax☆222Updated last year
- Modular, scalable library to train ML models☆182Updated last week
- ☆69Updated last year
- Minimal yet performant LLM examples in pure JAX☆219Updated 3 weeks ago
- seqax = sequence modeling + JAX☆169Updated 5 months ago
- Focused on fast experimentation and simplicity☆76Updated last year
- Efficient optimizers☆277Updated last week
- ☆286Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆299Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆436Updated 2 weeks ago
- Scalable and Performant Data Loading☆356Updated this week
- ☆314Updated last year
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- supporting pytorch FSDP for optimizers☆84Updated last year
- A Jax-based library for building transformers, includes implementations of GPT, Gemma, LlaMa, Mixtral, Whisper, SWin, ViT and more.☆297Updated last year