AI-Hypercomputer / maxdiffusionLinks
☆214Updated this week
Alternatives and similar repositories for maxdiffusion
Users that are interested in maxdiffusion are comparing it to the libraries listed below
Sorting:
- Google TPU optimizations for transformers models☆112Updated 4 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆334Updated this week
- JAX implementation of the Llama 2 model☆217Updated last year
- ☆138Updated 2 weeks ago
- ☆186Updated last week
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆498Updated last week
- jax-triton contains integrations between JAX and OpenAI Triton☆395Updated this week
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆60Updated 2 months ago
- Focused on fast experimentation and simplicity☆73Updated 5 months ago
- ☆267Updated 10 months ago
- PyTorch per step fault tolerance (actively under development)☆302Updated last week
- Scalable and Performant Data Loading☆269Updated this week
- ☆116Updated last week
- Implementation of Diffusion Transformer (DiT) in JAX☆276Updated 11 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆127Updated last year
- JAX-Toolbox☆308Updated this week
- An implementation of PSGD Kron second-order optimizer for PyTorch☆91Updated 2 months ago
- Faster generation with text-to-image diffusion models.☆214Updated 7 months ago
- supporting pytorch FSDP for optimizers☆79Updated 5 months ago
- Load compute kernels from the Hub☆139Updated this week
- A simple library for scaling up JAX programs☆137Updated 7 months ago
- seqax = sequence modeling + JAX☆155Updated last month
- Inference code for LLaMA models in JAX☆117Updated last year
- WIP☆93Updated 9 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆249Updated this week
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆126Updated 3 weeks ago
- ☆78Updated 10 months ago
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆586Updated this week
- Efficient optimizers☆206Updated this week
- DeMo: Decoupled Momentum Optimization☆188Updated 6 months ago