AI-Hypercomputer / maxdiffusionLinks
☆283Updated last week
Alternatives and similar repositories for maxdiffusion
Users that are interested in maxdiffusion are comparing it to the libraries listed below
Sorting:
- JAX-Toolbox☆364Updated this week
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆540Updated 2 weeks ago
- JAX implementation of the Llama 2 model☆216Updated last year
- a Jax quantization library☆68Updated last week
- Scalable and Performant Data Loading☆349Updated this week
- Google TPU optimizations for transformers models☆123Updated 10 months ago
- Load compute kernels from the Hub☆337Updated last week
- ☆148Updated 3 weeks ago
- ☆337Updated 2 weeks ago
- Implementation of Diffusion Transformer (DiT) in JAX☆297Updated last year
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆454Updated 3 weeks ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆392Updated 5 months ago
- Dion optimizer algorithm☆395Updated 2 weeks ago
- A simple library for scaling up JAX programs☆144Updated last month
- Minimal yet performant LLM examples in pure JAX☆204Updated 2 months ago
- Efficient optimizers☆276Updated 3 weeks ago
- seqax = sequence modeling + JAX☆168Updated 4 months ago
- ☆91Updated last year
- torchax is a PyTorch frontend for JAX. It gives JAX the ability to author JAX programs using familiar PyTorch syntax. It also provides JA…☆134Updated last week
- ☆285Updated last year
- Implementation of Flash Attention in Jax☆222Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆436Updated this week
- ☆190Updated 2 weeks ago
- Modular, scalable library to train ML models☆176Updated last week
- A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.☆149Updated 3 weeks ago
- supporting pytorch FSDP for optimizers☆84Updated 11 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- For optimization algorithm research and development.☆547Updated 2 weeks ago
- Inference code for LLaMA models in JAX☆120Updated last year
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆78Updated 2 months ago