AI-Hypercomputer / maxdiffusionLinks
☆268Updated this week
Alternatives and similar repositories for maxdiffusion
Users that are interested in maxdiffusion are comparing it to the libraries listed below
Sorting:
- Scalable and Performant Data Loading☆311Updated this week
- Google TPU optimizations for transformers models☆120Updated 9 months ago
- JAX-Toolbox☆355Updated this week
- JAX implementation of the Llama 2 model☆216Updated last year
- ☆145Updated last week
- Dion optimizer algorithm☆369Updated 3 weeks ago
- ☆91Updated last year
- ☆190Updated last month
- ☆335Updated last month
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆539Updated last month
- Implementation of Flash Attention in Jax☆219Updated last year
- Load compute kernels from the Hub☆304Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆420Updated last week
- a Jax quantization library☆52Updated this week
- Modular, scalable library to train ML models☆167Updated this week
- supporting pytorch FSDP for optimizers☆83Updated 10 months ago
- jax-triton contains integrations between JAX and OpenAI Triton☆428Updated last week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆385Updated 4 months ago
- Focused on fast experimentation and simplicity☆75Updated 10 months ago
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆76Updated last month
- ☆283Updated last year
- Efficient optimizers☆274Updated last week
- ☆309Updated last year
- A FlashAttention implementation for JAX with support for efficient document mask computation and context parallelism.☆147Updated 6 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆132Updated last year
- A simple library for scaling up JAX programs☆144Updated 11 months ago
- A library for unit scaling in PyTorch☆132Updated 3 months ago
- ☆67Updated 11 months ago
- Minimal yet performant LLM examples in pure JAX☆186Updated last month
- seqax = sequence modeling + JAX☆168Updated 3 months ago