AI-Hypercomputer / maxdiffusion
☆155Updated this week
Related projects ⓘ
Alternatives and complementary repositories for maxdiffusion
- Google TPU optimizations for transformers models☆74Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆229Updated this week
- ☆178Updated last week
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆112Updated 6 months ago
- JAX implementation of the Llama 2 model☆210Updated 9 months ago
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆40Updated this week
- ☆72Updated 4 months ago
- ☆121Updated this week
- ☆96Updated last month
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆457Updated 2 weeks ago
- ☆223Updated 4 months ago
- A simple library for scaling up JAX programs☆125Updated last week
- JAX-Toolbox☆241Updated this week
- seqax = sequence modeling + JAX☆132Updated 3 months ago
- ☆51Updated 6 months ago
- Train vision models using JAX and 🤗 transformers☆95Updated 3 weeks ago
- Inference code for LLaMA models in JAX☆112Updated 5 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆83Updated last week
- Megatron's multi-modal data loader☆130Updated this week
- ☆292Updated 4 months ago
- ☆197Updated 3 months ago
- A user-friendly tool chain that enables the seamless execution of ONNX models using JAX as the backend.☆98Updated last month
- A library for unit scaling in PyTorch☆105Updated this week
- ☆46Updated last month
- Implementation of Flash Attention in Jax☆196Updated 8 months ago
- Scalable neural net training via automatic normalization in the modular norm.☆119Updated 2 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆104Updated last month
- This repository contains the experimental PyTorch native float8 training UX☆211Updated 3 months ago
- Efficient optimizers☆58Updated this week
- ☆64Updated 2 years ago