AI-Hypercomputer / maxdiffusion
☆178Updated this week
Alternatives and similar repositories for maxdiffusion:
Users that are interested in maxdiffusion are comparing it to the libraries listed below
- Google TPU optimizations for transformers models☆86Updated this week
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆266Updated this week
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"☆46Updated this week
- ☆181Updated 3 weeks ago
- ☆75Updated 6 months ago
- ☆126Updated this week
- Scalable and Performant Data Loading☆207Updated this week
- Faster generation with text-to-image diffusion models.☆206Updated 3 months ago
- Inference code for LLaMA models in JAX☆114Updated 7 months ago
- PyTorch per step fault tolerance (actively under development)☆220Updated this week
- Accelerated inference of 🤗 models using FuriosaAI NPU chips.☆26Updated 7 months ago
- JAX implementation of the Llama 2 model☆213Updated 11 months ago
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆471Updated this week
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆121Updated 9 months ago
- ☆269Updated 6 months ago
- This repository contains the experimental PyTorch native float8 training UX☆219Updated 5 months ago
- Implementation of Flash Attention in Jax☆204Updated 10 months ago
- Megatron's multi-modal data loader☆157Updated this week
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆238Updated this week
- supporting pytorch FSDP for optimizers☆75Updated last month
- jax-triton contains integrations between JAX and OpenAI Triton☆363Updated this week
- Focused on fast experimentation and simplicity☆64Updated 3 weeks ago
- A library for unit scaling in PyTorch☆118Updated last month
- A simple library for scaling up JAX programs☆129Updated 2 months ago
- Experiment of using Tangent to autodiff triton☆74Updated 11 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆90Updated 2 months ago
- ☆85Updated 10 months ago
- ☆170Updated last week
- ☆275Updated this week