rwitten / HighPerfLLMs2024
☆407Updated 8 months ago
Alternatives and similar repositories for HighPerfLLMs2024:
Users that are interested in HighPerfLLMs2024 are comparing it to the libraries listed below
- Legible, Scalable, Reproducible Foundation Models with Named Tensors and Jax☆557Updated this week
- seqax = sequence modeling + JAX☆148Updated last week
- ☆214Updated 8 months ago
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆483Updated last week
- A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.☆524Updated last month
- Building blocks for foundation models.☆464Updated last year
- jax-triton contains integrations between JAX and OpenAI Triton☆384Updated last week
- Minimalistic 4D-parallelism distributed training framework for education purpose☆935Updated 2 weeks ago
- Deep learning for dummies. All the practical details and useful utilities that go into working with real models.☆783Updated 2 weeks ago
- What would you do with 1000 H100s...☆1,016Updated last year
- Implementation of Diffusion Transformer (DiT) in JAX☆269Updated 9 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆232Updated 2 weeks ago
- ☆137Updated this week
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆126Updated last year
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆297Updated this week
- Best practices & guides on how to write distributed pytorch training code☆368Updated 3 weeks ago
- ring-attention experiments☆127Updated 5 months ago
- ☆290Updated this week
- ☆420Updated 5 months ago
- Annotated version of the Mamba paper☆475Updated last year
- ☆149Updated last year
- Cataloging released Triton kernels.☆204Updated 2 months ago
- ☆158Updated last month
- ☆190Updated last month
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆234Updated this week
- PyTorch per step fault tolerance (actively under development)☆266Updated this week
- A repository for research on medium sized language models.☆493Updated 2 months ago
- Open weights language model from Google DeepMind, based on Griffin.☆627Updated last month