boweiliu / ncclLinks
Optimized primitives for collective multi-GPU communication
☆10Updated last year
Alternatives and similar repositories for nccl
Users that are interested in nccl are comparing it to the libraries listed below
Sorting:
- ☆21Updated 9 months ago
- ☆317Updated last year
- An implementation of the Llama architecture, to instruct and delight☆21Updated 6 months ago
- An experimental implementation of compiler-driven automatic sharding of models across a given device mesh.☆47Updated this week
- train with kittens!☆63Updated last year
- PyTorch centric eager mode debugger☆48Updated 11 months ago
- NVIDIA Resiliency Extension is a python package for framework developers and users to implement fault-tolerant features. It improves the …☆239Updated this week
- ☆20Updated 2 years ago
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆161Updated 2 months ago
- ☆42Updated this week
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆32Updated 6 months ago
- A bunch of kernels that might make stuff slower 😉☆65Updated last week
- JAX Scalify: end-to-end scaled arithmetics☆17Updated last year
- ☆121Updated last year
- ☆90Updated last year
- Two implementations of ZeRO-1 optimizer sharding in JAX☆14Updated 2 years ago
- torchcomms: a modern PyTorch communications API☆302Updated this week
- This repository contains the experimental PyTorch native float8 training UX☆227Updated last year
- ring-attention experiments☆160Updated last year
- LM engine is a library for pretraining/finetuning LLMs☆77Updated last week
- Custom triton kernels for training Karpathy's nanoGPT.☆19Updated last year
- Experiment of using Tangent to autodiff triton☆81Updated last year
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆456Updated last week
- Collection of kernels written in Triton language☆173Updated 8 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆64Updated 2 weeks ago
- A library for unit scaling in PyTorch☆132Updated 5 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆46Updated last year
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆271Updated 3 weeks ago
- ☆91Updated last year
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆153Updated 2 years ago