axonn-ai / axonnLinks
A parallel framework for training deep neural networks
☆61Updated 3 months ago
Alternatives and similar repositories for axonn
Users that are interested in axonn are comparing it to the libraries listed below
Sorting:
- extensible collectives library in triton☆86Updated 2 months ago
- Experiment of using Tangent to autodiff triton☆79Updated last year
- ☆28Updated 5 months ago
- ☆81Updated 7 months ago
- A Quirky Assortment of CuTe Kernels☆117Updated this week
- LLM training in simple, raw C/CUDA☆99Updated last year
- ☆105Updated 10 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆46Updated this week
- Collection of kernels written in Triton language☆132Updated 2 months ago
- A bunch of kernels that might make stuff slower 😉☆51Updated this week
- Learn CUDA with PyTorch☆27Updated this week
- Sparsity support for PyTorch☆35Updated 3 months ago
- This repository contains the experimental PyTorch native float8 training UX☆224Updated 10 months ago
- A curated collection of resources, tutorials, and best practices for learning and mastering NVIDIA CUTLASS☆189Updated last month
- High-Performance SGEMM on CUDA devices☆95Updated 5 months ago
- Small scale distributed training of sequential deep learning models, built on Numpy and MPI.☆134Updated last year
- Personal solutions to the Triton Puzzles☆19Updated 11 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆45Updated 11 months ago
- PyTorch bindings for CUTLASS grouped GEMM.☆100Updated 3 weeks ago
- ☆21Updated 3 months ago
- Effective transpose on Hopper GPU☆23Updated last month
- PyTorch Single Controller☆231Updated this week
- Complete GPU residency for ML.☆17Updated last week
- ☆109Updated last year
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆65Updated 3 years ago
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆167Updated this week
- ring-attention experiments☆144Updated 8 months ago
- ☆219Updated this week
- A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.☆166Updated this week
- End to End steps for adding custom ops in PyTorch.☆23Updated 4 years ago