awslabs / slapoLinks
A schedule language for large model training
☆148Updated last year
Alternatives and similar repositories for slapo
Users that are interested in slapo are comparing it to the libraries listed below
Sorting:
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆138Updated 2 years ago
- ☆144Updated 4 months ago
- ☆43Updated last year
- ☆147Updated 11 months ago
- Research and development for optimizing transformers☆128Updated 4 years ago
- System for automated integration of deep learning backends.☆47Updated 2 years ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆121Updated 3 years ago
- ☆74Updated 4 years ago
- Home for OctoML PyTorch Profiler☆113Updated 2 years ago
- ☆105Updated 9 months ago
- ☆79Updated 2 years ago
- ☆92Updated 2 years ago
- ☆81Updated 7 months ago
- extensible collectives library in triton☆86Updated 2 months ago
- Training neural networks in TensorFlow 2.0 with 5x less memory☆131Updated 3 years ago
- DietCode Code Release☆64Updated 2 years ago
- ☆23Updated 6 months ago
- ☆212Updated 11 months ago
- Automated Parallelization System and Infrastructure for Multiple Ecosystems☆79Updated 7 months ago
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆50Updated 11 months ago
- FTPipe and related pipeline model parallelism research.☆41Updated 2 years ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆39Updated 2 years ago
- MLIR-based partitioning system☆91Updated this week
- ☆91Updated 5 months ago
- A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.☆166Updated this week
- A baseline repository of Auto-Parallelism in Training Neural Networks☆144Updated 2 years ago
- Collection of kernels written in Triton language☆128Updated 2 months ago
- A Vectorized N:M Format for Unleashing the Power of Sparse Tensor Cores☆51Updated last year
- nnScaler: Compiling DNN models for Parallel Training☆113Updated this week
- Boost hardware utilization for ML training workloads via Inter-model Horizontal Fusion☆32Updated last year