mlsys-seo / ooo-backpropLinks
☆25Updated 2 years ago
Alternatives and similar repositories for ooo-backprop
Users that are interested in ooo-backprop are comparing it to the libraries listed below
Sorting:
- [ACM EuroSys 2023] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access☆57Updated this week
- ☆103Updated 2 years ago
- ☆73Updated 2 months ago
- Study Group of Deep Learning Compiler☆161Updated 2 years ago
- ☆12Updated 4 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆40Updated 2 years ago
- Lightweight and Parallel Deep Learning Framework☆264Updated 2 years ago
- Welcome to PeriFlow CLI ☁︎☆12Updated 2 years ago
- ☆25Updated last year
- ☆24Updated 6 years ago
- ☆80Updated 2 years ago
- ☆75Updated 4 years ago
- Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.☆66Updated 4 months ago
- ☆47Updated 11 months ago
- FTPipe and related pipeline model parallelism research.☆41Updated 2 years ago
- Proteus: A High-Throughput Inference-Serving System with Accuracy Scaling☆13Updated last year
- Synthesizer for optimal collective communication algorithms☆113Updated last year
- ☆54Updated 8 months ago
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆50Updated 2 years ago
- A resilient distributed training framework☆94Updated last year
- Microsoft Collective Communication Library☆65Updated 8 months ago
- An Efficient Pipelined Data Parallel Approach for Training Large Model☆77Updated 4 years ago
- ☆49Updated 7 months ago
- ☆38Updated 4 years ago
- Microsoft Collective Communication Library☆353Updated last year
- An interference-aware scheduler for fine-grained GPU sharing☆143Updated 6 months ago
- Know Your Enemy To Save Cloud Energy: Energy-Performance Characterization of Machine Learning Serving (HPCA '23)☆13Updated last month
- ☆55Updated 4 years ago
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆74Updated 2 years ago
- SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training☆35Updated 2 years ago