zhuangwang93 / EspressoLinks
Hi-Speed DNN Training with Espresso: Unleashing the Full Potential of Gradient Compression with Near-Optimal Usage Strategies (EuroSys '23)
☆15Updated 2 years ago
Alternatives and similar repositories for Espresso
Users that are interested in Espresso are comparing it to the libraries listed below
Sorting:
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆35Updated 2 years ago
- Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k c…☆27Updated 2 years ago
- THC: Accelerating Distributed Deep Learning Using Tensor Homomorphic Compression☆19Updated last year
- ☆58Updated last year
- Artifact for "Shockwave: Fair and Efficient Cluster Scheduling for Dynamic Adaptation in Machine Learning" [NSDI '23]☆45Updated 2 years ago
- Artifacts for our SIGCOMM'22 paper Muri☆43Updated last year
- Managed collective communication service☆22Updated last year
- ☆51Updated 2 years ago
- ☆40Updated last year
- Analyze network performance in distributed training☆19Updated 4 years ago
- ☆23Updated last year
- ☆68Updated 2 years ago
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆75Updated 2 years ago
- ☆25Updated 2 years ago
- ☆44Updated last year
- [TBD] "m4: A Learned Flow-level Network Simulator" by Chenning Li, Anton A. Zabreyko, Arash Nasr-Esfahany, Kevin Zhao, Prateesh Goyal, Mo…☆15Updated 2 weeks ago
- Artifacts for our ASPLOS'23 paper ElasticFlow☆53Updated last year
- [NSDI 2023] TopoOpt: Optimizing the Network Topology for Distributed DNN Training☆32Updated last year
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆51Updated 2 years ago
- Model-less Inference Serving☆92Updated last year
- ☆16Updated last year
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆67Updated 10 months ago
- AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving (OSDI 23)☆87Updated 2 years ago
- ☆83Updated 3 years ago
- ☆51Updated 9 months ago
- A Hybrid Framework to Build High-performance Adaptive Neural Networks for Kernel Datapath☆27Updated 2 years ago
- ☆22Updated 2 years ago
- ☆20Updated 2 years ago
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆127Updated 3 years ago
- Herald: Accelerating Neural Recommendation Training with Embedding Scheduling (NSDI 2024)☆23Updated last year