zhuangwang93 / Cupcake
Cupcake: A Compression Scheduler for Scalable Communication-Efficient Distributed Training (MLSys '23)
☆9Updated last year
Alternatives and similar repositories for Cupcake:
Users that are interested in Cupcake are comparing it to the libraries listed below
- Hi-Speed DNN Training with Espresso: Unleashing the Full Potential of Gradient Compression with Near-Optimal Usage Strategies (EuroSys '2…☆15Updated last year
- ☆16Updated 9 months ago
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆18Updated 2 months ago
- ☆14Updated 2 years ago
- Reading seminar in Harvard Cloud Networking and Systems Group☆16Updated 2 years ago
- Deferred Continuous Batching in Resource-Efficient Large Language Model Serving (EuroMLSys 2024)☆12Updated 8 months ago
- A rust-based benchmark for BlueField SmartNICs.☆27Updated last year
- Artifacts for our SIGCOMM'23 paper Ditto☆15Updated last year
- Ultra | Ultimate | Unified CCL☆32Updated this week
- Source code for OSDI 2023 paper titled "Cilantro - Performance-Aware Resource Allocation for General Objectives via Online Feedback"☆38Updated last year
- SocksDirect code repository☆19Updated 2 years ago
- ☆43Updated 3 years ago
- Primo: Practical Learning-Augmented Systems with Interpretable Models☆19Updated last year
- Code for "Shockwave: Fair and Efficient Cluster Scheduling for Dynamic Adaptation in Machine Learning" [NSDI '23]☆39Updated 2 years ago
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆48Updated 2 years ago
- Efficient Interactive LLM Serving with Proxy Model-based Sequence Length Prediction | A tiny BERT model can tell you the verbosity of an …☆26Updated 8 months ago
- Artifacts for our SIGCOMM'22 paper Muri☆41Updated last year
- ☆18Updated 7 months ago
- ☆23Updated last year
- ☆14Updated 3 years ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆40Updated 2 years ago
- Deduplication over dis-aggregated memory for Serverless Computing☆12Updated 2 years ago
- Herald: Accelerating Neural Recommendation Training with Embedding Scheduling (NSDI 2024)☆21Updated 9 months ago
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆22Updated 2 months ago
- ☆40Updated 7 months ago
- ☆18Updated last year
- SOTA Learning-augmented Systems☆34Updated 2 years ago
- The prototype for NSDI paper "NetHint: White-Box Networking for Multi-Tenant Data Centers"☆26Updated last year
- THC: Accelerating Distributed Deep Learning Using Tensor Homomorphic Compression☆14Updated 6 months ago
- Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k c…☆24Updated 2 years ago