SophiaLi06 / BytePS_THC
THC: Accelerating Distributed Deep Learning Using Tensor Homomorphic Compression
☆18Updated 9 months ago
Alternatives and similar repositories for BytePS_THC
Users that are interested in BytePS_THC are comparing it to the libraries listed below
Sorting:
- Hi-Speed DNN Training with Espresso: Unleashing the Full Potential of Gradient Compression with Near-Optimal Usage Strategies (EuroSys '2…☆15Updated last year
- Managed collective communication service☆21Updated 8 months ago
- ☆51Updated 10 months ago
- ☆38Updated 8 months ago
- [NSDI 2023] TopoOpt: Optimizing the Network Topology for Distributed DNN Training☆29Updated 8 months ago
- Aequitas enables RPC-level QoS in datacenter networks.☆16Updated 2 years ago
- Cupcake: A Compression Scheduler for Scalable Communication-Efficient Distributed Training (MLSys '23)☆9Updated last year
- ☆17Updated last year
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆73Updated last year
- An Automated Performance Optimization Framework for P4-Programmable SmartNICs☆26Updated last year
- [ACM SIGCOMM 2024] "m3: Accurate Flow-Level Performance Estimation using Machine Learning" by Chenning Li, Arash Nasr-Esfahany, Kevin Zha…☆24Updated 7 months ago
- Repository for MLCommons Chakra schema and tools☆39Updated last year
- ☆80Updated 3 years ago
- ☆19Updated 2 years ago
- ☆32Updated 4 years ago
- GPU-accelerated LLM Training Simulator☆29Updated 2 weeks ago
- NS3 implementation of Homa Transport Protocol☆24Updated last year
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆40Updated 5 months ago
- Artifacts for our SIGCOMM'22 paper Muri☆41Updated last year
- ☆23Updated 10 months ago
- Sources and examples for ASPLOS20 paper☆14Updated 4 years ago
- ☆49Updated 2 years ago
- The prototype for NSDI paper "NetHint: White-Box Networking for Multi-Tenant Data Centers"☆26Updated last year
- Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k c…☆26Updated 2 years ago
- A minimum demo for PyTorch distributed extension functionality for collectives.☆11Updated 9 months ago
- ☆42Updated 10 months ago
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆34Updated 2 years ago
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆24Updated 5 months ago
- A collection of tools, code, and documentation to understand the host network on real server hardware.☆35Updated 5 months ago
- The source code of the paper "Runtime Programmable Switches".☆27Updated 2 years ago