suquark / hopliteLinks
☆44Updated 4 years ago
Alternatives and similar repositories for hoplite
Users that are interested in hoplite are comparing it to the libraries listed below
Sorting:
- ☆15Updated 3 years ago
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆51Updated 2 years ago
- Artifact for "Shockwave: Fair and Efficient Cluster Scheduling for Dynamic Adaptation in Machine Learning" [NSDI '23]☆45Updated 2 years ago
- ☆56Updated 4 years ago
- ☆12Updated 4 months ago
- Dorylus: Affordable, Scalable, and Accurate GNN Training☆76Updated 4 years ago
- AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving (OSDI 23)☆85Updated 2 years ago
- PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications☆126Updated 3 years ago
- Analyze network performance in distributed training☆19Updated 4 years ago
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆74Updated 2 years ago
- ☆38Updated 4 years ago
- Artifacts for our ASPLOS'23 paper ElasticFlow☆52Updated last year
- Repository for MLCommons Chakra schema and tools☆39Updated last year
- Model-less Inference Serving☆92Updated last year
- Fine-grained GPU sharing primitives☆144Updated last month
- ☆24Updated 2 years ago
- A Federated Execution Engine for Fast Distributed Computation Over Slow Networks☆26Updated 4 years ago
- Reading seminar in Harvard Cloud Networking and Systems Group☆16Updated 3 years ago
- Tiresias is a GPU cluster manager for distributed deep learning training.☆158Updated 5 years ago
- This repository contains code for the paper: Bergsma S., Zeyl T., Senderovich A., and Beck J. C., "Generating Complex, Realistic Cloud Wo…☆43Updated 3 years ago
- ☆42Updated 2 years ago
- Artifact for "Apparate: Rethinking Early Exits to Tame Latency-Throughput Tensions in ML Serving" [SOSP '24]☆25Updated 9 months ago
- Virtual Memory Abstraction for Serverless Architectures☆48Updated 3 years ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆43Updated 3 years ago
- Source code for OSDI 2023 paper titled "Cilantro - Performance-Aware Resource Allocation for General Objectives via Online Feedback"☆40Updated 2 years ago
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆62Updated last year
- rFaaS: a high-performance FaaS platform with RDMA acceleration for low-latency invocations.☆53Updated 2 months ago
- ☆44Updated last year
- ☆37Updated last year
- The prototype for NSDI paper "NetHint: White-Box Networking for Multi-Tenant Data Centers"☆26Updated last year