asplos-contest / 2025
The ASPLOS 2025 / EuroSys 2025 Contest Track
☆35Updated this week
Alternatives and similar repositories for 2025:
Users that are interested in 2025 are comparing it to the libraries listed below
- Supplemental materials for The ASPLOS 2025 / EuroSys 2025 Contest on Intra-Operator Parallelism for Distributed Deep Learning☆23Updated 4 months ago
- ☆79Updated 2 years ago
- Compiler for Dynamic Neural Networks☆46Updated last year
- PerFlow-AI is a programmable performance analysis, modeling, prediction tool for AI system.☆19Updated this week
- ASPLOS'24: Optimal Kernel Orchestration for Tensor Programs with Korch☆33Updated last month
- ☆23Updated 9 months ago
- Ultra | Ultimate | Unified CCL☆59Updated 2 months ago
- ☆22Updated last year
- ☆28Updated 9 months ago
- ☆36Updated 5 months ago
- REEF is a GPU-accelerated DNN inference serving system that enables instant kernel preemption and biased concurrent execution in GPU sche…☆94Updated 2 years ago
- Horizontal Fusion☆23Updated 3 years ago
- TACCL: Guiding Collective Algorithm Synthesis using Communication Sketches☆73Updated last year
- ☆70Updated 3 years ago
- Bamboo is a system for running large pipeline-parallel DNNs affordably, reliably, and efficiently using spot instances.☆49Updated 2 years ago
- Microsoft Collective Communication Library☆65Updated 5 months ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆50Updated 10 months ago
- ☆34Updated 2 months ago
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆60Updated 10 months ago
- LLM serving cluster simulator☆97Updated last year
- ☆35Updated 9 months ago
- ☆23Updated 2 years ago
- ☆53Updated 4 years ago
- Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training☆23Updated last year
- A hierarchical collective communications library with portable optimizations☆34Updated 4 months ago
- Artifacts for our ASPLOS'23 paper ElasticFlow☆51Updated 11 months ago
- NEO is a LLM inference engine built to save the GPU memory crisis by CPU offloading☆21Updated last month
- FlexFlow Serve: Low-Latency, High-Performance LLM Serving☆34Updated last week
- Summary of some awesome work for optimizing LLM inference☆69Updated 2 weeks ago
- ☆32Updated 10 months ago