fredrickang / LaLaRANDLinks
LaLaRAND: Flexible Layer-by-Layer CPU/GPU Scheduling for Real-Time DNN Tasks
☆17Updated 3 years ago
Alternatives and similar repositories for LaLaRAND
Users that are interested in LaLaRAND are comparing it to the libraries listed below
Sorting:
- Source code for the paper: "A Latency-Predictable Multi-Dimensional Optimization Framework forDNN-driven Autonomous Systems"☆22Updated 5 years ago
- ☆38Updated 7 months ago
- LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale☆177Updated 6 months ago
- ☆53Updated last year
- ☆23Updated this week
- [ACM EuroSys 2023] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access☆56Updated 5 months ago
- ☆25Updated 3 years ago
- Proteus: A High-Throughput Inference-Serving System with Accuracy Scaling☆12Updated last year
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- An interference-aware scheduler for fine-grained GPU sharing☆159Updated 2 months ago
- MISO: Exploiting Multi-Instance GPU Capability on Multi-Tenant GPU Clusters☆20Updated 2 years ago
- ☆81Updated 8 months ago
- LLM serving cluster simulator☆134Updated last year
- ☆166Updated last year
- ☆52Updated 3 years ago
- This repository is established to store personal notes and annotated papers during daily research.☆179Updated 2 weeks ago
- ☆28Updated last year
- ☆223Updated 3 months ago
- Model-less Inference Serving☆93Updated 2 years ago
- ☆116Updated this week
- iGniter, an interference-aware GPU resource provisioning framework for achieving predictable performance of DNN inference in the cloud.☆39Updated last year
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆76Updated 3 months ago
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆44Updated 3 years ago
- REEF is a GPU-accelerated DNN inference serving system that enables instant kernel preemption and biased concurrent execution in GPU sche…☆104Updated 3 years ago
- Multi-DNN Inference Engine for Heterogeneous Mobile Processors☆37Updated last year
- ☆26Updated 2 years ago
- ☆26Updated last year
- Hi-Speed DNN Training with Espresso: Unleashing the Full Potential of Gradient Compression with Near-Optimal Usage Strategies (EuroSys '2…☆15Updated 2 years ago
- ☆41Updated 2 years ago
- ☆78Updated 2 years ago