casys-kaist / CoVALinks
Official code repository for "CoVA: Exploiting Compressed-Domain Analysis to Accelerate Video Analytics [USENIX ATC 22]"
☆18Updated last year
Alternatives and similar repositories for CoVA
Users that are interested in CoVA are comparing it to the libraries listed below
Sorting:
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- Source code and datasets for Ekya, a system for continuous learning on the edge.☆111Updated 3 years ago
- ☆47Updated 2 years ago
- ☆57Updated 4 years ago
- Adaptive Model Streaming for real-time video inference on edge devices☆41Updated 4 years ago
- ☆12Updated 5 years ago
- ☆213Updated last year
- [ACM EuroSys 2023] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access☆56Updated 5 months ago
- ☆20Updated 3 years ago
- Source code for the paper: "A Latency-Predictable Multi-Dimensional Optimization Framework forDNN-driven Autonomous Systems"☆22Updated 5 years ago
- MobiSys#114☆22Updated 2 years ago
- ☆22Updated last year
- ☆78Updated 2 years ago
- Source code for Jellyfish, a soft real-time inference serving system☆14Updated 3 years ago
- Model-less Inference Serving☆92Updated 2 years ago
- ☆38Updated 6 months ago
- ☆14Updated 4 years ago
- ☆107Updated last week
- ☆102Updated last year
- PacketGame: Multi-Stream Packet Gating for Concurrent Video Inference at Scale☆14Updated 2 years ago
- [DATE 2023] Pipe-BD: Pipelined Parallel Blockwise Distillation☆11Updated 2 years ago
- LaLaRAND: Flexible Layer-by-Layer CPU/GPU Scheduling for Real-Time DNN Tasks☆16Updated 3 years ago
- ☆30Updated 3 years ago
- To deploy Transformer models in CV to mobile devices.☆18Updated 3 years ago
- zTT: Learning-based DVFS with Zero Thermal Throttling for Mobile Devices [MobiSys'21] - Artifact Evaluation☆28Updated 4 years ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆57Updated last year
- ☆53Updated last year
- ☆25Updated 2 years ago
- LLMServingSim: A HW/SW Co-Simulation Infrastructure for LLM Inference Serving at Scale☆168Updated 5 months ago
- Official Repo for "SplitQuant / LLM-PQ: Resource-Efficient LLM Offline Serving on Heterogeneous GPUs via Phase-Aware Model Partition and …☆35Updated 4 months ago