caoting-dotcom / multiBranchModelLinks
Multi-branch model for concurrent execution
☆18Updated 2 years ago
Alternatives and similar repositories for multiBranchModel
Users that are interested in multiBranchModel are comparing it to the libraries listed below
Sorting:
- ☆78Updated 2 years ago
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- DISB is a new DNN inference serving benchmark with diverse workloads and models, as well as real-world traces.☆58Updated last year
- Paella: Low-latency Model Serving with Virtualized GPU Scheduling☆66Updated last year
- ☆15Updated 3 years ago
- hands on model tuning with TVM and profile it on a Mac M1, x86 CPU, and GTX-1080 GPU.☆49Updated 2 years ago
- play gemm with tvm☆92Updated 2 years ago
- ASPLOS'24: Optimal Kernel Orchestration for Tensor Programs with Korch☆40Updated 10 months ago
- From Minimal GEMM to Everything☆98Updated last month
- ☆38Updated 7 months ago
- SOTA Learning-augmented Systems☆37Updated 3 years ago
- Source code for the paper: "A Latency-Predictable Multi-Dimensional Optimization Framework forDNN-driven Autonomous Systems"☆22Updated 5 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆200Updated 3 years ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆124Updated 3 years ago
- Model-less Inference Serving☆93Updated 2 years ago
- The open-source project for "Mandheling: Mixed-Precision On-Device DNN Training with DSP Offloading"[MobiCom'2022]☆19Updated 3 years ago
- MobiSys#114☆23Updated 2 years ago
- ☆18Updated 2 weeks ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆56Updated last year
- An unofficial cuda assembler, for all generations of SASS, hopefully :)☆84Updated 2 years ago
- REEF is a GPU-accelerated DNN inference serving system that enables instant kernel preemption and biased concurrent execution in GPU sche…☆104Updated 3 years ago
- ☆53Updated last year
- A GPU-accelerated DNN inference serving system that supports instant kernel preemption and biased concurrent execution in GPU scheduling.☆44Updated 3 years ago
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆64Updated last year
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆96Updated 4 months ago
- LLM serving cluster simulator☆134Updated last year
- My study note for mlsys☆15Updated last year
- Triton Compiler related materials.☆42Updated last year
- code reading for tvm☆76Updated 4 years ago
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆123Updated last month