Medusa: Accelerating Serverless LLM Inference with Materialization [ASPLOS'25]
☆44May 13, 2025Updated 11 months ago
Alternatives and similar repositories for Medusa
Users that are interested in Medusa are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Medusa: Accelerating Serverless LLM Inference with Materialization [ASPLOS'25]☆12Nov 8, 2024Updated last year
- Deft: A Scalable Tree Index for Disaggregated Memory☆23Apr 23, 2025Updated 11 months ago
- The source code of INFless,a native serverless platform for AI inference.☆46Oct 10, 2022Updated 3 years ago
- Examples of usage for Mellanox HW offloads☆17Jan 18, 2022Updated 4 years ago
- The official implementation of OSDI'25 paper BlitzScale☆44Sep 20, 2025Updated 6 months ago
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- ☆11Aug 9, 2021Updated 4 years ago
- ☆101Apr 6, 2026Updated last week
- ☆28Jun 22, 2025Updated 9 months ago
- A caching framework for microservice applications☆24Apr 22, 2024Updated last year
- Open-source implementation for "Helix: Serving Large Language Models over Heterogeneous GPUs and Network via Max-Flow"☆81Oct 15, 2025Updated 6 months ago
- APEX+ is an LLM Serving Simulator☆44Jun 16, 2025Updated 9 months ago
- A fast and scalable distributed lock service using programmable switches.☆20Jul 30, 2024Updated last year
- [OSDI 2024] Motor: Enabling Multi-Versioning for Distributed Transactions on Disaggregated Memory☆50Mar 3, 2024Updated 2 years ago
- a simple API to use CUPTI☆10Aug 19, 2025Updated 7 months ago
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- Pacman: An Efficient Compaction Approach for Log-Structured Key-Value Store on Persistent Memory☆44Dec 12, 2022Updated 3 years ago
- Pluggable in-process caching engine to build and scale high performance services☆18Apr 8, 2026Updated last week
- STREAMer: Benchmarking remote volatile and non-volatile memory bandwidth☆18Aug 21, 2023Updated 2 years ago
- Nap - NUMA-Aware Persistent Indexes☆41May 27, 2021Updated 4 years ago
- Fast OS-level support for GPU checkpoint and restore☆280Sep 28, 2025Updated 6 months ago
- Efficient Long-context Language Model Training by Core Attention Disaggregation☆97Apr 7, 2026Updated last week
- Artifacts for our ASPLOS'23 paper dRAID☆30Feb 24, 2023Updated 3 years ago
- PipeRAG: Fast Retrieval-Augmented Generation via Algorithm-System Co-design (KDD 2025)☆31Jun 14, 2024Updated last year
- This is the implementation repository of our SOSP'24 paper: Aceso: Achieving Efficient Fault Tolerance in Memory-Disaggregated Key-Value …☆24Oct 20, 2024Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- An efficient implementation of the NSA (Native Sparse Attention) kernel☆133Jun 24, 2025Updated 9 months ago
- ☆38Aug 7, 2025Updated 8 months ago
- ☆41Nov 28, 2022Updated 3 years ago
- Accelerating Large-Scale Reasoning Model Inference with Sparse Self-Speculative Decoding☆98Dec 2, 2025Updated 4 months ago
- [FAST'25] ShiftLock: Mitigate One-sided RDMA Lock Contention via Handover.☆20Feb 11, 2025Updated last year
- [NeurIPS 2025] ClusterFusion: Expanding Operator Fusion Scope for LLM Inference via Cluster-Level Collective Primitive☆66Dec 11, 2025Updated 4 months ago
- Implementation from scratch in C of the Multi-head latent attention used in the Deepseek-v3 technical paper.☆18Jan 15, 2025Updated last year
- Ensō is a high-performance streaming interface for NIC-application communication.☆79Updated this week
- Deduplication over dis-aggregated memory for Serverless Computing☆14Mar 21, 2022Updated 4 years ago
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Deferred Continuous Batching in Resource-Efficient Large Language Model Serving (EuroMLSys 2024)☆19May 28, 2024Updated last year
- TeRM: Extending RDMA-Attached Memory with SSD [FAST'24]☆45Oct 21, 2024Updated last year
- An auxiliary project analysis of the characteristics of KV in DiT Attention.☆34Nov 29, 2024Updated last year
- λ-IO: a unified I/O stack for computational storage [FAST'23]☆79Apr 29, 2025Updated 11 months ago
- Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of pap…☆282Mar 6, 2025Updated last year
- C++ interfaces for RDMA access☆83Mar 30, 2026Updated 2 weeks ago
- LoRAFusion: Efficient LoRA Fine-Tuning for LLMs☆26Apr 8, 2026Updated last week