taco-project / FlexKVLinks
☆138Updated this week
Alternatives and similar repositories for FlexKV
Users that are interested in FlexKV are comparing it to the libraries listed below
Sorting:
- KV cache store for distributed LLM inference☆384Updated last month
- NVIDIA Inference Xfer Library (NIXL)☆788Updated this week
- Efficient and easy multi-instance LLM serving☆520Updated 4 months ago
- GLake: optimizing GPU memory management and IO transmission.☆494Updated 9 months ago
- Virtualized Elastic KV Cache for Dynamic GPU Sharing and Beyond☆735Updated last month
- Offline optimization of your disaggregated Dynamo graph☆137Updated this week
- NVIDIA NCCL Tests for Distributed Training☆132Updated this week
- A low-latency & high-throughput serving engine for LLMs☆462Updated 2 months ago
- Open Model Engine (OME) — Kubernetes operator for LLM serving, GPU scheduling, and model lifecycle management. Works with SGLang, vLLM, T…☆355Updated this week
- Fast OS-level support for GPU checkpoint and restore☆267Updated 3 months ago
- Disaggregated serving system for Large Language Models (LLMs).☆761Updated 9 months ago
- ☆337Updated last week
- CUDA checkpoint and restore utility☆401Updated 3 months ago
- DeepSeek-V3/R1 inference performance simulator☆175Updated 9 months ago
- ☆518Updated this week
- GPUd automates monitoring, diagnostics, and issue identification for GPUs☆468Updated last week
- MSCCL++: A GPU-driven communication stack for scalable AI applications☆449Updated this week
- A workload for deploying LLM inference services on Kubernetes☆153Updated this week
- High Performance KV Cache Store for LLM☆45Updated last month
- Venus Collective Communication Library, supported by SII and Infrawaves.☆132Updated this week
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆123Updated 2 weeks ago
- Dynamic Memory Management for Serving LLMs without PagedAttention☆454Updated 7 months ago
- Fast and memory-efficient exact attention☆107Updated 3 weeks ago
- The driver for LMCache core to run in vLLM☆59Updated 11 months ago
- RDMA and SHARP plugins for nccl library☆218Updated last month
- Perplexity GPU Kernels☆548Updated 2 months ago
- Materials for learning SGLang☆714Updated this week
- TePDist (TEnsor Program DISTributed) is an HLO-level automatic distributed system for DL models.☆99Updated 2 years ago
- A high-performance RL training-inference weight synchronization framework, designed to enable second-level parameter updates from trainin…☆123Updated 2 weeks ago
- DeepXTrace is a lightweight tool for precisely diagnosing slow ranks in DeepEP-based environments.☆85Updated 3 weeks ago