NVIDIA Inference Xfer Library (NIXL)
☆970Apr 10, 2026Updated this week
Alternatives and similar repositories for nixl
Users that are interested in nixl are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- A Datacenter Scale Distributed Inference Serving Framework☆6,527Updated this week
- Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.☆5,071Updated this week
- Perplexity GPU Kernels☆565Nov 7, 2025Updated 5 months ago
- KV cache store for distributed LLM inference☆405Nov 13, 2025Updated 5 months ago
- FlashInfer: Kernel Library for LLM Serving☆5,372Updated this week
- Deploy open-source AI quickly and easily - Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- A fast communication-overlapping library for tensor/expert parallelism on GPUs.☆1,286Aug 28, 2025Updated 7 months ago
- A lightweight design for computation-communication overlap.☆226Jan 20, 2026Updated 2 months ago
- Distributed Compiler based on Triton for Parallel Systems☆1,403Updated this week
- DeepXTrace is a lightweight tool for precisely diagnosing slow ranks in DeepEP-based environments.☆95Jan 16, 2026Updated 3 months ago
- Efficient and easy multi-instance LLM serving☆543Mar 12, 2026Updated last month
- Supercharge Your LLM with the Fastest KV Cache Layer