llm-d / llm-d-inference-schedulerLinks
Inference scheduler for llm-d
☆113Updated this week
Alternatives and similar repositories for llm-d-inference-scheduler
Users that are interested in llm-d-inference-scheduler are comparing it to the libraries listed below
Sorting:
- Distributed KV cache scheduling & offloading libraries☆94Updated this week
- Simplified model deployment on llm-d☆28Updated 6 months ago
- Kubernetes-native AI serving platform for scalable model serving.☆154Updated this week
- knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.☆73Updated 5 months ago
- Example DRA driver that developers can fork and modify to get them started writing their own.☆111Updated this week
- A toolkit for discovering cluster network topology.☆89Updated last month
- GenAI inference performance benchmarking tool☆140Updated 2 weeks ago
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆284Updated 3 weeks ago
- Gateway API Inference Extension☆559Updated this week
- WG Serving☆32Updated 3 weeks ago
- JobSet: a k8s native API for distributed ML training and HPC workloads☆296Updated last week
- Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscaling☆138Updated this week
- Holistic job manager on Kubernetes☆115Updated last year
- Cloud Native Artifacial Intelligence Model Format Specification☆166Updated this week
- Incubating P/D sidecar for llm-d☆16Updated last month
- ☆192Updated this week
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆140Updated 3 weeks ago
- 🧯 Kubernetes coverage for fault awareness and recovery, works for any LLMOps, MLOps, AI workloads.☆33Updated 3 weeks ago
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆645Updated 2 weeks ago
- Go Abstraction for Allocating NVIDIA GPUs with Custom Policies☆120Updated last month
- llm-d helm charts and deployment examples☆48Updated 3 weeks ago
- Command-line tools for managing OCI model artifacts, which are bundled based on Model Spec☆59Updated last week
- A workload for deploying LLM inference services on Kubernetes☆153Updated this week
- A collection of community maintained NRI plugins☆100Updated 3 weeks ago
- Golang bindings for Nvidia Datacenter GPU Manager (DCGM)☆145Updated this week
- A light weight vLLM simulator, for mocking out replicas.☆76Updated last week
- Open Model Engine (OME) — Kubernetes operator for LLM serving, GPU scheduling, and model lifecycle management. Works with SGLang, vLLM, T…☆355Updated this week
- All the things to make the scheduler extendable with wasm.☆129Updated last month
- ☆123Updated 3 years ago
- 💫 A lightweight p2p-based cache system for model distributions on Kubernetes. Reframing now to make it an unified cache system with POSI…☆25Updated last year