NVIDIA / KAI-Scheduler
KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale
☆189Updated this week
Alternatives and similar repositories for KAI-Scheduler:
Users that are interested in KAI-Scheduler are comparing it to the libraries listed below
- JobSet: a k8s native API for distributed ML training and HPC workloads☆211Updated this week
- Gateway API Inference Extension☆189Updated this week
- Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes☆335Updated this week
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆354Updated this week
- K8s device plugin for GPU sharing☆100Updated last year
- Example DRA driver that developers can fork and modify to get them started writing their own.☆65Updated 2 weeks ago
- A toolkit for discovering cluster network topology.☆40Updated last week
- Holistic job manager on Kubernetes☆114Updated last year
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆92Updated this week
- knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.☆64Updated last week
- ☆100Updated last week
- ☆50Updated last year
- ☆176Updated this week
- InstaSlice Operator facilitates slicing of accelerators using stable APIs☆30Updated this week
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆113Updated this week
- Enabling Kubernetes to make pod placement decisions with platform intelligence.☆174Updated 2 months ago
- Envoy AI Gateway is an open source project for using Envoy Gateway to handle request traffic from application clients to Generative AI se…☆197Updated this week
- GenAI inference performance benchmarking tool☆21Updated this week
- GPU plugin to the node feature discovery for Kubernetes☆299Updated 10 months ago
- This project provides a framework that runs Slurm in Kubernetes.☆70Updated last week
- AWS virtual gpu device plugin provides capability to use smaller virtual gpus for your machine learning inference workloads☆204Updated last year
- A federation scheduler for multi-cluster☆36Updated last month
- ☆123Updated 3 weeks ago
- CUDA checkpoint and restore utility☆319Updated 2 months ago
- ☆114Updated this week
- Repository for open inference protocol specification☆52Updated 8 months ago
- A Topology-Aware Custom Scheduler For Kubernetes☆63Updated last year
- 🧯 Kubernetes coverage for fault awareness and recovery, works for any LLMOps, MLOps, AI workloads.☆29Updated 3 months ago
- NVIDIA Network Operator☆244Updated this week
- MIG Partition Editor for NVIDIA GPUs☆192Updated this week