NVIDIA / groveLinks
Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscaling
☆59Updated this week
Alternatives and similar repositories for grove
Users that are interested in grove are comparing it to the libraries listed below
Sorting:
- A toolkit for discovering cluster network topology.☆69Updated this week
- Inference scheduler for llm-d☆94Updated last week
- OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)☆273Updated last week
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆815Updated last week
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆129Updated this week
- Gateway API Inference Extension☆486Updated this week
- GenAI inference performance benchmarking tool☆97Updated this week
- NVIDIA DRA Driver for GPUs☆446Updated this week
- A light weight vLLM simulator, for mocking out replicas.☆48Updated last week
- Distributed KV cache coordinator☆71Updated this week
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆583Updated this week
- JobSet: a k8s native API for distributed ML training and HPC workloads☆262Updated this week
- CUDA checkpoint and restore utility☆371Updated last week
- A tool to detect infrastructure issues on cloud native AI systems☆47Updated last week
- MIG Partition Editor for NVIDIA GPUs☆213Updated 2 weeks ago
- knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.☆69Updated 2 months ago
- NVIDIA NCCL Tests for Distributed Training☆111Updated last week
- llm-d benchmark scripts and tooling☆28Updated this week
- WG Serving☆30Updated 2 weeks ago
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆254Updated last week
- GPUd automates monitoring, diagnostics, and issue identification for GPUs☆433Updated this week
- Holistic job manager on Kubernetes☆116Updated last year
- Simplified model deployment on llm-d☆27Updated 2 months ago
- llm-d enables high-performance distributed LLM inference on Kubernetes☆1,781Updated this week
- ☆144Updated last week
- A workload for deploying LLM inference services on Kubernetes☆43Updated this week
- Cloud Native Benchmarking of Foundation Models☆42Updated last month
- NVIDIA Inference Xfer Library (NIXL)☆633Updated this week
- ☆254Updated last week
- This repo includes everything you need to know about deploying GPU nodes on OCI☆35Updated last week