NVIDIA / topographLinks
A toolkit for discovering cluster network topology.
☆76Updated last week
Alternatives and similar repositories for topograph
Users that are interested in topograph are comparing it to the libraries listed below
Sorting:
- Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscaling☆86Updated this week
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆131Updated this week
- Inference scheduler for llm-d☆102Updated last week
- JobSet: a k8s native API for distributed ML training and HPC workloads☆276Updated last week
- GenAI inference performance benchmarking tool☆110Updated this week
- knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.☆71Updated 3 months ago
- Holistic job manager on Kubernetes☆116Updated last year
- Distributed KV cache coordinator☆82Updated last week
- NVIDIA DRA Driver for GPUs☆477Updated this week
- Example DRA driver that developers can fork and modify to get them started writing their own.☆100Updated 2 weeks ago
- Golang bindings for Nvidia Datacenter GPU Manager (DCGM)☆138Updated 2 weeks ago
- Gateway API Inference Extension☆514Updated this week
- Go Abstraction for Allocating NVIDIA GPUs with Custom Policies☆117Updated last month
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆899Updated this week
- ☆159Updated 2 weeks ago
- MIG Partition Editor for NVIDIA GPUs☆224Updated this week
- Enabling Kubernetes to make pod placement decisions with platform intelligence.☆176Updated 9 months ago
- NVIDIA NCCL Tests for Distributed Training☆121Updated last week
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆604Updated last week
- CUDA checkpoint and restore utility☆381Updated last month
- WG Serving☆31Updated 3 weeks ago
- ☆267Updated 3 weeks ago
- 🧯 Kubernetes coverage for fault awareness and recovery, works for any LLMOps, MLOps, AI workloads.☆33Updated 3 weeks ago
- OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)☆307Updated last week
- Simplified model deployment on llm-d☆27Updated 4 months ago
- A tool to detect infrastructure issues on cloud native AI systems☆49Updated last month
- NVIDIA Network Operator☆289Updated this week
- GPUd automates monitoring, diagnostics, and issue identification for GPUs☆442Updated last week
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆264Updated last week
- The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.☆138Updated 2 weeks ago