Build and run containers leveraging NVIDIA GPUs
☆4,088Updated this week
Alternatives and similar repositories for nvidia-container-toolkit
Users that are interested in nvidia-container-toolkit are comparing it to the libraries listed below
Sorting:
- NVIDIA container runtime library☆1,072Updated this week
- Build and run Docker containers leveraging NVIDIA GPUs☆17,498Dec 6, 2023Updated 2 years ago
- NVIDIA device plugin for Kubernetes☆3,671Updated this week
- NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes☆2,549Updated this week
- NVIDIA GPU metrics exporter for Prometheus leveraging DCGM☆1,626Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆71,234Updated this week
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizat…☆12,938Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆10,393Updated this week
- NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs☆671Feb 17, 2026Updated last week
- Samples for CUDA Developers which demonstrates features in CUDA Toolkit☆8,870Jan 6, 2026Updated last month
- NVIDIA container runtime☆1,124Oct 27, 2023Updated 2 years ago
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆12,723Updated this week
- Heterogeneous AI Computing Virtualization Middleware(Project under CNCF)☆3,032Feb 18, 2026Updated last week
- Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.☆163,632Updated this week
- An open and reliable container runtime☆20,385Updated this week
- ☆285Feb 20, 2026Updated last week
- GPU plugin to the node feature discovery for Kubernetes☆307May 27, 2024Updated last year
- The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.☆160Updated this week
- AIStore: scalable storage for AI applications☆1,766Updated this week
- LLM inference in C/C++☆95,726Updated this week
- NVIDIA Linux open GPU kernel module source☆16,743Updated this week
- A Cloud Native Batch System (Project under CNCF)☆5,340Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.☆23,905Updated this week
- ☆336Feb 22, 2026Updated last week
- A toolkit to run Ray applications on Kubernetes☆2,341Updated this week
- ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator☆19,389Updated this week
- A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Auto…☆16,807Updated this week
- Podman: A tool for managing OCI containers and pods.☆30,824Updated this week
- Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search☆43,056Updated this week
- contaiNERD CTL - Docker-compatible CLI for containerd, with support for Compose, Rootless, eStargz, OCIcrypt, IPFS, ...☆9,858Updated this week
- Optimized primitives for collective multi-GPU communication☆4,474Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆41,706Updated this week
- Kubernetes Virtualization API and runtime in order to define and manage virtual machines.☆6,675Updated this week
- Standardized Distributed Generative and Predictive AI Inference Platform for Scalable, Multi-Framework Deployment on Kubernetes☆5,135Updated this week
- Go Bindings for the NVIDIA Management Library (NVML)☆425Feb 12, 2026Updated 2 weeks ago
- A collection of useful Go libraries for use with NVIDIA GPU management tools☆50Jan 15, 2026Updated last month
- An Open Source Machine Learning Framework for Everyone☆1,150Jul 31, 2025Updated 7 months ago
- Development repository for the Triton language and compiler☆18,501Updated this week
- Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.☆41,516Updated this week