NVIDIA / nvidia-container-toolkitLinks
Build and run containers leveraging NVIDIA GPUs
☆3,901Updated last week
Alternatives and similar repositories for nvidia-container-toolkit
Users that are interested in nvidia-container-toolkit are comparing it to the libraries listed below
Sorting:
- NVIDIA container runtime library☆1,049Updated last week
- NVIDIA device plugin for Kubernetes☆3,584Updated this week
- NVIDIA GPU metrics exporter for Prometheus leveraging DCGM☆1,521Updated last week
- NVIDIA container runtime☆1,124Updated 2 years ago
- NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes☆2,448Updated this week
- An Open Source Machine Learning Framework for Everyone☆1,149Updated 4 months ago
- Simple, safe way to store and distribute tensors☆3,547Updated this week
- Multi-GPU CUDA stress test☆2,025Updated last month
- Nvidia GPU exporter for prometheus using nvidia-smi binary☆1,336Updated last week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,995Updated last week
- Docker CLI plugin for extended build capabilities with BuildKit☆4,191Updated this week
- Samples for CUDA Developers which demonstrates features in CUDA Toolkit☆8,564Updated 3 months ago
- NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs☆627Updated last week
- Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.☆2,203Updated this week
- dockerd as a compliant Container Runtime Interface for Kubernetes☆1,318Updated last month
- Heterogeneous AI Computing Virtualization Middleware(Project under CNCF)☆2,747Updated this week
- A Datacenter Scale Distributed Inference Serving Framework☆5,617Updated this week
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizat…☆12,312Updated last week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,203Updated this week
- Optimized primitives for collective multi-GPU communication☆4,289Updated last week
- GPU & Accelerator process monitoring for AMD, Apple, Huawei, Intel, NVIDIA and Qualcomm☆9,834Updated last month
- Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.☆667Updated this week
- AIStore: scalable storage for AI applications☆1,691Updated this week
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆3,794Updated this week
- OpenVINO™ is an open source toolkit for optimizing and deploying AI inference☆9,325Updated this week
- A toolkit to run Ray applications on Kubernetes☆2,190Updated this week
- On-device AI across mobile, embedded and edge for PyTorch☆3,634Updated last week
- Python bindings for llama.cpp☆9,800Updated 3 months ago
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆10,106Updated this week
- Large Language Model Text Generation Inference☆10,693Updated this week