NVIDIA / nvidia-container-toolkit
Build and run containers leveraging NVIDIA GPUs
☆3,032Updated this week
Alternatives and similar repositories for nvidia-container-toolkit:
Users that are interested in nvidia-container-toolkit are comparing it to the libraries listed below
- NVIDIA container runtime library☆918Updated this week
- NVIDIA container runtime☆1,116Updated last year
- NVIDIA device plugin for Kubernetes☆3,098Updated this week
- NVIDIA GPU Operator creates, configures, and manages GPUs in Kubernetes☆2,047Updated this week
- An Open Source Machine Learning Framework for Everyone☆1,112Updated 6 months ago
- AIStore: scalable storage for AI applications☆1,441Updated this week
- NVIDIA GPU metrics exporter for Prometheus leveraging DCGM☆1,101Updated last week
- Simple, safe way to store and distribute tensors☆3,185Updated last week
- Multi-GPU CUDA stress test☆1,610Updated 7 months ago
- Optimized primitives for collective multi-GPU communication☆3,587Updated last week
- A Python package for extending the official PyTorch that can easily obtain performance on Intel platform☆1,793Updated this week
- DirectML is a high-performance, hardware-accelerated DirectX 12 library for machine learning. DirectML provides GPU acceleration for comm…☆2,392Updated 4 months ago
- Samples for CUDA Developers which demonstrates features in CUDA Toolkit☆7,194Updated 2 weeks ago
- NVIDIA Data Center GPU Manager (DCGM) is a project for gathering telemetry and measuring the health of NVIDIA GPUs☆480Updated last month
- SGLang is a fast serving framework for large language models and vision language models.☆12,427Updated this week
- Transformer related optimization, including BERT, GPT☆6,095Updated last year
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs…☆2,299Updated this week
- CUDA Python: Performance meets Productivity☆1,245Updated this week
- A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM☆2,930Updated 7 months ago
- A machine learning compiler for GPUs, CPUs, and ML accelerators☆3,045Updated this week
- An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.☆5,324Updated last week
- Accessible large language models via k-bit quantization for PyTorch.☆6,836Updated last week
- NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source compone…☆11,373Updated 2 weeks ago
- Hackable and optimized Transformers building blocks, supporting a composable construction.☆9,215Updated last week
- TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain…☆9,975Updated this week
- cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it☆531Updated last week
- AMD ROCm™ Software - GitHub Home☆5,125Updated this week
- Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.☆496Updated 3 weeks ago
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆8,958Updated this week
- FlashInfer: Kernel Library for LLM Serving☆2,483Updated this week