NVIDIA / nim-deploy
A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deployment.
☆179Updated this week
Alternatives and similar repositories for nim-deploy:
Users that are interested in nim-deploy are comparing it to the libraries listed below
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆159Updated last week
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆100Updated this week
- ☆207Updated this week
- This NVIDIA RAG blueprint serves as a reference solution for a foundational Retrieval Augmented Generation (RAG) pipeline.☆89Updated this week
- ☆152Updated this week
- Infrastructure as code for GPU accelerated managed Kubernetes clusters.☆55Updated last week
- Controller for ModelMesh☆229Updated this week
- Self-host LLMs with vLLM and BentoML☆109Updated this week
- Run cloud native workloads on NVIDIA GPUs☆168Updated 2 weeks ago
- Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inferen…☆62Updated 2 weeks ago
- Repository for open inference protocol specification☆54Updated 9 months ago
- Tutorial for building LLM router☆202Updated 9 months ago
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆291Updated this week
- Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.☆229Updated this week
- Using LlamaIndex with Ray for productionizing LLM applications☆71Updated last year
- Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes☆353Updated this week
- Helm charts for the KubeRay project☆43Updated last month
- 🚀 Use NVIDIA NIMs with Haystack pipelines☆31Updated 8 months ago
- Fine-tune an LLM to perform batch inference and online serving.☆110Updated last week
- Containerization and cloud native suite for OPEA☆57Updated this week
- Model Registry provides a single pane of glass for ML model developers to index and manage models, versions, and ML artifacts metadata. I…☆119Updated this week
- ☆46Updated last week
- ☆253Updated this week
- Kubernetes Operator, ansible playbooks, and production scripts for large-scale AIStore deployments on Kubernetes.☆94Updated this week
- Collection of reference workflows for building intelligent agents with NIMs☆155Updated 3 months ago
- MIG Partition Editor for NVIDIA GPUs☆198Updated this week
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆539Updated this week
- AI21 Python SDK☆63Updated this week
- Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.☆501Updated 3 weeks ago
- Distributed Model Serving Framework☆165Updated this week