nebuly-ai / nos
Module to Automatically maximize the utilization of GPU resources in a Kubernetes cluster through real-time dynamic partitioning and elastic quotas - Effortless optimization at its finest!
☆651Updated 11 months ago
Alternatives and similar repositories for nos:
Users that are interested in nos are comparing it to the libraries listed below
- NVIDIA device plugin for Kubernetes☆48Updated last year
- GPU environment and cluster management with LLM support☆595Updated 10 months ago
- JobSet: a k8s native API for distributed ML training and HPC workloads☆211Updated this week
- Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes☆332Updated this week
- User documentation for KServe.☆105Updated 2 weeks ago
- Controller for ModelMesh☆226Updated last week
- GPU plugin to the node feature discovery for Kubernetes☆299Updated 10 months ago
- Distributed Model Serving Framework☆159Updated 2 weeks ago
- A curated list of awesome projects and resources related to Kubeflow (a CNCF incubating project)☆207Updated 4 months ago
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆354Updated this week
- Module, Model, and Tensor Serialization/Deserialization☆220Updated last month
- Kubernetes Operator for MPI-based applications (distributed training, HPC, etc.)☆469Updated 2 weeks ago
- An inference server for your machine learning models, including support for multiple frameworks, multi-model serving and more☆793Updated this week
- Kubeflow Deployment Manifests☆894Updated this week
- elastic-gpu-scheduler is a Kubernetes scheduler extender for GPU resources scheduling.☆140Updated 2 years ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆465Updated 3 weeks ago
- deployKF builds machine learning platforms on Kubernetes. We combine the best of Kubeflow, Airflow†, and MLflow† into a complete platform…☆415Updated 8 months ago
- PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.☆785Updated last month
- AWS virtual gpu device plugin provides capability to use smaller virtual gpus for your machine learning inference workloads☆204Updated last year
- Repository for open inference protocol specification☆52Updated 8 months ago
- Curated list of awesome material on optimization techniques to make artificial intelligence faster and more efficient 🚀☆113Updated last year
- Run Slurm in Kubernetes☆200Updated this week
- A toolkit to run Ray applications on Kubernetes☆1,619Updated this week
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆199Updated 2 months ago
- MIG Partition Editor for NVIDIA GPUs☆192Updated this week
- AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-te…☆862Updated this week
- A top-like tool for monitoring GPUs in a cluster☆87Updated last year
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆189Updated this week
- A multi-cluster batch queuing system for high-throughput workloads on Kubernetes.☆508Updated this week
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆356Updated this week