kserve / modelmesh
Distributed Model Serving Framework
☆157Updated 3 months ago
Alternatives and similar repositories for modelmesh:
Users that are interested in modelmesh are comparing it to the libraries listed below
- Controller for ModelMesh☆214Updated 3 weeks ago
- User documentation for KServe.☆103Updated this week
- ☆100Updated this week
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆176Updated this week
- Kubeflow Pipelines on Tekton☆175Updated 2 months ago
- Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes☆305Updated this week
- Holistic job manager on Kubernetes☆111Updated 11 months ago
- Repository for open inference protocol specification☆45Updated 6 months ago
- KServe models web UI☆35Updated 4 months ago
- JobSet: a k8s native API for distributed ML training and HPC workloads☆175Updated this week
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆193Updated 2 weeks ago
- Unified runtime-adapter image of the sidecar containers which run in the modelmesh pods☆21Updated last week
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆79Updated this week
- elastic-gpu-scheduler is a Kubernetes scheduler extender for GPU resources scheduling.☆137Updated 2 years ago
- GPU plugin to the node feature discovery for Kubernetes☆296Updated 8 months ago
- Fork of NVIDIA device plugin for Kubernetes with support for shared GPUs by declaring GPUs multiple times☆88Updated 2 years ago
- Kubernetes Operator for MPI-based applications (distributed training, HPC, etc.)☆452Updated last week
- Device plugins for Volcano, e.g. GPU☆113Updated 4 months ago
- Gateway API Inference Extension☆129Updated this week
- Helm charts for the KubeRay project☆37Updated last week
- AWS virtual gpu device plugin provides capability to use smaller virtual gpus for your machine learning inference workloads☆203Updated last year
- Automatic tuning for ML model deployment on Kubernetes☆80Updated 2 months ago
- MIG Partition Editor for NVIDIA GPUs☆185Updated this week
- ☆113Updated 2 years ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆448Updated 2 weeks ago
- markdown docs☆75Updated this week
- NVIDIA device plugin for Kubernetes☆47Updated 11 months ago
- Kubernetes Operator, ansible playbooks, and production scripts for large-scale AIStore deployments on Kubernetes.☆88Updated last week
- InstaSlice Operator facilitates slicing of accelerators using stable APIs☆27Updated this week
- Collection of tools and examples for managing Accelerated workloads in Kubernetes Engine☆223Updated last week