ray-project / serve_config_examplesLinks
☆10Updated last year
Alternatives and similar repositories for serve_config_examples
Users that are interested in serve_config_examples are comparing it to the libraries listed below
Sorting:
- A curated list of awesome projects and resources related to Kubeflow (a CNCF incubating project)☆222Updated last month
- Helm charts for the KubeRay project☆59Updated 2 months ago
- llm-d helm charts and deployment examples☆48Updated last month
- KServe models web UI☆47Updated last week
- markdown docs☆93Updated last week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆221Updated last week
- Repository for open inference protocol specification☆64Updated 8 months ago
- Controller for ModelMesh☆242Updated 8 months ago
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆146Updated last week
- Model Registry provides a single pane of glass for ML model developers to index and manage models, versions, and ML artifacts metadata. I…☆161Updated this week
- Run cloud native workloads on NVIDIA GPUs☆223Updated 2 weeks ago
- Distributed Model Serving Framework☆185Updated 4 months ago
- NVIDIA DRA Driver for GPUs☆557Updated this week
- ☆134Updated last week
- AWS virtual gpu device plugin provides capability to use smaller virtual gpus for your machine learning inference workloads☆204Updated 2 years ago
- ModelMesh Performance Scripts, Dashboard and Pipelines☆12Updated 8 months ago
- deployKF builds machine learning platforms on Kubernetes. We combine the best of Kubeflow, Airflow†, and MLflow† into a complete platform…☆466Updated last year
- User documentation for KServe.☆109Updated last week
- Helm charts for llm-d☆52Updated 6 months ago
- KubeFlow on AWS☆186Updated last week
- Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inferen…☆73Updated last week
- GenAI inference performance benchmarking tool☆142Updated last week
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆130Updated 4 months ago
- Charmed Kubeflow examples☆45Updated last year
- Fork of NVIDIA device plugin for Kubernetes with support for shared GPUs by declaring GPUs multiple times☆87Updated 3 years ago
- JobSet: a k8s native API for distributed ML training and HPC workloads☆308Updated this week
- Argoflow has been superseded by deployKF☆134Updated 2 years ago
- ☆191Updated 3 weeks ago
- This repository aims to showcase how to finetune a FM model in Amazon EKS cluster using, JupyterHub to provision notebooks and craft both…☆52Updated 7 months ago
- Gateway API Inference Extension☆576Updated this week