Open Model Engine (OME) — Kubernetes operator for LLM serving, GPU scheduling, and model lifecycle management. Works with SGLang, vLLM, TensorRT-LLM, and Triton
☆435Apr 28, 2026Updated this week
Alternatives and similar repositories for ome
Users that are interested in ome are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Genai-bench is a powerful benchmark tool designed for comprehensive token-level performance evaluation of large language model (LLM) serv…☆294Updated this week
- A workload for deploying LLM inference services on Kubernetes☆210Updated this week
- Following the same workflows as Kubernetes. Widely used in InftyAI community.☆13Dec 5, 2025Updated 4 months ago
- Virtualized Elastic KV Cache for Dynamic GPU Sharing and Beyond☆888Apr 21, 2026Updated last week
- Gateway API Inference Extension☆657Updated this week
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆704Apr 21, 2026Updated last week
- NVIDIA Inference Xfer Library (NIXL)☆1,003Updated this week
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆301Jan 26, 2026Updated 3 months ago
- Achieve state of the art inference performance with modern accelerators on Kubernetes☆3,069Updated this week
- A lightweight, configurable, and real-time simulator designed to mimic the behavior of vLLM without the need for GPUs or running actual h…☆119Updated this week
- An Envoy inspired, ultimate LLM-first gateway for LLM serving and downstream application developers and enterprises☆26Apr 24, 2025Updated last year
- A Datacenter Scale Distributed Inference Serving Framework☆6,634Updated this week
- Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.☆5,186Updated this week
- WG Serving☆35Mar 24, 2026Updated last month
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- 💫 A lightweight p2p-based cache system for model distributions on Kubernetes. Reframing now to make it an unified cache system with POSI…☆26Dec 6, 2024Updated last year
- Materials for learning SGLang☆806Jan 5, 2026Updated 3 months ago
- vLLM’s reference system for K8S-native cluster-wide deployment with community-driven performance optimization☆2,299Updated this week
- Offline optimization of your disaggregated Dynamo graph☆274Updated this week
- The main purpose of runtime copilot is to assist with node runtime management tasks such as configuring registries, upgrading versions, i…☆12May 16, 2023Updated 2 years ago
- https://bbuf.github.io/gpu-glossary-zh/☆26Nov 7, 2025Updated 5 months ago
- Kubernetes-native AI serving platform for scalable model serving.☆315Updated this week
- Fast and memory-efficient exact attention☆21Apr 10, 2026Updated 2 weeks ago
- KV cache store for distributed LLM inference☆410Nov 13, 2025Updated 5 months ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Efficient and easy multi-instance LLM serving☆547Mar 12, 2026Updated last month
- Cost-efficient and pluggable Infrastructure components for GenAI inference☆4,756Updated this week
- DRA Driver for NVIDIA GPUs☆633Updated this week
- The Intelligent Inference Scheduler for Large-scale Inference Services.☆67Feb 12, 2026Updated 2 months ago
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆1,245Updated this week
- Supercharge Your LLM with the Fastest KV Cache Layer☆8,132Updated this week
- Kubernetes APIServer 高性能代理组件,代理 APIServer 的 List 请求,其它类型的请求会直接反向代理到原生 APIServer。 CKube 还额外支持了分页、搜索和索引等功能。 并且,CKube 100% 兼容原生 kubectl 和 ku…☆19Sep 16, 2022Updated 3 years ago
- Simplified Data Management and Sharing for Kubernetes☆18Apr 23, 2026Updated last week
- GenAI inference performance benchmarking tool☆178Updated this week
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆945Feb 28, 2026Updated 2 months ago
- Train speculative decoding models effortlessly and port them smoothly to SGLang serving.☆801Apr 2, 2026Updated 3 weeks ago
- 🎉 An awesome & curated list of best LLMOps tools.☆227Updated this week
- AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-te…☆1,186Mar 31, 2026Updated 3 weeks ago
- A toolkit to run Ray applications on Kubernetes☆2,466Updated this week
- Standardized Distributed Generative and Predictive AI Inference Platform for Scalable, Multi-Framework Deployment on Kubernetes☆5,395Updated this week
- ☆196Updated this week