InftyAI / Awesome-LLMOpsLinks
π An awesome & curated list of best LLMOps tools.
β190Updated this week
Alternatives and similar repositories for Awesome-LLMOps
Users that are interested in Awesome-LLMOps are comparing it to the libraries listed below
Sorting:
- βΈοΈ Easy, advanced inference platform for large language models on Kubernetes. π Star to support our work!β287Updated 2 weeks ago
- agent-sandbox enables easy management of isolated, stateful, singleton workloads, ideal for use cases like AI agent runtimes.β875Updated this week
- Kubernetes-native AI serving platform for scalable model serving.β198Updated last week
- π« A lightweight p2p-based cache system for model distributions on Kubernetes. Reframing now to make it an unified cache system with POSIβ¦β25Updated last year
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.β146Updated this week
- WG Servingβ34Updated last month
- A diverse, simple, and secure all-in-one LLMOps platformβ109Updated last year
- llm-d helm charts and deployment examplesβ48Updated last month
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replicationβ656Updated last week
- GenAI inference performance benchmarking toolβ142Updated last week
- Cloud Native Artifacial Intelligence Model Format Specificationβ175Updated last week
- knavigator is a development, testing, and optimization toolkit for AI/ML scheduling systems at scale on Kubernetes.β74Updated 6 months ago
- Containerization and cloud native suite for OPEAβ74Updated last month
- A workload for deploying LLM inference services on Kubernetesβ168Updated last week
- Large language model fine-tuning capabilities based on cloud native and distributed computing.β92Updated last year
- Command-line tools for managing OCI model artifacts, which are bundled based on Model Specβ61Updated this week
- A toolkit for discovering cluster network topology.β96Updated last week
- This is a landscape of the infrastructure that powers the generative AI ecosystemβ151Updated last year
- Inference scheduler for llm-dβ124Updated last week
- Open Model Engine (OME) β Kubernetes operator for LLM serving, GPU scheduling, and model lifecycle management. Works with SGLang, vLLM, Tβ¦β365Updated this week
- Distributed KV cache scheduling & offloading librariesβ101Updated last week
- Device-plugin for volcano vgpu which support hard resource isolationβ143Updated last month
- Kubernetes enhancements for Network Topology Aware Gang Scheduling & Autoscalingβ159Updated this week
- Rapid and cost-effective operator and best practice for agent sandbox lifecycle management.β99Updated this week
- π§― Kubernetes coverage for fault awareness and recovery, works for any LLMOps, MLOps, AI workloads.β35Updated this week
- A federation scheduler for multi-clusterβ61Updated last week
- The main purpose of runtime copilot is to assist with node runtime management tasks such as configuring registries, upgrading versions, iβ¦β12Updated 2 years ago
- CLI tool and Kubernetes Controller for building, testing and deploying MCP serversβ417Updated this week
- Extensible generative AI platform on Kubernetes with OpenAI-compatible APIs.β93Updated 3 months ago
- Gateway API Inference Extensionβ576Updated this week