sgl-project / omeLinks
OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)
☆174Updated this week
Alternatives and similar repositories for ome
Users that are interested in ome are comparing it to the libraries listed below
Sorting:
- GPUd automates monitoring, diagnostics, and issue identification for GPUs☆387Updated this week
- KV cache store for distributed LLM inference☆290Updated last month
- Genai-bench is a powerful benchmark tool designed for comprehensive token-level performance evaluation of large language model (LLM) serv…☆169Updated this week
- ☆29Updated 2 months ago
- ☸️ Easy, advanced inference platform for large language models on Kubernetes. 🌟 Star to support our work!☆224Updated last week
- The driver for LMCache core to run in vLLM☆44Updated 5 months ago
- Cloud Native Benchmarking of Foundation Models☆38Updated last month
- 🧯 Kubernetes coverage for fault awareness and recovery, works for any LLMOps, MLOps, AI workloads.☆30Updated last week
- NVIDIA Inference Xfer Library (NIXL)☆473Updated this week
- NVIDIA NCCL Tests for Distributed Training☆97Updated 3 weeks ago
- CUDA checkpoint and restore utility☆346Updated 5 months ago
- A light weight vLLM simulator, for mocking out replicas.☆30Updated this week
- Efficient and easy multi-instance LLM serving☆448Updated this week
- A benchmarking tool for comparing different LLM API providers' DeepSeek model deployments.☆28Updated 3 months ago
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆129Updated last week
- A toolkit for discovering cluster network topology.☆56Updated last week