Cost-efficient and pluggable Infrastructure components for GenAI inference
☆4,714Apr 10, 2026Updated this week
Alternatives and similar repositories for aibrix
Users that are interested in aibrix are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- A Datacenter Scale Distributed Inference Serving Framework☆6,527Updated this week
- vLLM’s reference system for K8S-native cluster-wide deployment with community-driven performance optimization☆2,267Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.☆25,643Updated this week
- Gateway API Inference Extension☆639Updated this week
- LeaderWorkerSet: An API for deploying a group of pods as a unit of replication☆693Updated this week
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Supercharge Your LLM with the Fastest KV Cache Layer☆7,969Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆75,637Updated this week
- Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.☆5,071Updated this week
- A toolkit to run Ray applications on Kubernetes☆2,432Updated this week
- FlashInfer: Kernel Library for LLM Serving☆5,372Updated this week
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, Slurm, 20+ cl…☆9,822Updated this week
- NVIDIA Inference Xfer Library (NIXL)☆970Updated this week
- A high-performance distributed file system designed to address the challenges of AI training and inference workloads.☆9,800Mar 30, 2026Updated 2 weeks ago
- Standardized Distributed Generative and Predictive AI Inference Platform for Scalable, Multi-Framework Deployment on Kubernetes☆5,305Updated this week
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- KV cache store for distributed LLM inference☆402Nov 13, 2025Updated 5 months ago
- Achieve state of the art inference performance with modern accelerators on Kubernetes☆2,957Updated this week
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizat…☆13,354Updated this week
- DeepEP: an efficient expert-parallel communication library☆9,105Updated this week
- Heterogeneous GPU Sharing on Kubernetes☆3,257Updated this week
- Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation☆7,971May 15, 2025Updated 10 months ago
- Composable building blocks to build LLM Apps☆8,324Updated this week
- AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports VLMs, LLMs, embeddings, and speech-to-te…☆1,179Mar 31, 2026Updated last week
- Unsloth Studio is a web UI for training and running open models like Gemma 4, Qwen3.5, DeepSeek, gpt-oss locally.☆61,312Updated this week
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- A Cloud Native Batch System (Project under CNCF)☆5,440Updated this week
- KAI Scheduler is an open source Kubernetes Native scheduler for AI workloads at large scale☆1,221Updated this week
- FlashMLA: Efficient Multi-head Latent Attention Kernels☆12,558Apr 7, 2026Updated last week
- Large Language Model Text Generation Inference☆10,830Mar 21, 2026Updated 3 weeks ago
- Open Model Engine (OME) — Kubernetes operator for LLM serving, GPU scheduling, and model lifecycle management. Works with SGLang, vLLM, T…☆413Updated this week
- verl: Volcano Engine Reinforcement Learning for LLMs☆20,603Updated this week
- Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing a…☆42,652Updated this week
- Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM☆2,996Updated this week
- A QoS-based scheduling system brings optimal layout and status to workloads such as microservices, web services, big data jobs, AI jobs, …☆1,669Updated this week
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- A minimal Python framework for building custom AI inference servers with full control over logic, batching, and scaling.☆3,854Updated this week
- A lightweight data processing framework built on DuckDB and 3FS.☆4,946Mar 5, 2025Updated last year
- A throughput-oriented high-performance serving framework for LLMs☆953Mar 29, 2026Updated 2 weeks ago
- LMDeploy is a toolkit for compressing, deploying, and serving LLMs.☆7,775Updated this week
- Easily fine-tune, evaluate and deploy gpt-oss, Qwen3, DeepSeek-R1, or any open source LLM / VLM!☆9,171Updated this week
- An invoice generator app built using Next.js, Typescript, and Shadcn☆6,205Apr 2, 2026Updated last week
- Efficient and easy multi-instance LLM serving☆541Mar 12, 2026Updated last month