vllm-project / vllm-metalLinks
Community maintained hardware plugin for vLLM on Apple Silicon
☆62Updated this week
Alternatives and similar repositories for vllm-metal
Users that are interested in vllm-metal are comparing it to the libraries listed below
Sorting:
- A modern web interface for managing and interacting with vLLM servers (www.github.com/vllm-project/vllm). Supports both GPU and CPU modes…☆172Updated this week
- vLLM adapter for a TGIS-compatible gRPC server.☆45Updated this week
- Common recipes to run vLLM☆283Updated last week
- A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM☆160Updated last week
- Benchmark suite for LLMs from Fireworks.ai☆84Updated last month
- A command-line interface tool for serving LLM using vLLM.☆456Updated 3 weeks ago
- Verify Precision of all Kimi K2 API Vendor☆489Updated last month
- ArcticInference: vLLM plugin for high-throughput, low-latency inference☆354Updated this week
- Inference server benchmarking tool☆132Updated 2 months ago
- A collection of all available inference solutions for the LLMs☆93Updated 9 months ago
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang/tree/main/docs.☆95Updated this week
- Easy, Fast, and Scalable Multimodal AI☆81Updated this week
- ☆63Updated 7 months ago
- Self-host LLMs with vLLM and BentoML☆162Updated 3 weeks ago
- The LLM abstraction layer for modern AI agent applications.☆496Updated this week
- An early research stage expert-parallel load balancer for MoE models based on linear programming.☆469Updated last month
- Memory optimized Mixture of Experts☆72Updated 4 months ago
- The driver for LMCache core to run in vLLM☆59Updated 10 months ago
- Fused Qwen3 MoE layer for faster training, compatible with HF Transformers, LoRA, 4-bit quant, Unsloth☆217Updated last month
- Self-host LLMs with LMDeploy and BentoML☆21Updated 5 months ago
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆871Updated this week
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆261Updated this week
- ☆51Updated last year
- [NeurIPS'25 Oral] Query-agnostic KV cache eviction: 3–4× reduction in memory and 2× decrease in latency (Qwen3/2.5, Gemma3, LLaMA3)☆169Updated 3 weeks ago
- vLLM performance dashboard☆39Updated last year
- KV cache compression for high-throughput LLM inference☆148Updated 10 months ago
- [NeurIPS 2025] Simple extension on vLLM to help you speed up reasoning model without training.☆215Updated 6 months ago
- Library for model distillation☆158Updated 3 months ago
- ☆219Updated 11 months ago
- Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache☆136Updated 4 months ago