Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
☆2,506Mar 2, 2026Updated this week
Alternatives and similar repositories for llama-swap
Users that are interested in llama-swap are comparing it to the libraries listed below
Sorting:
- llama.cpp fork with additional SOTA quants and improved performance☆1,696Updated this week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,139Feb 9, 2026Updated 3 weeks ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆88Updated this week
- One command brings a complete pre-wired LLM stack with hundreds of services to explore.☆2,482Updated this week
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,444Dec 9, 2025Updated 2 months ago
- Run GGUF models easily with a KoboldAI UI. One File. Zero Install.☆9,594Updated this week
- Large-scale LLM inference engine☆1,666Updated this week
- Open-source LLM load balancer and serving platform for self-hosting LLMs at scale 🏓🦙 Alternative to projects like llm-d, Docker Model R…☆1,474Updated this week
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆52Aug 21, 2025Updated 6 months ago
- Manifold is an experimental platform for enabling long horizon workflow automation using teams of AI assistants.☆482Feb 28, 2026Updated last week
- Fast, flexible LLM inference☆6,653Feb 27, 2026Updated last week
- Optimizing inference proxy for LLMs☆3,352Jan 28, 2026Updated last month
- VS Code extension for LLM-assisted code/text completion☆1,186Updated this week
- LLM inference in C/C++☆96,322Updated this week
- Plano is an AI-native proxy and data plane for agentic apps — with built-in orchestration, safety, observability, and smart LLM routing s…☆5,841Updated this week
- Docker/podman container for llama.cpp/vllm/exllamav{2,3} orchestrated using llama-swap☆17Feb 22, 2026Updated last week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆638Updated this week
- ☆53Oct 10, 2025Updated 4 months ago
- ☆92Dec 9, 2025Updated 2 months ago
- InferX: Inference as a Service Platform☆163Feb 28, 2026Updated last week
- WilmerAI is one of the oldest LLM semantic routers. It uses multi-layer prompt routing and complex workflows to allow you to not only cre…☆804Feb 9, 2026Updated 3 weeks ago
- LLM Frontend in a single html file☆704Dec 27, 2025Updated 2 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆1,009Dec 17, 2025Updated 2 months ago
- ☆229May 7, 2025Updated 9 months ago
- Speech-to-speech AI assistant with natural conversation flow, mid-speech interruption, vision capabilities and AI-initiated follow-ups. F…☆289Apr 14, 2025Updated 10 months ago
- tl/dw (Too Long, Didn't Watch): Your Personal Research Multi-Tool - a naive attempt at 'A Young Lady's Illustrated Primer' (Open Source N…☆1,277Updated this week
- ☆3,002Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆167May 16, 2024Updated last year
- Easy to use interface for the Whisper model optimized for all GPUs!☆488Feb 15, 2026Updated 2 weeks ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆283Jan 6, 2026Updated 2 months ago
- Web UI for ExLlamaV2☆512Feb 5, 2025Updated last year
- Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek, Qwen, Llama, Gemma, TTS 2x faster with 70% less VRAM.☆53,029Updated this week
- RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for in…☆2,624Updated this week
- Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs. Join our discord: https…☆2,315Updated this week
- ☆177Aug 10, 2025Updated 6 months ago
- The Fastest Way to Fine-Tune LLMs Locally☆334Dec 18, 2025Updated 2 months ago
- Claraverse is a opesource privacy focused ecosystem to replace ChatGPT, Claude, N8N, ImageGen with your own hosted llm, keys and compute.…☆3,721Feb 28, 2026Updated last week
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,842Feb 10, 2026Updated 3 weeks ago
- llama-swap + a minimal ollama compatible api☆51Feb 13, 2026Updated 3 weeks ago