Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
☆3,772May 1, 2026Updated this week
Alternatives and similar repositories for llama-swap
Users that are interested in llama-swap are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- llama.cpp fork with additional SOTA quants and improved performance☆2,276Updated this week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,205Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆90Apr 20, 2026Updated 2 weeks ago
- Stop configuring your AI stack. Start using it. One command brings a complete pre-wired LLM stack with hundreds of services to explore.☆2,902Updated this week
- Large-scale LLM inference engine☆1,719Updated this week
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- LLM inference in C/C++