pwilkin / llama-runnerLinks
Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends
☆48Updated 2 months ago
Alternatives and similar repositories for llama-runner
Users that are interested in llama-runner are comparing it to the libraries listed below
Sorting:
- llm client, server and agent☆73Updated this week
- The easiest & fastest way to run LLMs in your home lab☆70Updated 2 months ago
- A local front-end for open-weight LLMs with memory, RAG, TTS/STT, Elo ratings, and dynamic research tools. Built with React and FastAPI.☆37Updated 2 months ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆107Updated last week
- A web application that converts speech to speech 100% private☆77Updated 4 months ago
- ☆168Updated 2 months ago
- Orpheus Chat WebUI☆74Updated 7 months ago
- 🚀 FlexLLama - Lightweight self-hosted tool for running multiple llama.cpp server instances with OpenAI v1 API compatibility and multi-GP…☆39Updated this week
- ☆18Updated 3 months ago
- ☆83Updated 8 months ago
- Cognito: Supercharge your Chrome browser with AI. Guide, query, and control everything using natural language.☆55Updated last month
- A persistent local memory for AI, LLMs, or Copilot in VS Code.☆163Updated this week
- A real-time shared memory layer for multi-agent LLM systems.☆48Updated 4 months ago
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆224Updated 2 months ago
- Capture, tag, and search images locally with OSS models.☆44Updated 9 months ago
- Open source tool for transcirption and subtitling, alternative to happyscribe.☆30Updated 8 months ago
- Autonomous, agentic, creative story writing system that incorporates stored embeddings and Knowledge Graphs.☆80Updated 3 weeks ago
- ☆28Updated 4 months ago
- Give your local LLM a real memory with a lightweight, fully local memory system. 100% offline and under your control.☆59Updated last month
- Python language chat with Ollama models locally, anthropic and openai☆24Updated 6 months ago
- ☆206Updated last month
- Simple node proxy for llama-server that enables MCP use☆13Updated 5 months ago
- A natural language file search tool that uses LLMs to help you find files by describing what you're looking for.☆28Updated 7 months ago
- Run Orpheus 3B Locally With LM Studio☆31Updated 7 months ago
- Run Orpheus 3B Locally with Gradio UI, Standalone App☆21Updated 6 months ago
- A TTS model capable of generating ultra-realistic dialogue in one pass.☆31Updated 5 months ago
- Personal voice assistant, with voice interruption and Twilio support☆18Updated 8 months ago
- 🗣️ Real‑time, low‑latency voice, vision, and conversational‑memory AI assistant built on LiveKit and local LLMs ✨☆95Updated 4 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated last week
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆46Updated 2 months ago