thushan / ollaLinks
Lightweight & fast AI inference proxy for self-hosted LLMs backends like Ollama, LM Studio and others. Designed for speed, simplicity and local-first deployments.
☆92Updated last week
Alternatives and similar repositories for olla
Users that are interested in olla are comparing it to the libraries listed below
Sorting:
- A web application that converts speech to speech 100% private☆76Updated 4 months ago
- A real-time shared memory layer for multi-agent LLM systems.☆48Updated 3 months ago
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆224Updated last month
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆46Updated last month
- A sleek web interface for Ollama, making local LLM management and usage simple. WebOllama provides an intuitive UI to manage Ollama model…☆57Updated last month
- OLLama IMage CAtegorizer☆69Updated 8 months ago
- Local AI voice assistant stack for Home Assistant (GPU-accelerated) with persistent memory, follow-up conversation, and Ollama model reco…☆206Updated 2 months ago
- Eternal is an experimental platform for machine learning models and workflows.☆67Updated 6 months ago
- ☆165Updated last month
- A platform to self-host AI on easy mode☆171Updated this week
- AI powered Chatbot with real time updates.☆66Updated 11 months ago
- The PyVisionAI Official Repo☆105Updated 2 months ago
- 🗣️ Real‑time, low‑latency voice, vision, and conversational‑memory AI assistant built on LiveKit and local LLMs ✨☆95Updated 3 months ago
- ☆18Updated 3 months ago
- Personal voice assistant, with voice interruption and Twilio support☆18Updated 7 months ago
- 🚀 FlexLLama - Lightweight self-hosted tool for running multiple llama.cpp server instances with OpenAI v1 API compatibility and multi-GP…☆34Updated this week
- A local front-end for open-weight LLMs with memory, RAG, TTS/STT, Elo ratings, and dynamic research tools. Built with React and FastAPI.☆38Updated last month
- A lightweight UI for chatting with Ollama models. Streaming responses, conversation history, and multi-model support.☆122Updated 6 months ago
- A MCP server allowing LLM agents to easily connect and retrieve data from any database☆98Updated 2 months ago
- Shared Memory Storage for Multi-Agent Systems☆125Updated 3 months ago
- Retrieval-augmented generation (RAG) for remote & local LLM use☆45Updated 4 months ago
- the IDE for research, built from the ground up with AI integrations☆115Updated last week
- ☆187Updated 6 months ago
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆92Updated 3 months ago
- fully local, temporally aware natural language file search on your pc! even without a GPU. find relevant files using natural language i…☆117Updated last week
- ☆178Updated 3 weeks ago
- ☆60Updated 3 months ago
- ☆20Updated last year
- Give your local LLM a real memory with a lightweight, fully local memory system. 100% offline and under your control.☆58Updated 2 weeks ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆82Updated this week