teabranch / open-responses-serverLinks
Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm compliant.
☆67Updated this week
Alternatives and similar repositories for open-responses-server
Users that are interested in open-responses-server are comparing it to the libraries listed below
Sorting:
- InferX is a Inference Function as a Service Platform☆114Updated this week
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 9 months ago
- A meta-framework for self-improving LLMs with transparent reasoning☆19Updated this week
- An extension that lets the AI take the wheel, allowing it to use the mouse and keyboard, recognize UI elements, and prompt itself :3...no…☆124Updated 8 months ago
- Open source tool for transcirption and subtitling, alternative to happyscribe.☆26Updated 4 months ago
- A simple tool to anonymize LLM prompts.☆63Updated 5 months ago
- MockLLM, when you want it to do what you tell it to do!☆55Updated last week
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆26Updated 3 weeks ago
- A web application that converts speech to speech 100% private☆71Updated last month
- Retrieval-augmented generation (RAG) for remote & local LLM use☆45Updated last month
- Give your local LLM a real memory with a lightweight, fully local memory system — just like a human recalling past discussions. 100% off…☆49Updated 2 weeks ago
- ☆97Updated last month
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆31Updated 3 months ago
- AI web agent to find answers to any question☆33Updated last month
- Shared Memory Storage for Multi-Agent Systems☆81Updated this week
- Chat with OpenAI models from Claude Desktop☆54Updated 6 months ago
- ☆100Updated 2 weeks ago
- ☆24Updated 5 months ago
- ☆62Updated 6 months ago
- 🗣️ Real‑time, low‑latency voice, vision, and conversational‑memory AI assistant built on LiveKit and local LLMs ✨☆60Updated last week
- Local LLM Server with GPU and NPU Acceleration☆160Updated this week
- Smart proxy for LLM APIs that enables model-specific parameter control, automatic mode switching (like Qwen3's /think and /no_think), and…☆48Updated last month
- A simple experiment on letting two local LLM have a conversation about anything!☆110Updated last year
- ☆79Updated 4 months ago
- Super simple python connectors for llama.cpp, including vision models (Gemma 3, Qwen2-VL). Compile llama.cpp and run!☆25Updated last month
- ☆145Updated this week
- ☆128Updated last week
- CaSIL is an advanced natural language processing system that implements a sophisticated four-layer semantic analysis architecture. It pro…☆66Updated 7 months ago
- Call another MCP client from your MCP client. Offload context windows, delegate tasks, split between models☆26Updated 4 months ago
- Conduct in-depth research with AI-driven insights : DeepDive is a command-line tool that leverages web searches and AI models to generate…☆42Updated 10 months ago