lemonade-sdk / lemonadeLinks
Local LLM Server with GPU and NPU Acceleration
☆206Updated this week
Alternatives and similar repositories for lemonade
Users that are interested in lemonade are comparing it to the libraries listed below
Sorting:
- Lightweight Inference server for OpenVINO☆188Updated this week
- InferX is a Inference Function as a Service Platform☆116Updated 2 weeks ago
- Minimal Linux OS with a Model Context Protocol (MCP) gateway to expose local capabilities to LLMs.☆257Updated 3 weeks ago
- Turns devices into a scalable LLM platform☆150Updated last week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆72Updated 2 weeks ago
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆255Updated 2 weeks ago
- Run LLM Agents on Ryzen AI PCs in Minutes☆454Updated 2 weeks ago
- ☆267Updated this week
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆221Updated last month
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆185Updated this week
- ☆204Updated last month
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆67Updated 2 weeks ago
- A web application that converts speech to speech 100% private☆72Updated last month
- Fully Open Language Models with Stellar Performance☆234Updated last month
- MockLLM, when you want it to do what you tell it to do!☆56Updated this week
- Phi4 Multimodal Instruct - OpenAI endpoint and Docker Image for self-hosting☆37Updated 4 months ago
- ☆146Updated last week
- A simple tool to anonymize LLM prompts.☆63Updated 5 months ago
- Official python implementation of the UTCP☆184Updated this week
- Download models from the Ollama library, without Ollama☆89Updated 8 months ago
- Run and manage MCP servers as Docker containers with a unified HTTP endpoint. Inspired by Docker compose.☆39Updated this week
- ☆28Updated last month
- ☆95Updated 6 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆214Updated last year
- Neo AI integrates into the Linux terminal, capable of executing system commands and providing helpful information.☆112Updated 2 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆89Updated last week
- The Fastest Way to Fine-Tune LLMs Locally☆312Updated 3 months ago
- A MCP server allowing LLM agents to easily connect and retrieve data from any database☆95Updated last month
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 10 months ago
- Dia-JAX: A JAX port of Dia, the text-to-speech model for generating realistic dialogue from text with emotion and tone control.☆27Updated 2 months ago