eugr / spark-vllm-dockerLinks
Docker configuration for running VLLM on dual DGX Sparks
☆227Updated this week
Alternatives and similar repositories for spark-vllm-docker
Users that are interested in spark-vllm-docker are comparing it to the libraries listed below
Sorting:
- A persistent local memory for AI, LLMs, or Copilot in VS Code.☆193Updated 3 months ago
- ☆207Updated 5 months ago
- InferX: Inference as a Service Platform☆156Updated this week
- Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.☆435Updated this week
- Docs for GGUF quantization (unofficial)☆366Updated 6 months ago
- ☆442Updated 2 months ago
- Plug-and-play memory for LLMs in 3 lines of code. Add persistent, intelligent, human-like memory and recall to any model in minutes.☆254Updated last week
- Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs. Join our discord: https…☆2,113Updated this week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆626Updated 2 weeks ago
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆2,374Updated this week
- An MCP server that executes Python code in isolated rootless containers with optional MCP server proxying. Implementation of Anthropic's …☆306Updated 2 months ago
- Enhancing LLMs with LoRA☆206Updated 3 months ago
- Manifold is an experimental platform for enabling long horizon workflow automation using teams of AI assistants.☆479Updated last week
- ☆178Updated 6 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,605Updated this week
- ☆911Updated last week
- Build AI applications that can see, hear, and speak using your screens, microphones, and cameras as inputs.☆1,078Updated last month
- We believe that every SOTA result is only valid on its own dataset. RAGView provides a unified evaluation platform to benchmark different…☆78Updated 2 months ago
- Multi-agent autonomous research system using LangGraph and LangChain. Generates citation-backed reports with credibility scoring and web …☆136Updated last month
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆282Updated last month
- ☆88Updated 3 weeks ago
- OpenAI-compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP too…☆314Updated this week
- ☆1,205Updated last week
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆595Updated this week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆149Updated 3 months ago
- ☆230Updated 9 months ago
- Official python implementation of UTCP. UTCP is an open standard that lets AI agents call any API directly, without extra middleware.☆636Updated 2 months ago
- Interactive launcher and benchmarking harness for llama.cpp server throughput, with tests, sweeps, and round‑robin load tools.☆211Updated last week
- CoexistAI is a modular, developer-friendly research assistant framework . It enables you to build, search, summarize, and automate resear…☆425Updated 3 months ago
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆227Updated 6 months ago