itsmostafa / inference-speed-testsLinks
Local LLM inference speed tests on various devices
☆116Updated last month
Alternatives and similar repositories for inference-speed-tests
Users that are interested in inference-speed-tests are comparing it to the libraries listed below
Sorting:
- Optimized Ollama LLM server configuration for Mac Studio and other Apple Silicon Macs. Headless setup with automatic startup, resource op…☆281Updated 2 weeks ago
- Your gateway to both Ollama & Apple MlX models☆150Updated 11 months ago
- A wannabe Ollama equivalent for Apple MlX models☆83Updated 11 months ago
- High-performance MLX-based LLM inference engine for macOS with native Swift implementation☆478Updated 3 weeks ago
- Accessing Apple Intelligence and ChatGPT desktop through OpenAI / Ollama API☆339Updated 5 months ago
- Local Apple Notes + LLM Chat☆98Updated last month
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆661Updated last month
- An MCP server that securely interfaces with your iMessage database via the Model Context Protocol (MCP), allowing LLMs to query and analy…☆238Updated 3 weeks ago
- MCP server that execute applescript giving you full control of your Mac☆414Updated 2 months ago
- CoexistAI is a modular, developer-friendly research assistant framework . It enables you to build, search, summarize, and automate resear…☆425Updated 3 months ago
- A native desktop app for Open WebUI built with Tauri v2☆82Updated this week
- macOS whisper dictation app☆610Updated this week
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆227Updated 6 months ago
- ☆258Updated 9 months ago
- 'afm' command cli: macOS server and single prompt mode that exposes Apple's Foundation Models and other APIs running on your Mac through …☆131Updated last week
- Local image and music generation for Apple Silicon☆72Updated 10 months ago
- MacOS menu‑bar utility to adjust Apple Silicon GPU VRAM allocation☆248Updated 9 months ago
- A macOS AppleScript MCP server☆356Updated 9 months ago
- ☆296Updated last year
- A multi-agent AI architecture that connects 25+ specialized agents through n8n and MCP servers. Project NOVA routes requests to domain-sp…☆255Updated 8 months ago
- AI agent that controls computer with OS-level tools, MCP compatible, works with any model☆132Updated 5 months ago
- MLX-GUI MLX Inference Server for Apple Silicone☆184Updated 3 weeks ago
- Ollama desktop client for everyday use☆89Updated 8 months ago
- OpenAI-compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP too…☆314Updated this week
- Local coding agent with neat UI☆344Updated 8 months ago
- a magical LLM desktop client that makes it easy for *anyone* to use LLMs and MCP☆587Updated 3 months ago
- MCP server for enabling LLM applications to perform deep research via the MCP protocol☆313Updated 3 months ago
- Link you Ollama models to LM-Studio☆150Updated last year
- ☆230Updated 9 months ago
- Tool for scraping and consolidating documentation websites into a single MD file.☆255Updated last month