SearchSavior / OpenArcLinks
Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.
☆238Updated last week
Alternatives and similar repositories for OpenArc
Users that are interested in OpenArc are comparing it to the libraries listed below
Sorting:
- A platform to self-host AI on easy mode☆176Updated this week
- InferX: Inference as a Service Platform☆138Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆427Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆89Updated last week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆83Updated 2 weeks ago
- ☆85Updated last week
- ☆83Updated 8 months ago
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- Enhancing LLMs with LoRA☆174Updated 3 weeks ago
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆236Updated last month
- ☆207Updated 2 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆277Updated 2 months ago
- ☆173Updated 3 months ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆112Updated last week
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆225Updated 3 months ago
- ☆192Updated 2 months ago
- A tool to determine whether or not your PC can run a given LLM☆164Updated 9 months ago
- Code for Papeg.ai☆226Updated 10 months ago
- Privacy-first agentic framework with powerful reasoning & task automation capabilities. Natively distributed and fully ISO 27XXX complian…☆67Updated 7 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆97Updated 2 weeks ago
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆48Updated 2 months ago
- GPU Power and Performance Manager☆61Updated last year
- A local front-end for open-weight LLMs with memory, RAG, TTS/STT, Elo ratings, and dynamic research tools. Built with React and FastAPI.☆38Updated 3 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆571Updated this week
- ☆106Updated 2 months ago
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆255Updated last month
- ☆49Updated last month
- A web application that converts speech to speech 100% private☆80Updated 5 months ago
- The Fastest Way to Fine-Tune LLMs Locally☆325Updated 7 months ago
- ☆226Updated 6 months ago