SearchSavior / OpenArcLinks
Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.
☆211Updated this week
Alternatives and similar repositories for OpenArc
Users that are interested in OpenArc are comparing it to the libraries listed below
Sorting:
- A platform to self-host AI on easy mode☆171Updated this week
- InferX: Inference as a Service Platform☆136Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆82Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆280Updated this week
- ☆83Updated this week
- No-code CLI designed for accelerating ONNX workflows☆214Updated 3 months ago
- Lightweight & fast AI inference proxy for self-hosted LLMs backends like Ollama, LM Studio and others. Designed for speed, simplicity and…☆92Updated last week
- A web application that converts speech to speech 100% private☆76Updated 4 months ago
- Enhancing LLMs with LoRA☆159Updated 3 weeks ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆279Updated last month
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆254Updated 7 months ago
- ☆207Updated last month
- Sparse Inferencing for transformer based LLMs☆201Updated last month
- ☆83Updated 7 months ago
- ☆178Updated last month
- ☆165Updated last month
- fully local, temporally aware natural language file search on your pc! even without a GPU. find relevant files using natural language i…☆120Updated last week
- Open source LLM UI, compatible with all local LLM providers.☆175Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆1,220Updated last week
- reddacted lets you analyze & sanitize your online footprint using LLMs, PII detection & sentiment analysis to identify anything that migh…☆108Updated 2 months ago
- The Fastest Way to Fine-Tune LLMs Locally☆321Updated 6 months ago
- Docs for GGUF quantization (unofficial)☆275Updated 2 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆21Updated this week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆110Updated 3 months ago
- GPU Power and Performance Manager☆61Updated 11 months ago
- ☆225Updated 5 months ago
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆46Updated last month
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆20Updated this week
- A open webui function for better R1 experience☆78Updated 7 months ago
- interactive semantic search demo using Qwen3-0.6B-Embedding in your browser☆52Updated 3 months ago