SearchSavior / OpenArcLinks
Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.
☆226Updated this week
Alternatives and similar repositories for OpenArc
Users that are interested in OpenArc are comparing it to the libraries listed below
Sorting:
- A platform to self-host AI on easy mode☆171Updated this week
- InferX: Inference as a Service Platform☆137Updated this week
- ☆84Updated 3 weeks ago
- Enhancing LLMs with LoRA☆172Updated last week
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆276Updated 2 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated last week
- A web application that converts speech to speech 100% private☆76Updated 4 months ago
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆378Updated this week
- ☆180Updated last month
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆224Updated 2 months ago
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- No-code CLI designed for accelerating ONNX workflows☆215Updated 4 months ago
- The Fastest Way to Fine-Tune LLMs Locally☆323Updated 7 months ago
- ☆168Updated 2 months ago
- Docs for GGUF quantization (unofficial)☆293Updated 3 months ago
- ☆83Updated 8 months ago
- Sparse Inferencing for transformer based LLMs☆201Updated 2 months ago
- ☆206Updated last month
- Manifold is a platform for enabling workflow automation using AI assistants.☆464Updated this week
- fully local, temporally aware natural language file search on your pc! even without a GPU. find relevant files using natural language i…☆129Updated last month
- ☆226Updated 5 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆21Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆79Updated this week
- GPU Power and Performance Manager☆60Updated last year
- Code for Papeg.ai☆225Updated 9 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆541Updated last week
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆255Updated last week
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆47Updated 2 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,266Updated last week
- Easy to use interface for the Whisper model optimized for all GPUs!☆376Updated 2 months ago