SearchSavior / OpenArcLinks
Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.
☆274Updated this week
Alternatives and similar repositories for OpenArc
Users that are interested in OpenArc are comparing it to the libraries listed below
Sorting:
- InferX: Inference as a Service Platform☆146Updated this week
- Aggregates compute from spare GPU capacity☆184Updated 2 weeks ago
- ☆88Updated last month
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆637Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆87Updated this week
- ☆83Updated 10 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆617Updated this week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆164Updated this week
- LLM Fine Tuning Toolbox images for Ryzen AI 395+ Strix Halo☆42Updated 4 months ago
- Sparse Inferencing for transformer based LLMs☆218Updated 5 months ago
- A tool to determine whether or not your PC can run a given LLM☆167Updated 11 months ago
- Easy to use interface for the Whisper model optimized for all GPUs!☆441Updated last week
- Manifold is an experimental platform for enabling long horizon workflow automation using teams of AI assistants.☆477Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,511Updated this week
- ☆178Updated 5 months ago
- ☆229Updated 8 months ago
- ☆204Updated 4 months ago
- ☆50Updated 3 months ago
- A web application that converts speech to speech 100% private☆82Updated 7 months ago
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆225Updated 5 months ago
- ☆210Updated 2 weeks ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆127Updated this week
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆279Updated 2 weeks ago
- GPU Power and Performance Manager☆65Updated last year
- Privacy-first agentic framework with powerful reasoning & task automation capabilities. Natively distributed and fully ISO 27XXX complian…☆68Updated 9 months ago
- ☆119Updated last week
- Open source LLM UI, compatible with all local LLM providers.☆177Updated last year
- The Fastest Way to Fine-Tune LLMs Locally☆333Updated last month
- Produce your own Dynamic 3.0 Quants and achieve optimum accuracy & SOTA quantization performance! Input your VRAM and RAM and the toolcha…☆76Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆21Updated last week