SearchSavior / OpenArcLinks
Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.
☆260Updated last week
Alternatives and similar repositories for OpenArc
Users that are interested in OpenArc are comparing it to the libraries listed below
Sorting:
- A platform to self-host AI on easy mode☆178Updated this week
- InferX: Inference as a Service Platform☆142Updated this week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆488Updated last week
- ☆87Updated 2 weeks ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆84Updated last week
- A web application that converts speech to speech 100% private☆81Updated 6 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆278Updated 3 months ago
- ☆176Updated 3 months ago
- ☆50Updated last month
- ☆83Updated 9 months ago
- ☆228Updated 7 months ago
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- ☆195Updated 3 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆588Updated this week
- Enhancing LLMs with LoRA☆191Updated last month
- Easy to use interface for the Whisper model optimized for all GPUs!☆399Updated 4 months ago
- Sparse Inferencing for transformer based LLMs☆215Updated 3 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,358Updated this week
- reddacted lets you analyze & sanitize your online footprint using LLMs, PII detection & sentiment analysis to identify anything that migh…☆113Updated 4 months ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆129Updated this week
- ☆209Updated 3 months ago
- Docs for GGUF quantization (unofficial)☆330Updated 4 months ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆119Updated 2 weeks ago
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆226Updated 3 months ago
- No-code CLI designed for accelerating ONNX workflows☆219Updated 5 months ago
- ☆108Updated 3 months ago
- Cohere Toolkit is a collection of prebuilt components enabling users to quickly build and deploy RAG applications.☆29Updated 10 months ago
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆236Updated last month
- The Fastest Way to Fine-Tune LLMs Locally☆328Updated 8 months ago
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆257Updated last month