zhaohb / ollama_ovLinks
Add genai backend for ollama to run generative AI models using OpenVINO Runtime.
☆18Updated 5 months ago
Alternatives and similar repositories for ollama_ov
Users that are interested in ollama_ov are comparing it to the libraries listed below
Sorting:
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated 2 weeks ago
- ☆257Updated 5 months ago
- Make use of Intel Arc Series GPU to Run Ollama, StableDiffusion, Whisper and Open WebUI, for image generation, speech recognition and int…☆184Updated 5 months ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆103Updated this week
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆692Updated 2 weeks ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆241Updated last week
- Ollama with intel (i)GPU acceleration in docker and benchmark☆27Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆451Updated this week
- This is a one-click install script to enable Gen 12/13 Intel vGPU using SR-IOV Technology so up to 7 Client VMs can enjoy hardware GPU de…☆75Updated last year
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆327Updated last month
- See how to play with ROCm, run it with AMD GPUs!☆36Updated 6 months ago
- The main repository for building Pascal-compatible versions of ML applications and libraries.☆147Updated 2 months ago
- Make PyTorch models at least run on APUs.☆57Updated last year
- ☆52Updated 2 years ago
- llama-swap + a minimal ollama compatible api☆33Updated 2 weeks ago
- Everything you need to setup on your AMD system for Machine Learning Stuff☆19Updated 3 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆97Updated 2 weeks ago
- A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Cons…☆40Updated 3 weeks ago
- Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, DeepSeek, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V,…☆66Updated 6 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆31Updated last year
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆656Updated this week
- ROCm docker images with fixes/support for extra architectures, such as gfx803/gfx1010.☆31Updated 2 years ago
- Faster whisper Running on AMD GPUs with modified CTranslate 2 Libraries served up with Wyoming protocol☆29Updated last year
- Automated script to convert Huggingface and GGUF models to rkllm format for running on Rockchip NPU☆37Updated last year
- Open WebUI Client for Android is a mobile app for using Open WebUI interfaces with local or remote AI models.☆121Updated 3 months ago
- Wyoming protocol server for whisper speech to text system on AMD GPUs☆12Updated last year
- Fast inference engine for Transformer models☆54Updated last year
- ZFS Snapshot Browser Based GUI☆46Updated last year
- Run LLM Agents on Ryzen AI PCs in Minutes☆744Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,329Updated this week