lemonade-sdk / llamacpp-rocmLinks
Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration
☆31Updated this week
Alternatives and similar repositories for llamacpp-rocm
Users that are interested in llamacpp-rocm are comparing it to the libraries listed below
Sorting:
- Lightweight Inference server for OpenVINO☆198Updated this week
- AMD APU compatible Ollama. Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, Mistral Small 3.1 and other large language mod…☆80Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆129Updated last week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆209Updated 6 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆93Updated 2 months ago
- ☆253Updated 2 months ago
- ☆381Updated 4 months ago
- Make PyTorch models at least run on APUs.☆56Updated last year
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆22Updated 10 months ago
- ☆51Updated last week
- GPU Power and Performance Manager☆61Updated 10 months ago
- Add genai backend for ollama to run generative AI models using OpenVINO Runtime.☆13Updated 2 months ago
- Model swapping for llama.cpp (or any local OpenAPI compatible server)☆1,370Updated this week
- Input text from speech in any Linux window, the lean, fast and accurate way, using whisper.cpp OFFLINE. Speak with local LLMs via llama.c…☆126Updated last month
- Docs for GGUF quantization (unofficial)☆251Updated last month
- ☆63Updated 3 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆28Updated 10 months ago
- A platform to self-host AI on easy mode☆159Updated 2 weeks ago
- A open webui function for better R1 experience☆79Updated 5 months ago
- ☆129Updated this week
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆73Updated 3 months ago
- llama-swap + a minimal ollama compatible api☆23Updated this week
- ☆221Updated 3 months ago
- General Site for the GFX803 ROCm Stuff☆103Updated 3 weeks ago
- ☆461Updated this week
- Run LLM Agents on Ryzen AI PCs in Minutes☆518Updated last week
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆280Updated 2 weeks ago
- Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPU…☆1,098Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,096Updated this week
- reddacted lets you analyze & sanitize your online footprint using LLMs, PII detection & sentiment analysis to identify anything that migh…☆108Updated last month