lemonade-sdk / llamacpp-rocmLinks
Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration
☆164Updated this week
Alternatives and similar repositories for llamacpp-rocm
Users that are interested in llamacpp-rocm are comparing it to the libraries listed below
Sorting:
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆637Updated this week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆274Updated this week
- ☆186Updated 2 months ago
- ☆770Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆21Updated last week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated last month
- Build AI agents for your PC☆885Updated this week
- Produce your own Dynamic 3.0 Quants and achieve optimum accuracy & SOTA quantization performance! Input your VRAM and RAM and the toolcha…☆76Updated this week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆617Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,511Updated this week
- ☆50Updated 3 months ago
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆724Updated this week
- LLM Fine Tuning Toolbox images for Ryzen AI 395+ Strix Halo☆42Updated 4 months ago
- ☆119Updated last week
- ML software (llama.cpp, ComfyUI, vLLM) builds for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆98Updated 2 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆107Updated 2 months ago
- ☆102Updated 3 weeks ago
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆362Updated 3 weeks ago
- ☆257Updated 7 months ago
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆50Updated 4 months ago
- GPU Power and Performance Manager☆65Updated last year
- ☆88Updated last month
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆2,176Updated last week
- Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs. Join our discord: https…☆2,008Updated this week
- ☆204Updated 4 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆32Updated last year
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆127Updated this week
- llama.cpp-gfx906☆84Updated this week
- Docs for GGUF quantization (unofficial)☆348Updated 6 months ago
- Download models from the Ollama library, without Ollama☆119Updated last year