mixa3607 / ML-gfx906Links
ML software (llama.cpp, ComfyUI, vLLM) builds for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60
☆34Updated last week
Alternatives and similar repositories for ML-gfx906
Users that are interested in ML-gfx906 are comparing it to the libraries listed below
Sorting:
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆307Updated 3 weeks ago
- llama.cpp-gfx906☆45Updated last month
- FORK of VLLM for AMD MI25/50/60. A high-throughput and memory-efficient inference and serving engine for LLMs☆64Updated 5 months ago
- ☆409Updated 6 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,277Updated this week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆514Updated this week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆226Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆212Updated this week
- Triton for AMD MI25/50/60. Development repository for the Triton language and compiler☆32Updated last month
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆378Updated this week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆79Updated this week
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆702Updated 2 weeks ago
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆677Updated last week
- ☆234Updated 2 years ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆29Updated last year
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated last year
- GPU Power and Performance Manager☆60Updated last year
- No-code CLI designed for accelerating ONNX workflows☆215Updated 4 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆541Updated 2 weeks ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆96Updated last month
- General Site for the GFX803 ROCm Stuff☆120Updated 2 months ago
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆640Updated last week
- Privacy-first agentic framework with powerful reasoning & task automation capabilities. Natively distributed and fully ISO 27XXX complian…☆66Updated 6 months ago
- ROCm Library Files for gfx1103 and update with others arches based on AMD GPUs for use in Windows.☆660Updated last month
- Stable Diffusion and Flux in pure C/C++☆21Updated this week
- triton for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆32Updated last month
- Run LLM Agents on Ryzen AI PCs in Minutes☆684Updated last week
- LM inference server implementation based on *.cpp.☆286Updated 2 months ago
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆62Updated last week
- InferX: Inference as a Service Platform☆137Updated this week