nktice / AMD-AILinks
AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1
☆204Updated 3 months ago
Alternatives and similar repositories for AMD-AI
Users that are interested in AMD-AI are comparing it to the libraries listed below
Sorting:
- ☆326Updated 2 months ago
- Stable Diffusion Docker image preconfigured for usage with AMD Radeon cards☆133Updated last year
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆624Updated this week
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆969Updated this week
- ☆226Updated 2 years ago
- Web UI for ExLlamaV2☆495Updated 4 months ago
- Lightweight Inference server for OpenVINO☆176Updated last week
- build scripts for ROCm☆186Updated last year
- A zero dependency web UI for any LLM backend, including KoboldCpp, OpenAI and AI Horde☆121Updated this week
- A manual for helping using tesla p40 gpu☆126Updated 6 months ago
- Croco.Cpp is a 3rd party testground for KoboldCPP, a simple one-file way to run various GGML/GGUF models with KoboldAI's UI. (for Croco.C…☆107Updated this week
- Simple monkeypatch to boost AMD Navi 3 GPUs☆42Updated last month
- 8-bit CUDA functions for PyTorch, ported to HIP for use in AMD GPUs☆49Updated 2 years ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- ☆71Updated last week
- ☆525Updated 2 weeks ago
- ROCm docker images with fixes/support for extra architectures, such as gfx803/gfx1010.☆30Updated last year
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆86Updated last month
- Make PyTorch models at least run on APUs.☆55Updated last year
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆385Updated this week
- Run stable-diffusion-webui with Radeon RX 580 8GB on Ubuntu 22.04.2 LTS☆63Updated last year
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆42Updated 3 weeks ago
- Prebuilt Windows ROCm Libs for gfx1031 and gfx1032☆138Updated 2 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆519Updated this week
- Writing Extension for Text Generation WebUI☆55Updated 4 months ago
- CUDA on AMD GPUs☆502Updated 3 weeks ago
- GPU Power and Performance Manager☆58Updated 7 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆153Updated last year
- Dolphin System Messages☆311Updated 3 months ago
- Fast and memory-efficient exact attention☆172Updated this week