nktice / AMD-AI
AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1
☆177Updated last month
Alternatives and similar repositories for AMD-AI:
Users that are interested in AMD-AI are comparing it to the libraries listed below
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆478Updated last week
- 8-bit CUDA functions for PyTorch☆39Updated last month
- 8-bit CUDA functions for PyTorch, ported to HIP for use in AMD GPUs☆44Updated last year
- 8-bit CUDA functions for PyTorch Rocm compatible☆39Updated 8 months ago
- An OAI compatible exllamav2 API that's both lightweight and fast☆674Updated this week
- Fast and memory-efficient exact attention☆148Updated this week
- The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface. Now ZLUDA enhanced for better AMD GPU p…☆185Updated this week
- LLM Frontend in a single html file☆272Updated 2 weeks ago
- Simple monkeypatch to boost AMD Navi 3 GPUs☆24Updated 7 months ago
- Stable Diffusion Docker image preconfigured for usage with AMD Radeon cards☆128Updated 6 months ago
- Open source LLM UI, compatible with all local LLM providers.☆168Updated 2 months ago
- ☆205Updated last year
- CUDA on AMD GPUs☆318Updated 3 months ago
- Web UI for ExLlamaV2☆450Updated 2 months ago
- build scripts for ROCm☆181Updated 11 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆44Updated last month
- Dolphin System Messages☆216Updated 2 months ago
- A multimodal, function calling powered LLM webui.☆209Updated 2 months ago
- An OpenAI API compatible text to speech server using Coqui AI's xtts_v2 and/or piper tts as the backend.☆538Updated 3 months ago
- On-demand model switching with llama.cpp (or other OpenAI compatible backends)☆109Updated this week
- ☆145Updated this week
- 4 bits quantization of LLaMA using GPTQ, ported to HIP for use in AMD GPUs.☆32Updated last year
- ROCm docker images with fixes/support for extra architectures, such as gfx803/gfx1010.☆25Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆131Updated 7 months ago
- A fast batching API to serve LLM models☆175Updated 7 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆112Updated this week
- An AI assistant beyond the chat box.☆316Updated 9 months ago
- GPU Power and Performance Manager☆51Updated 2 months ago
- A manual for helping using tesla p40 gpu☆106Updated last month
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆141Updated 3 months ago