set-soft / sd_webui_rx5500Links
Docker images for Stable Diffusion WebUI (AUTOMATIC1111) for AMD Radeon RX5500XT and similar boards
☆13Updated last year
Alternatives and similar repositories for sd_webui_rx5500
Users that are interested in sd_webui_rx5500 are comparing it to the libraries listed below
Sorting:
- Produce your own Dynamic 3.0 Quants and achieve optimum accuracy & SOTA quantization performance! Input your VRAM and RAM and the toolcha…☆76Updated last week
- Run stable-diffusion-webui with Radeon RX 580 8GB on Ubuntu 22.04.2 LTS☆68Updated 2 years ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆88Updated last week
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆21Updated 5 months ago
- ☆49Updated 2 years ago
- ☆90Updated 2 months ago
- A simple GUI utility for gathering LIMA-like chat data.☆23Updated 4 months ago
- tinygrad port of the RWKV large language model.☆45Updated 11 months ago
- Lower Precision Floating Point Operations☆66Updated last month
- Port of Microsoft's BioGPT in C/C++ using ggml☆85Updated last year
- MiniLM (BERT) embeddings from scratch☆17Updated 5 months ago
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆51Updated 11 months ago
- Generate a llama-quantize command to copy the quantization parameters of any GGUF☆30Updated 2 weeks ago
- Make abliterated models with transformers, easy and fast☆116Updated 2 weeks ago
- C/C++ implementation of PygmalionAI/pygmalion-6b☆55Updated 2 years ago
- Single-file, pure CUDA C implementation for running inference on Qwen3 0.6B GGUF. No Dependencies.☆22Updated 2 months ago
- fast state-of-the-art speech models and a runtime that runs anywhere 💥☆57Updated 7 months ago
- ☆238Updated 2 years ago
- Experimental LLM Inference UX to aid in creative writing☆128Updated last year
- Loader extension for tabbyAPI in SillyTavern☆26Updated 7 months ago
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆138Updated 2 years ago
- No-messing-around sh client for llama.cpp's server☆30Updated last year
- A c++ framework on efficient training & fine-tuning LLMs☆27Updated last week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆626Updated 2 weeks ago
- A shared memory message based broadcast system using atomic interlocking.☆40Updated 4 months ago
- An unsupervised model merging algorithm for Transformers-based language models.☆108Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA☆125Updated 2 years ago
- A simple, "Ollama-like" tool for managing and running GGUF language models from your terminal.☆23Updated last month
- My personal fork of koboldcpp where I hack in experimental samplers.☆44Updated last year
- Stable Diffusion and Flux in pure C/C++☆24Updated last week