mixa3607 / ML-gfx906Links
ML software (llama.cpp, ComfyUI, vLLM) builds for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60
☆103Updated 2 months ago
Alternatives and similar repositories for ML-gfx906
Users that are interested in ML-gfx906 are comparing it to the libraries listed below
Sorting:
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆365Updated last month
- llama.cpp-gfx906☆85Updated 2 weeks ago
- ☆426Updated 9 months ago
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆742Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,553Updated this week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆283Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆665Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆167Updated last week
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆751Updated last week
- FORK of VLLM for AMD MI25/50/60. A high-throughput and memory-efficient inference and serving engine for LLMs☆65Updated 8 months ago
- Triton for AMD MI25/50/60. Development repository for the Triton language and compiler☆32Updated last month
- ☆857Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆217Updated 2 months ago
- triton for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆40Updated last month
- Build AI agents for your PC☆894Updated last week
- AI Inferencing at the Edge. A simple one-file way to run various GGML models with KoboldAI's UI with AMD ROCm offloading☆735Updated last month
- ☆192Updated 3 months ago
- ☆237Updated 2 years ago
- build scripts for ROCm☆186Updated 2 years ago
- ROCm Library Files for gfx1103 and update with others arches based on AMD GPUs for use in Windows.☆737Updated 4 months ago
- ☆520Updated this week
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆108Updated 9 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆622Updated this week
- The main repository for building Pascal-compatible versions of ML applications and libraries.☆163Updated 5 months ago
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆720Updated last week
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆2,260Updated last week
- LM inference server implementation based on *.cpp.☆294Updated 2 months ago
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆136Updated 2 years ago
- 8-bit CUDA functions for PyTorch☆70Updated 4 months ago
- GPU Power and Performance Manager☆66Updated last year