pomoke / torch-apu-helperLinks
Make PyTorch models at least run on APUs.
☆56Updated 2 years ago
Alternatives and similar repositories for torch-apu-helper
Users that are interested in torch-apu-helper are comparing it to the libraries listed below
Sorting:
- ☆63Updated 8 months ago
- ☆514Updated this week
- Deep Learning Primitives and Mini-Framework for OpenCL☆205Updated last year
- Fork of ollama for vulkan support☆108Updated 11 months ago
- Because RKNPU only knows 4D☆40Updated last year
- ☆236Updated 2 years ago
- DLPrimitives/OpenCL out of tree backend for pytorch☆384Updated last month
- ☆53Updated last year
- ☆49Updated 2 years ago
- ☆422Updated 9 months ago
- My develoopment fork of llama.cpp. For now working on RK3588 NPU and Tenstorrent backend☆114Updated 2 months ago
- Reverse engineering the rk3588 npu☆108Updated last year
- Run Large Language Models on RK3588 with GPU-acceleration☆121Updated 2 years ago
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆107Updated 8 months ago
- build scripts for ROCm☆188Updated 2 years ago
- ☆180Updated last month
- Download models from the Ollama library, without Ollama☆119Updated last year
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆164Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated last month
- Run stable-diffusion-webui with Radeon RX 580 8GB on Ubuntu 22.04.2 LTS☆67Updated 2 years ago
- ☆66Updated last year
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆274Updated this week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆724Updated this week
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆735Updated last month
- ☆119Updated last week
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆52Updated 10 months ago
- FORK of VLLM for AMD MI25/50/60. A high-throughput and memory-efficient inference and serving engine for LLMs☆65Updated 8 months ago
- Efficient Inference of Transformer models☆478Updated last year
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆82Updated 7 months ago
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆134Updated 2 years ago