pomoke / torch-apu-helperLinks
Make PyTorch models at least run on APUs.
☆56Updated 2 years ago
Alternatives and similar repositories for torch-apu-helper
Users that are interested in torch-apu-helper are comparing it to the libraries listed below
Sorting:
- ☆63Updated 7 months ago
- ☆502Updated this week
- build scripts for ROCm☆188Updated last year
- Reverse engineering the rk3588 npu☆104Updated last year
- ☆420Updated 8 months ago
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆107Updated 8 months ago
- ☆236Updated 2 years ago
- No-code CLI designed for accelerating ONNX workflows☆222Updated 6 months ago
- FORK of VLLM for AMD MI25/50/60. A high-throughput and memory-efficient inference and serving engine for LLMs☆65Updated 7 months ago
- Because RKNPU only knows 4D☆39Updated last year
- ☆53Updated last year
- My develoopment fork of llama.cpp. For now working on RK3588 NPU and Tenstorrent backend☆112Updated last month
- 8-bit CUDA functions for PyTorch☆69Updated 3 months ago
- ☆18Updated 11 months ago
- ☆65Updated last year
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆216Updated last month
- ☆48Updated 2 years ago
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆80Updated 7 months ago
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆88Updated 2 weeks ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆266Updated last week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆560Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆149Updated this week
- ☆111Updated this week
- Download models from the Ollama library, without Ollama☆118Updated last year
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆132Updated 2 years ago
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆71Updated this week
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆52Updated 10 months ago
- Run stable-diffusion-webui with Radeon RX 580 8GB on Ubuntu 22.04.2 LTS☆67Updated 2 years ago
- GPU Power and Performance Manager☆64Updated last year
- Kernel/Qemu Patches for Venus.☆35Updated last year