amd / RyzenAI-SW
☆495Updated this week
Alternatives and similar repositories for RyzenAI-SW:
Users that are interested in RyzenAI-SW are comparing it to the libraries listed below
- ☆399Updated this week
- Intel® NPU (Neural Processing Unit) Driver☆237Updated this week
- Intel® NPU Acceleration Library☆658Updated 2 months ago
- Fork of LLVM to support AMD AIEngine processors☆129Updated this week
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆54Updated this week
- A collection of examples for the ROCm software stack☆198Updated this week
- AI Tensor Engine for ROCm☆142Updated this week
- ☆104Updated 3 weeks ago
- ☆278Updated this week
- LLM SDK for OnnxRuntime GenAI (OGA)☆119Updated this week
- HIPIFY: Convert CUDA to Portable C++ Code☆567Updated last week
- AMD related optimizations for transformer models☆72Updated 5 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆200Updated last month
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆373Updated this week
- OpenAI Triton backend for Intel® GPUs☆172Updated this week
- chipStar is a tool for compiling and running HIP/CUDA on SPIR-V via OpenCL or Level Zero APIs.☆260Updated last week
- ☆122Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆249Updated this week
- An MLIR-based toolchain for AMD AI Engine-enabled devices.☆354Updated this week
- Dockerfiles for the various software layers defined in the ROCm software platform☆455Updated last month
- BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.☆572Updated last month
- build scripts for ROCm☆189Updated last year
- ☆60Updated last year
- AMD's graph optimization engine.☆213Updated this week
- Next generation BLAS implementation for ROCm platform☆361Updated this week
- Generative AI extensions for onnxruntime☆667Updated this week
- Fast and memory-efficient exact attention☆163Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆351Updated 7 months ago
- Low-bit LLM inference on CPU with lookup table☆705Updated 2 months ago
- cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it☆534Updated 2 weeks ago