ROCm / AITemplateLinks
AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.
☆12Updated last year
Alternatives and similar repositories for AITemplate
Users that are interested in AITemplate are comparing it to the libraries listed below
Sorting:
- a simple Flash Attention v2 implementation with ROCM (RDNA3 GPU, roc wmma), mainly used for stable diffusion(ComfyUI) in Windows ZLUDA en…☆47Updated last year
- AI Tensor Engine for ROCm☆276Updated this week
- Development repository for the Triton language and compiler☆130Updated last week
- Fast and memory-efficient exact attention☆188Updated last week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆101Updated this week
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆460Updated this week
- OpenAI Triton backend for Intel® GPUs☆207Updated this week
- ☆27Updated this week
- ☆231Updated last year
- ☆55Updated this week
- Model Compression Toolbox for Large Language Models and Diffusion Models☆628Updated last month
- ☆139Updated 4 months ago
- collection of benchmarks to measure basic GPU capabilities☆416Updated 7 months ago
- Ahead of Time (AOT) Triton Math Library☆76Updated 2 weeks ago
- FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.☆898Updated last year
- An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).☆265Updated 2 months ago
- SYCL based CUTLASS implementation for Intel GPUs☆39Updated this week
- QuickReduce is a performant all-reduce library designed for AMD ROCm that supports inline compression.☆33Updated 2 weeks ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆62Updated 2 months ago
- ☆43Updated last week
- Shared Middle-Layer for Triton Compilation☆286Updated 2 weeks ago
- AMD's graph optimization engine.☆249Updated this week
- Fastest kernels written from scratch☆343Updated 5 months ago
- monorepo for rocm libraries☆112Updated this week
- cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it☆614Updated 2 weeks ago
- BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.☆673Updated last month
- Fast low-bit matmul kernels in Triton☆365Updated this week
- [MLSys'25] QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving; [MLSys'25] LServe: Efficient Long-sequence LLM Se…☆753Updated 6 months ago
- A Easy-to-understand TensorOp Matmul Tutorial☆376Updated 11 months ago
- [DEPRECATED] Moved to ROCm/rocm-libraries repo☆111Updated this week