Simple example of how to write an Implicit GEMM Convolution in CUDA using the tensor core WMMA API and bindings for PyTorch.
☆18Jun 29, 2023Updated 2 years ago
Alternatives and similar repositories for implicit-gemm-tensor-core-convolution
Users that are interested in implicit-gemm-tensor-core-convolution are comparing it to the libraries listed below
Sorting:
- ☆14May 28, 2019Updated 6 years ago
- a tiny distro-independent package manager written in Rust.☆16Jun 22, 2024Updated last year
- ☆32Aug 24, 2022Updated 3 years ago
- Artifact for PPoPP22 QGTC: Accelerating Quantized GNN via GPU Tensor Core.☆30Feb 12, 2022Updated 4 years ago
- ☆17Oct 28, 2025Updated 4 months ago
- General Matrix Multiplication using NVIDIA Tensor Cores☆28Jan 25, 2025Updated last year
- ☆11Apr 16, 2023Updated 2 years ago
- A toy C Compiler implemented by Rust.☆19Feb 4, 2023Updated 3 years ago
- An extension library of WMMA API (Tensor Core API)☆111Jul 12, 2024Updated last year
- LLVM alternative in Rust☆15May 20, 2024Updated last year
- Instruction Pointer Classifier and Dynamic Degree Stream based Hardware Cache Prefetching☆16Nov 16, 2019Updated 6 years ago
- Source code of the PPoPP '22 paper: "TileSpGEMM: A Tiled Algorithm for Parallel Sparse General Matrix-Matrix Multiplication on GPUs" by Y…☆46May 22, 2024Updated last year
- ☆12Jan 19, 2020Updated 6 years ago
- ☆13Sep 30, 2023Updated 2 years ago
- ☆13Jan 18, 2020Updated 6 years ago
- A patch of clang format for #pragma with no patch to clang-format☆19Dec 27, 2021Updated 4 years ago
- ☆18Oct 3, 2022Updated 3 years ago
- [ISCA'25] LIA: A Single-GPU LLM Inference Acceleration with Cooperative AMX-Enabled CPU-GPU Computation and CXL Offloading☆24Jan 6, 2026Updated 2 months ago
- Fast GPU based tensor core reductions☆13Jan 13, 2023Updated 3 years ago
- CUDA 8-bit Tensor Core Matrix Multiplication based on m16n16k16 WMMA API☆34Sep 15, 2023Updated 2 years ago
- Source code of the paper "OpSparse: a Highly Optimized Framework for Sparse General Matrix Multiplication on GPUs"☆16Aug 23, 2022Updated 3 years ago
- A wide array of parallel programs using CUDA, OpenCL, MPI, OpenMP and pthreads.☆14Jan 6, 2015Updated 11 years ago
- Massively Scalable Parallel GMRES C-code for Sparse System of Equations☆13Feb 16, 2016Updated 10 years ago
- ☆21Aug 21, 2023Updated 2 years ago
- A new QR decomposition algorithm implemented in CUDA☆18Jun 24, 2024Updated last year
- ☆11Dec 13, 2014Updated 11 years ago
- A Easy-to-understand TensorOp Matmul Tutorial☆409Mar 5, 2026Updated 2 weeks ago
- [Mlsys'22] Understanding gnn computational graph: A coordinated computation, io, and memory perspective☆22Sep 11, 2023Updated 2 years ago
- NVIDIA Compute Unified Device Architecture Toolkit☆15Feb 2, 2026Updated last month
- Custom error type of nom to improve accuracy of error position☆11Mar 23, 2023Updated 2 years ago
- Implementing a parallelized conjugate gradient algorithm using a hybrid of distributed (MPI) and shared (OpenMP) memory approach.☆11Dec 8, 2018Updated 7 years ago
- a fast and customizable CUDA int4 tensor core gemm☆15Aug 2, 2024Updated last year
- High performance RMSNorm Implement by using SM Core Storage(Registers and Shared Memory)☆29Jan 22, 2026Updated last month
- ☆16Nov 22, 2022Updated 3 years ago
- Article: GPU-accelerated Proximity Graph Approximate Nearest Neighbor Search and Construction by Authors Yuanhang Yu, Dong Wen, Ying Zhan…☆24Jun 20, 2025Updated 9 months ago
- Library for exact linear algebra, a C++ template-library based originally on LinBox intended for F4-like implementations☆18Dec 15, 2012Updated 13 years ago
- ☆24Apr 4, 2024Updated last year
- A highly efficient library for GEMM operations on Sunway TaihuLight☆18Sep 7, 2020Updated 5 years ago
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆92Nov 23, 2022Updated 3 years ago