Hardware-Alchemy / cuDNN-sampleLinks
cuDNN sample codes provided by Nvidia
☆46Updated 6 years ago
Alternatives and similar repositories for cuDNN-sample
Users that are interested in cuDNN-sample are comparing it to the libraries listed below
Sorting:
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆146Updated 5 years ago
- CUDA Matrix Multiplication Optimization☆247Updated last year
- Assembler for NVIDIA Volta and Turing GPUs☆236Updated 3 years ago
- Fast CUDA Kernels for ResNet Inference.☆182Updated 6 years ago
- Efficient Top-K implementation on the GPU☆192Updated 6 years ago
- Training material for Nsight developer tools☆173Updated last year
- A tool for examining GPU scheduling behavior.☆89Updated last year
- ☆41Updated 3 years ago
- ☆110Updated last year
- Instructions, Docker images, and examples for Nsight Compute and Nsight Systems☆134Updated 5 years ago
- A library of GPU kernels for sparse matrix operations.☆280Updated 5 years ago
- THIS REPOSITORY HAS MOVED TO github.com/nvidia/cub, WHICH IS AUTOMATICALLY MIRRORED HERE.☆85Updated last year
- MatMul Performance Benchmarks for a Single CPU Core comparing both hand engineered and codegen kernels.☆138Updated 2 years ago
- Dissecting NVIDIA GPU Architecture☆115Updated 3 years ago
- heterogeneity-aware-lowering-and-optimization☆257Updated last year
- ☆117Updated last year
- A Winograd Minimal Filter Implementation in CUDA☆28Updated 4 years ago
- An extension library of WMMA API (Tensor Core API)☆109Updated last year
- [DEPRECATED] Moved to ROCm/rocm-libraries repo☆253Updated 2 weeks ago
- ☆480Updated 10 years ago
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆70Updated last year
- tophub autotvm log collections☆69Updated 3 years ago
- Benchmark code for the "Online normalizer calculation for softmax" paper☆103Updated 7 years ago
- Automatic Schedule Exploration and Optimization Framework for Tensor Computations☆181Updated 3 years ago
- Several optimization methods of half-precision general matrix multiplication (HGEMM) using tensor core with WMMA API and MMA PTX instruct…☆511Updated last year
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆500Updated this week
- A simple high performance CUDA GEMM implementation.☆421Updated last year
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆200Updated 3 years ago
- A home for the final text of all TVM RFCs.☆108Updated last year
- Optimizing SGEMM kernel functions on NVIDIA GPUs to a close-to-cuBLAS performance.☆398Updated 11 months ago