bdhirsh / pytorch_open_registration_exampleLinks
Example of using pytorch's open device registration API
☆30Updated 2 years ago
Alternatives and similar repositories for pytorch_open_registration_example
Users that are interested in pytorch_open_registration_example are comparing it to the libraries listed below
Sorting:
- An extension library of WMMA API (Tensor Core API)☆105Updated last year
- Benchmark code for the "Online normalizer calculation for softmax" paper☆99Updated 7 years ago
- Benchmark scripts for TVM☆74Updated 3 years ago
- System for automated integration of deep learning backends.☆47Updated 3 years ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆94Updated last week
- Ahead of Time (AOT) Triton Math Library☆76Updated 2 weeks ago
- ☆150Updated 8 months ago
- ☆98Updated last year
- ☆107Updated last year
- A home for the final text of all TVM RFCs.☆106Updated 11 months ago
- MatMul Performance Benchmarks for a Single CPU Core comparing both hand engineered and codegen kernels.☆134Updated last year
- An extention of TVMScript to write simple and high performance GPU kernels with tensorcore.☆51Updated last year
- Standalone Flash Attention v2 kernel without libtorch dependency☆110Updated last year
- An unofficial cuda assembler, for all generations of SASS, hopefully :)☆84Updated 2 years ago
- Assembler for NVIDIA Volta and Turing GPUs☆230Updated 3 years ago
- ☆116Updated 8 months ago
- ☆50Updated last year
- Performance of the C++ interface of flash attention and flash attention v2 in large language model (LLM) inference scenarios.☆40Updated 6 months ago
- ☆39Updated 5 years ago
- PET: Optimizing Tensor Programs with Partially Equivalent Transformations and Automated Corrections☆122Updated 3 years ago
- play gemm with tvm☆91Updated 2 years ago
- This is a demo how to write a high performance convolution run on apple silicon☆54Updated 3 years ago
- ☆139Updated 4 months ago
- A lightweight design for computation-communication overlap.☆167Updated last week
- Matrix Multiply-Accumulate with CUDA and WMMA( Tensor Core)☆143Updated 5 years ago
- ☆62Updated 9 months ago
- llama INT4 cuda inference with AWQ☆54Updated 8 months ago
- ☆44Updated this week
- ☆37Updated 2 months ago
- High Performance Grouped GEMM in PyTorch☆30Updated 3 years ago