MegEngine / cutlass
CUDA Templates for Linear Algebra Subroutines
☆96Updated 10 months ago
Alternatives and similar repositories for cutlass:
Users that are interested in cutlass are comparing it to the libraries listed below
- MegEngine到其他框架的转换器☆69Updated last year
- ☆109Updated 11 months ago
- A set of examples around MegEngine☆31Updated last year
- ☆95Updated 3 years ago
- ☆82Updated last year
- Slides with modifications for a course at Tsinghua University.☆58Updated 2 years ago
- Offline Quantization Tools for Deploy.☆124Updated last year
- ☆224Updated 2 years ago
- play gemm with tvm☆89Updated last year
- ☆141Updated 2 years ago
- ☆132Updated 2 months ago
- TensorRT 2022复赛方案: 首个基于Transformer的图像重建模型MST++的TensorRT模型推断优化☆138Updated 2 years ago
- Fast CUDA Kernels for ResNet Inference.☆172Updated 5 years ago
- Benchmark code for the "Online normalizer calculation for softmax" paper☆85Updated 6 years ago
- Manually implemented quantization-aware training☆21Updated 2 years ago
- Examples of CUDA implementations by Cutlass CuTe☆143Updated last month
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆33Updated 3 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆197Updated 2 years ago
- An unofficial cuda assembler, for all generations of SASS, hopefully :)☆82Updated last year
- A Winograd Minimal Filter Implementation in CUDA☆24Updated 3 years ago
- ☆44Updated 3 years ago
- ☆34Updated last year
- ☆112Updated 11 months ago
- ☆35Updated 5 months ago
- A standalone GEMM kernel for fp16 activation and quantized weight, extracted from FasterTransformer☆89Updated 2 weeks ago
- ☆60Updated 2 months ago
- This is a demo how to write a high performance convolution run on apple silicon☆54Updated 3 years ago
- High Performance Grouped GEMM in PyTorch☆28Updated 2 years ago
- NART = NART is not A RunTime, a deep learning inference framework.☆38Updated 2 years ago
- how to design cpu gemm on x86 with avx256, that can beat openblas.☆68Updated 5 years ago