dianhsu / swin-transformer-cppLinks
Swin Transformer C++ Implementation
☆64Updated 4 years ago
Alternatives and similar repositories for swin-transformer-cpp
Users that are interested in swin-transformer-cpp are comparing it to the libraries listed below
Sorting:
- 用C++实现一个简单的Transformer模型。 Attention Is All You Need。☆52Updated 4 years ago
- CUDA Templates for Linear Algebra Subroutines☆100Updated last year
- ☆38Updated last year
- Code for ACM MobiCom 2024 paper "FlexNN: Efficient and Adaptive DNN Inference on Memory-Constrained Edge Devices"☆56Updated 10 months ago
- PyTorch Quantization Aware Training Example☆144Updated last year
- Several optimization methods of half-precision general matrix vector multiplication (HGEMV) using CUDA core.☆69Updated last year
- ☆21Updated 4 years ago
- ☆19Updated last year
- A Winograd Minimal Filter Implementation in CUDA☆28Updated 4 years ago
- ☆165Updated 2 years ago
- Benchmark code for the "Online normalizer calculation for softmax" paper☆102Updated 7 years ago
- ☆207Updated 4 years ago
- CUDA 6大并行计算模式 代码与笔记☆61Updated 5 years ago
- CPU Memory Compiler and Parallel programing☆26Updated last year
- Common libraries for PPL projects☆30Updated 8 months ago
- A set of examples around MegEngine☆31Updated last year
- CUDA 8-bit Tensor Core Matrix Multiplication based on m16n16k16 WMMA API☆33Updated 2 years ago
- SGEMM optimization with cuda step by step☆20Updated last year
- ☆98Updated 4 years ago
- 使用 cutlass 仓库在 ada 架构上 实现 fp8 的 flash attention☆78Updated last year
- Standalone Flash Attention v2 kernel without libtorch dependency☆112Updated last year
- This is a demo how to write a high performance convolution run on apple silicon☆57Updated 3 years ago
- Tencent Distribution of TVM☆15Updated 2 years ago
- PyTorch extension for emulating FP8 data formats on standard FP32 Xeon/GPU hardware.☆112Updated 11 months ago
- CUDA Matrix Multiplication Optimization☆239Updated last year
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆36Updated 3 years ago
- how to design cpu gemm on x86 with avx256, that can beat openblas.☆73Updated 6 years ago
- Efficient operation implementation based on the Cambricon Machine Learning Unit (MLU) .☆140Updated last week
- ☆60Updated last year
- ☆143Updated last year