MegEngine / examples
A set of examples around MegEngine
☆31Updated last year
Alternatives and similar repositories for examples:
Users that are interested in examples are comparing it to the libraries listed below
- MegEngine到其他框架的转换器☆69Updated last year
- ☆34Updated 2 years ago
- TensorRT 2022复赛方案: 首个基于Transformer的图像重建模型MST++的TensorRT模型推断优化☆138Updated 2 years ago
- Slides with modifications for a course at Tsinghua University.☆59Updated 2 years ago
- Offline Quantization Tools for Deploy.☆126Updated last year
- Pytorch implementation of RAPQ, IJCAI 2022☆21Updated last year
- base quantization methods including: QAT, PTQ, per_channel, per_tensor, dorefa, lsq, adaround, omse, Histogram, bias_correction.etc☆43Updated 2 years ago
- 🐱 ncnn int8 模型量化评估☆14Updated 2 years ago
- CUDA Templates for Linear Algebra Subroutines☆98Updated 11 months ago
- NART = NART is not A RunTime, a deep learning inference framework.☆38Updated 2 years ago
- The official implementation of the NeurIPS 2022 paper Q-ViT.☆87Updated last year
- ☆28Updated 3 years ago
- ☆44Updated 3 years ago
- quantize aware training package for NCNN on pytorch☆70Updated 3 years ago
- [ICML 2022] "DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural Networks", by Yonggan …☆35Updated 2 years ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆33Updated 3 years ago
- Collections of model quantization algorithms. Any issues, please contact Peng Chen (blueardour@gmail.com)☆42Updated 3 years ago
- LSQ+ or LSQplus☆64Updated 2 months ago
- CUDA 6大并行计算模式 代码与笔记☆60Updated 4 years ago
- BitSplit Post-trining Quantization☆49Updated 3 years ago
- [CVPRW 2021] Dynamic-OFA: Runtime DNN Architecture Switching for Performance Scaling on Heterogeneous Embedded Platforms☆29Updated 2 years ago
- Based of paper "Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference"☆63Updated 4 years ago
- TensorRT 2022 亚军方案,tensorrt加速mobilevit模型☆64Updated 2 years ago
- A PyTorch Framework for Efficient Pruning and Quantization for specialized accelerators.☆34Updated 3 years ago
- [ICLR 2022 Oral] F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization☆94Updated 2 years ago
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- PyTorch Quantization Aware Training(QAT,量化感知训练)☆30Updated last year
- ☆24Updated last year
- arm-neon☆90Updated 8 months ago
- Post-Training Quantization for Vision transformers.☆214Updated 2 years ago