NVIDIA / sampleQATLinks
Inference of quantization aware trained networks using TensorRT
☆82Updated 2 years ago
Alternatives and similar repositories for sampleQAT
Users that are interested in sampleQAT are comparing it to the libraries listed below
Sorting:
- ☆69Updated 2 years ago
- PyTorch Quantization Aware Training Example☆136Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆93Updated 7 months ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆34Updated 3 years ago
- Benchmark scripts for TVM☆74Updated 3 years ago
- Offline Quantization Tools for Deploy.☆129Updated last year
- Benchmark of TVM quantized model on CUDA☆111Updated 5 years ago
- benchmark for embededded-ai deep learning inference engines, such as NCNN / TNN / MNN / TensorFlow Lite etc.☆204Updated 4 years ago
- Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming☆97Updated 4 years ago
- ☆149Updated 2 years ago
- Pytorch implementation of BRECQ, ICLR 2021☆276Updated 3 years ago
- MegEngine到其他框架的转换器☆70Updated 2 years ago
- A code generator from ONNX to PyTorch code☆138Updated 2 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆200Updated 3 years ago
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆262Updated last year
- A sample for onnxparser working with trt user defined plugins for TRT7.0☆168Updated 4 years ago
- tophub autotvm log collections☆69Updated 2 years ago
- symmetric int8 gemm☆66Updated 5 years ago
- NART = NART is not A RunTime, a deep learning inference framework.☆37Updated 2 years ago
- PyTorch Static Quantization Example☆38Updated 4 years ago
- code reading for tvm☆76Updated 3 years ago
- quantize aware training package for NCNN on pytorch☆70Updated 3 years ago
- play gemm with tvm☆91Updated last year
- Code for our paper at ECCV 2020: Post-Training Piecewise Linear Quantization for Deep Neural Networks☆69Updated 3 years ago
- A set of examples around MegEngine☆31Updated last year
- ☆44Updated 3 years ago
- ☆36Updated 8 months ago
- Parallel CUDA implementation of NON maximum Suppression☆79Updated 4 years ago
- ☆236Updated 2 years ago
- [ICLR 2022 Oral] F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization☆95Updated 3 years ago