sony / model_optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆400Updated this week
Alternatives and similar repositories for model_optimization
Users that are interested in model_optimization are comparing it to the libraries listed below
Sorting:
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆296Updated last year
- A parser, editor and profiler tool for ONNX models.☆442Updated 2 weeks ago
- ☆328Updated last year
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆342Updated 2 years ago
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆113Updated 2 months ago
- TFLite model analyzer & memory optimizer☆127Updated last year
- ☆149Updated 2 years ago
- ONNX Optimizer☆723Updated last week
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,051Updated this week
- ☆21Updated last month
- Conversion of PyTorch Models into TFLite☆382Updated 2 years ago
- ☆205Updated 3 years ago
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆814Updated last week
- Pytorch to Keras/Tensorflow/TFLite conversion made intuitive☆313Updated 3 months ago
- Transform ONNX model to PyTorch representation☆338Updated 7 months ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆579Updated last year
- Offline Quantization Tools for Deploy.☆129Updated last year
- Inference of quantization aware trained networks using TensorRT☆82Updated 2 years ago
- PyTorch Quantization Aware Training Example☆136Updated last year
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆332Updated 3 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆200Updated last year
- Common utilities for ONNX converters☆272Updated 6 months ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆441Updated 2 years ago
- C++ Helper Class for Deep Learning Inference Frameworks: TensorFlow Lite, TensorRT, OpenCV, OpenVINO, ncnn, MNN, SNPE, Arm NN, NNabla, ON…☆289Updated 3 years ago
- Pytorch implementation of BRECQ, ICLR 2021☆276Updated 3 years ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆150Updated this week
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆262Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆93Updated 8 months ago
- μNAS is a neural architecture search (NAS) system that designs small-yet-powerful microcontroller-compatible neural networks.☆80Updated 4 years ago
- Model compression for ONNX☆96Updated 7 months ago