SonySemiconductorSolutions / mct-model-optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆418Updated last week
Alternatives and similar repositories for mct-model-optimization
Users that are interested in mct-model-optimization are comparing it to the libraries listed below
Sorting:
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆297Updated last year
- ☆335Updated last year
- A parser, editor and profiler tool for ONNX models.☆457Updated 2 months ago
- TFLite model analyzer & memory optimizer☆131Updated last year
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆600Updated last year
- Conversion of PyTorch Models into TFLite☆391Updated 2 years ago
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- PyTorch Quantization Aware Training Example☆140Updated last year
- A code generator from ONNX to PyTorch code☆141Updated 2 years ago
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆857Updated 2 months ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆311Updated last year
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆126Updated 5 months ago
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆343Updated 2 years ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆94Updated 11 months ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,085Updated this week
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆161Updated this week
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆35Updated 3 years ago
- ☆206Updated 3 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆218Updated last year
- Transform ONNX model to PyTorch representation☆340Updated 10 months ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆450Updated 2 years ago
- ☆161Updated 2 years ago
- ☆157Updated 3 months ago
- ONNX Optimizer☆760Updated this week
- Common utilities for ONNX converters☆281Updated last month
- Convert tflite to JSON and make it editable in the IDE. It also converts the edited JSON back to tflite binary.☆27Updated 2 years ago
- Convert ONNX models to PyTorch.☆702Updated last year
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆895Updated 10 months ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆793Updated this week
- A library for researching neural networks compression and acceleration methods.☆139Updated last month