SonySemiconductorSolutions / mct-model-optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆428Updated this week
Alternatives and similar repositories for mct-model-optimization
Users that are interested in mct-model-optimization are comparing it to the libraries listed below
Sorting:
- TFLite model analyzer & memory optimizer☆135Updated last year
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆301Updated last year
- ☆339Updated 2 years ago
- Conversion of PyTorch Models into TFLite☆398Updated 2 years ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆633Updated last year
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆897Updated this week
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆344Updated 3 years ago
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- Pytorch to Keras/Tensorflow/TFLite conversion made intuitive☆335Updated 9 months ago
- A code generator from ONNX to PyTorch code☆141Updated 3 years ago
- Transform ONNX model to PyTorch representation☆344Updated last month
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆310Updated last year
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,112Updated this week
- A parser, editor and profiler tool for ONNX models.☆468Updated last month
- Count number of parameters / MACs / FLOPS for ONNX models.☆95Updated last year
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆168Updated this week
- Common utilities for ONNX converters☆289Updated last week
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆129Updated 8 months ago
- ☆23Updated 2 weeks ago
- PyTorch Quantization Aware Training Example☆146Updated last year
- Convert ONNX models to PyTorch.☆716Updated 2 months ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆37Updated 4 years ago
- ONNX Optimizer☆780Updated last month
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆224Updated last year
- ☆208Updated 4 years ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆913Updated last year
- ☆170Updated 2 weeks ago
- Convert tflite to JSON and make it editable in the IDE. It also converts the edited JSON back to tflite binary.☆28Updated 2 years ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆414Updated this week
- ☆168Updated 2 years ago