SonySemiconductorSolutions / mct-model-optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆414Updated 2 months ago
Alternatives and similar repositories for mct-model-optimization
Users that are interested in mct-model-optimization are comparing it to the libraries listed below
Sorting:
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆297Updated last year
- ☆334Updated last year
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆599Updated last year
- TFLite model analyzer & memory optimizer☆131Updated last year
- A parser, editor and profiler tool for ONNX models.☆454Updated last month
- Count number of parameters / MACs / FLOPS for ONNX models.☆94Updated 10 months ago
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- Conversion of PyTorch Models into TFLite☆390Updated 2 years ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆157Updated last month
- A code generator from ONNX to PyTorch code☆140Updated 2 years ago
- PyTorch Quantization Aware Training Example☆140Updated last year
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆213Updated last year
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆343Updated 2 years ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆310Updated 11 months ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,075Updated last week
- Common utilities for ONNX converters☆279Updated last week
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆850Updated last month
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆123Updated 4 months ago
- ☆206Updated 3 years ago
- ☆155Updated 2 months ago
- ☆158Updated 2 years ago
- Transform ONNX model to PyTorch representation☆339Updated 10 months ago
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆62Updated last week
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆889Updated 9 months ago
- Convert ONNX models to PyTorch.☆698Updated last year
- μNAS is a neural architecture search (NAS) system that designs small-yet-powerful microcontroller-compatible neural networks.☆81Updated 4 years ago
- Pytorch to Keras/Tensorflow/TFLite conversion made intuitive☆324Updated 6 months ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆445Updated 2 years ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆777Updated this week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆378Updated last week