SonySemiconductorSolutions / mct-model-optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆432Updated this week
Alternatives and similar repositories for mct-model-optimization
Users that are interested in mct-model-optimization are comparing it to the libraries listed below
Sorting:
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆303Updated last year
- ☆342Updated 2 years ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆648Updated last year
- PyTorch Quantization Aware Training Example☆150Updated last year
- A parser, editor and profiler tool for ONNX models.☆478Updated 3 months ago
- Inference of quantization aware trained networks using TensorRT☆83Updated 3 years ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆310Updated last year
- TFLite model analyzer & memory optimizer☆135Updated 2 years ago
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆919Updated last week
- Count number of parameters / MACs / FLOPS for ONNX models.☆95Updated last year
- A code generator from ONNX to PyTorch code☆142Updated 3 years ago
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆345Updated 3 years ago
- ☆208Updated 4 years ago
- Conversion of PyTorch Models into TFLite☆399Updated 2 years ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆922Updated last year
- μNAS is a neural architecture search (NAS) system that designs small-yet-powerful microcontroller-compatible neural networks.☆82Updated 5 years ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆175Updated this week
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆129Updated 9 months ago
- Common utilities for ONNX converters☆294Updated last month
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆453Updated 2 years ago
- ☆23Updated last month
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆37Updated 4 years ago
- ☆170Updated 2 years ago
- Transform ONNX model to PyTorch representation☆345Updated 3 months ago
- ONNX Optimizer☆795Updated this week
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆225Updated last year
- On-Device Training Under 256KB Memory [NeurIPS'22]☆510Updated last year
- Convert ONNX models to PyTorch.☆725Updated 3 months ago
- A DNN inference latency prediction toolkit for accurately modeling and predicting the latency on diverse edge devices.☆364Updated last year
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆96Updated this week