SonySemiconductorSolutions / mct-model-optimizationLinks
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆424Updated last week
Alternatives and similar repositories for mct-model-optimization
Users that are interested in mct-model-optimization are comparing it to the libraries listed below
Sorting:
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆299Updated last year
- ☆338Updated last year
- TFLite model analyzer & memory optimizer☆132Updated last year
- A parser, editor and profiler tool for ONNX models.☆465Updated last week
- Conversion of PyTorch Models into TFLite☆395Updated 2 years ago
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆129Updated 6 months ago
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆875Updated 3 weeks ago
- Convert ONNX models to PyTorch.☆709Updated last month
- PyTorch Quantization Aware Training Example☆144Updated last year
- Inference of quantization aware trained networks using TensorRT☆83Updated 2 years ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆309Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆95Updated last year
- ☆207Updated 4 years ago
- Common utilities for ONNX converters☆283Updated 2 months ago
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆343Updated 3 years ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,100Updated this week
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆619Updated last year
- Transform ONNX model to PyTorch representation☆342Updated last week
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆164Updated this week
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆219Updated last year
- ONNX Optimizer☆770Updated 2 weeks ago
- ☆166Updated this week
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆828Updated this week
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆36Updated 3 years ago
- ☆23Updated last week
- A code generator from ONNX to PyTorch code☆141Updated 3 years ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆451Updated 2 years ago
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆356Updated 3 years ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆408Updated this week
- ☆163Updated 2 years ago