sony / model_optimization
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆388Updated this week
Alternatives and similar repositories for model_optimization:
Users that are interested in model_optimization are comparing it to the libraries listed below
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆290Updated last year
- Common utilities for ONNX converters☆267Updated 4 months ago
- A parser, editor and profiler tool for ONNX models.☆425Updated 3 months ago
- ☆143Updated 2 years ago
- ☆321Updated last year
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆340Updated 2 years ago
- ONNX Optimizer☆696Updated 3 weeks ago
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆431Updated last year
- PyTorch Quantization Aware Training Example☆135Updated 11 months ago
- ☆21Updated last week
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆559Updated last year
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,002Updated this week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆341Updated this week
- A code generator from ONNX to PyTorch code☆136Updated 2 years ago
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆193Updated 10 months ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆91Updated 6 months ago
- ☆203Updated 3 years ago
- TFLite model analyzer & memory optimizer☆125Updated last year
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆326Updated 2 years ago
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆555Updated this week
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆107Updated 2 months ago
- Transform ONNX model to PyTorch representation☆332Updated 5 months ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆302Updated 7 months ago
- Conversion of PyTorch Models into TFLite☆375Updated 2 years ago
- Pytorch to Keras/Tensorflow/TFLite conversion made intuitive☆304Updated last month
- ☆222Updated 2 years ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆145Updated this week
- Convert ONNX models to PyTorch.☆666Updated 8 months ago
- Model compression for ONNX☆91Updated 5 months ago