sony / model_optimization
Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. This project provides researchers, developers, and engineers advanced quantization and compression tools for deploying state-of-the-art neural networks.
☆366Updated this week
Alternatives and similar repositories for model_optimization:
Users that are interested in model_optimization are comparing it to the libraries listed below
- A set of simple tools for splitting, merging, OP deletion, size compression, rewriting attributes and constants, OP generation, change op…☆286Updated 10 months ago
- Supporting PyTorch models with the Google AI Edge TFLite runtime.☆462Updated this week
- ☆313Updated last year
- A parser, editor and profiler tool for ONNX models.☆418Updated last month
- This script converts the ONNX/OpenVINO IR model to Tensorflow's saved_model, tflite, h5, tfjs, tftrt(TensorRT), CoreML, EdgeTPU, ONNX and…☆340Updated 2 years ago
- ☆202Updated 3 years ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆978Updated this week
- TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillati…☆755Updated last week
- ☆139Updated last year
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆521Updated 11 months ago
- Conversion of PyTorch Models into TFLite☆369Updated last year
- Count number of parameters / MACs / FLOPS for ONNX models.☆89Updated 4 months ago
- Actively maintained ONNX Optimizer☆673Updated last month
- Quantization library for PyTorch. Support low-precision and mixed-precision quantization, with hardware implementation through TVM.☆427Updated last year
- NVIDIA DLA-SW, the recipes and tools for running deep learning workloads on NVIDIA DLA cores for inference applications.☆189Updated 8 months ago
- TFLite model analyzer & memory optimizer☆122Updated last year
- Common utilities for ONNX converters☆259Updated 3 months ago
- A code generator from ONNX to PyTorch code☆135Updated 2 years ago
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆300Updated 5 months ago
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- Implementation of YOLOv9 QAT optimized for deployment on TensorRT platforms.☆100Updated last week
- A tutorial for getting started with the Deep Learning Accelerator (DLA) on NVIDIA Jetson☆319Updated 2 years ago
- PyTorch emulation library for Microscaling (MX)-compatible data formats☆204Updated 5 months ago
- ☆19Updated this week
- PyTorch Quantization Aware Training Example☆130Updated 9 months ago
- Self-Created Tools to convert ONNX files (NCHW) to TensorFlow/TFLite/Keras format (NHWC). The purpose of this tool is to solve the massiv…☆755Updated 2 weeks ago
- Pytorch implementation of BRECQ, ICLR 2021☆266Updated 3 years ago
- Offline Quantization Tools for Deploy.☆123Updated last year
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆260Updated last year
- PyTorch implementation for the APoT quantization (ICLR 2020)☆272Updated 2 months ago