SonySemiconductorSolutions / mct-quantization-layers
☆21Updated this week
Alternatives and similar repositories for mct-quantization-layers
Users that are interested in mct-quantization-layers are comparing it to the libraries listed below
Sorting:
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆394Updated last week
- TFLite model analyzer & memory optimizer☆126Updated last year
- AI Edge Quantizer: flexible post training quantization for LiteRT models.☆32Updated last week
- [ICASSP'22] Integer-only Zero-shot Quantization for Efficient Speech Recognition☆31Updated 3 years ago
- ResNet Implementation, Training, and Inference Using LibTorch C++ API☆40Updated 11 months ago
- Scailable ONNX python tools☆97Updated 6 months ago
- Converting a deep neural network to integer-only inference in native C via uniform quantization and the fixed-point representation.☆23Updated 3 years ago
- PyTorch Quantization Aware Training Example☆135Updated 11 months ago
- ncnn HiFi-GAN☆26Updated 7 months ago
- ☆225Updated 2 years ago
- Neural Architecture Search for Neural Network Libraries☆59Updated last year
- This repository contains the results and code for the MLPerf™ Tiny Inference v0.7 benchmark.☆17Updated last year
- Awesome Quantization Paper lists with Codes☆11Updated 4 years ago
- Fork of Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://nervanasystems.githu…☆15Updated 9 months ago
- The official, proof-of-concept C++ implementation of PocketNN.☆32Updated 11 months ago
- ONNX and TensorRT implementation of Whisper☆61Updated last year
- Inference of quantization aware trained networks using TensorRT☆80Updated 2 years ago
- PyTorch to TensorFlow Lite converter☆183Updated 9 months ago
- ONNXモデルをpyca/cryptographyを用いて暗号化/復号化するサンプル☆16Updated 3 years ago
- Count number of parameters / MACs / FLOPS for ONNX models.☆92Updated 6 months ago
- A very simple tool for situations where optimization with onnx-simplifier would exceed the Protocol Buffers upper file size limit of 2GB,…☆16Updated last year
- Convert tflite to JSON and make it editable in the IDE. It also converts the edited JSON back to tflite binary.☆27Updated 2 years ago
- CMix-NN: Mixed Low-Precision CNN Library for Memory-Constrained Edge Devices☆41Updated 5 years ago
- In this repository, we explore model compression for transformer architectures via quantization. We specifically explore quantization awa…☆24Updated 4 years ago
- Snapdragon Neural Processing Engine (SNPE) SDKThe Snapdragon Neural Processing Engine (SNPE) is a Qualcomm Snapdragon software accelerate…☆34Updated 3 years ago
- Python library to work with the Visual Wake Words Dataset.☆35Updated 4 years ago
- Test Framework for few-shot open set KWS☆31Updated 6 months ago
- FakeQuantize with Learned Step Size(LSQ+) as Observer in PyTorch☆34Updated 3 years ago
- silero-vad pytorch implement☆17Updated 5 months ago
- QONNX: Arbitrary-Precision Quantized Neural Networks in ONNX☆148Updated this week