pytorch / QNNPACK
Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators
☆1,538Updated 5 years ago
Alternatives and similar repositories for QNNPACK:
Users that are interested in QNNPACK are comparing it to the libraries listed below
- Low-precision matrix multiplication☆1,799Updated last year
- Tensorflow Backend for ONNX☆1,302Updated last year
- A domain specific language to express machine learning workloads.☆1,759Updated last year
- TVM integration into PyTorch☆452Updated 5 years ago
- An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.☆2,862Updated 2 years ago
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆940Updated last week
- FeatherCNN is a high performance inference engine for convolutional neural networks.☆1,217Updated 5 years ago
- Acceleration package for neural networks on multi-core CPUs☆1,686Updated 10 months ago
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆533Updated 2 years ago
- FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) - https://code.fb.com/ml-applications/fbgemm/☆1,299Updated this week
- nGraph has moved to OpenVINO☆1,349Updated 4 years ago
- Arm NN ML Software. The code here is a read-only mirror of https://review.mlplatform.org/admin/repos/ml/armnn☆1,249Updated last week
- Memory consumption and FLOP count estimates for convnets☆918Updated 6 years ago
- Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distille…☆4,388Updated last year
- Code for: "And the bit goes down: Revisiting the quantization of neural networks"☆633Updated 4 years ago
- Benchmarking Neural Network Inference on Mobile Devices☆370Updated 2 years ago
- Mobile AI Compute Engine Model Zoo☆377Updated 3 years ago
- ☆1,659Updated 6 years ago
- Winograd minimal convolution algorithm generator for convolutional neural networks.☆615Updated 4 years ago
- TensorFlow/TensorRT integration☆741Updated last year
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆1,998Updated last week
- A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.☆1,531Updated 2 months ago
- MMdnn is a set of tools to help users inter-operate among different deep learning frameworks. E.g. model conversion and visualization. Co…☆5,811Updated 10 months ago
- Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)☆1,083Updated 11 months ago
- Compiler for Neural Network hardware accelerators☆3,279Updated 11 months ago
- A performant and modular runtime for TensorFlow☆759Updated this week
- [ICLR 2019] ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware☆1,437Updated 7 months ago
- AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.☆2,273Updated this week
- The convertor/conversion of deep learning models for different deep learning frameworks/softwares.☆3,247Updated last year
- Caffe Implementation of Google's MobileNets (v1 and v2)☆1,265Updated 3 years ago