pytorch / QNNPACKLinks
Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators
☆1,547Updated 6 years ago
Alternatives and similar repositories for QNNPACK
Users that are interested in QNNPACK are comparing it to the libraries listed below
Sorting:
- Low-precision matrix multiplication☆1,819Updated last year
- FeatherCNN is a high performance inference engine for convolutional neural networks.☆1,223Updated 6 years ago
- An Automatic Model Compression (AutoMC) framework for developing smaller and faster AI applications.☆2,909Updated 2 years ago
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆954Updated 7 months ago
- TVM integration into PyTorch☆455Updated 5 years ago
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆535Updated 3 years ago
- Tensorflow Backend for ONNX☆1,327Updated last year
- nGraph has moved to OpenVINO☆1,345Updated 5 years ago
- Memory consumption and FLOP count estimates for convnets☆931Updated 6 years ago
- [ICLR 2019] ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware☆1,451Updated last year
- Code for: "And the bit goes down: Revisiting the quantization of neural networks"☆631Updated 5 years ago
- Benchmarking Neural Network Inference on Mobile Devices☆383Updated 2 years ago
- Acceleration package for neural networks on multi-core CPUs☆1,702Updated last year
- Arm NN ML Software.☆1,290Updated last week
- Winograd minimal convolution algorithm generator for convolutional neural networks.☆624Updated 5 years ago
- Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)☆1,085Updated last year
- A domain specific language to express machine learning workloads.☆1,759Updated 2 years ago
- Embedded and mobile deep learning research resources☆758Updated 2 years ago
- Fast & Simple Resource-Constrained Learning of Deep Network Structure☆1,032Updated 5 months ago
- ☆1,655Updated 7 years ago
- A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.☆1,559Updated last week
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,187Updated this week
- TensorFlow/TensorRT integration☆744Updated 2 years ago
- Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distille…☆4,398Updated 2 years ago
- Dive into Deep Learning Compiler☆646Updated 3 years ago
- Benchmarking Deep Learning operations on different hardware☆1,102Updated 4 years ago
- A performant and modular runtime for TensorFlow☆757Updated 3 months ago
- Facebook AI Performance Evaluation Platform☆391Updated this week
- ImageNet classification using binary Convolutional Neural Networks☆868Updated 8 years ago
- Mobile AI Compute Engine Model Zoo☆376Updated 4 years ago