larq / compute-engine
Highly optimized inference engine for Binarized Neural Networks
☆248Updated last week
Alternatives and similar repositories for compute-engine:
Users that are interested in compute-engine are comparing it to the libraries listed below
- Reference implementations of popular Binarized Neural Networks☆107Updated last week
- An Open-Source Library for Training Binarized Neural Networks☆713Updated 7 months ago
- Simple Training and Deployment of Fast End-to-End Binary Networks☆158Updated 3 years ago
- Graph Transforms to Quantize and Retrain Deep Neural Nets in TensorFlow☆168Updated 5 years ago
- Low Precision Arithmetic Simulation in PyTorch☆272Updated 10 months ago
- TFLite model analyzer & memory optimizer☆124Updated last year
- A small library for managing deep learning models, hyperparameters and datasets☆23Updated last year
- Customized matrix multiplication kernels☆53Updated 3 years ago
- A PyTorch implementation of "Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights"☆167Updated 5 years ago
- Scailable ONNX python tools☆97Updated 4 months ago
- ☆69Updated 2 years ago
- Training neural networks in TensorFlow 2.0 with 5x less memory☆130Updated 3 years ago
- Efficient Sparse-Winograd Convolutional Neural Networks (ICLR 2018)☆190Updated 5 years ago
- End-to-end training of sparse deep neural networks with little-to-no performance loss.☆319Updated 2 years ago
- Quantization of Convolutional Neural networks.☆244Updated 7 months ago
- Programmable Neural Network Compression☆148Updated 2 years ago
- TVM integration into PyTorch☆452Updated 5 years ago
- BMXNet 2: An Open-Source Binary Neural Network Implementation Based on MXNet☆231Updated 2 years ago
- A self-contained version of the tutorial which can be easily cloned and viewed by others.☆24Updated 5 years ago
- Parse TFLite models (*.tflite) EASILY with Python. Check the API at https://zhenhuaw.me/tflite/docs/☆98Updated last month
- A research library for pytorch-based neural network pruning, compression, and more.☆160Updated 2 years ago
- ☆67Updated 5 years ago
- Fast matrix multiplication for few-bit integer matrices on CPUs.☆27Updated 6 years ago
- tophub autotvm log collections☆70Updated 2 years ago
- Quantized Neural Networks - networks trained for inference at arbitrary low precision.☆146Updated 7 years ago
- A Pytorch implementation of Neural Network Compression (pruning, deep compression, channel pruning)☆155Updated 8 months ago
- PyTorch implementation of Data Free Quantization Through Weight Equalization and Bias Correction.☆260Updated last year
- Train neural networks with joint quantization and pruning on both weights and activations using any pytorch modules☆40Updated 2 years ago
- Fast sparse deep learning on CPUs☆52Updated 2 years ago
- A code generator from ONNX to PyTorch code☆135Updated 2 years ago