apple / ml-quantLinks
Research publication code for "Least Squares Binary Quantization of Neural Networks"
☆83Updated 2 years ago
Alternatives and similar repositories for ml-quant
Users that are interested in ml-quant are comparing it to the libraries listed below
Sorting:
- ☆57Updated 3 years ago
- Reference implementations of popular Binarized Neural Networks☆109Updated 3 weeks ago
- ☆19Updated 4 years ago
- PyProf2: PyTorch Profiling tool☆82Updated 5 years ago
- Official implementation of "UNAS: Differentiable Architecture Search Meets Reinforcement Learning", CVPR 2020 Oral☆63Updated 2 years ago
- Using ideas from product quantization for state-of-the-art neural network compression.☆146Updated 4 years ago
- Programmable Neural Network Compression☆149Updated 3 years ago
- Butterfly matrix multiplication in PyTorch☆177Updated 2 years ago
- A curated list of binary neural network research papers and software packages.☆28Updated 5 years ago
- ☆75Updated 3 years ago
- Customized matrix multiplication kernels☆57Updated 3 years ago
- ☆46Updated last year
- [JMLR'20] NeurIPS 2019 MicroNet Challenge Efficient Language Modeling, Champion☆41Updated 4 years ago
- Export utility for unconstrained channel pruned models☆72Updated 2 years ago
- A highly modular PyTorch framework with a focus on Neural Architecture Search (NAS).☆23Updated 4 years ago
- Implementation for the paper "Latent Weights Do Not Exist: Rethinking Binarized Neural Network Optimization"☆75Updated 6 years ago
- ☆22Updated 7 years ago
- ☆23Updated 3 years ago
- Code for High-Capacity Expert Binary Networks (ICLR 2021).☆27Updated 4 years ago
- A research library for pytorch-based neural network pruning, compression, and more.☆163Updated 3 years ago
- pytest plugin for a better developer experience when working with the PyTorch test suite☆44Updated 4 years ago
- 3rd place solution for NeurIPS 2019 MicroNet challenge☆37Updated 6 years ago
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆237Updated 2 years ago
- EfficientNet, MobileNetV3, MobileNetV2, MixNet, etc in JAX w/ Flax Linen and Objax☆129Updated last year
- PyTorch model training and layer saturation monitor☆83Updated 2 years ago
- End-to-end training of sparse deep neural networks with little-to-no performance loss.☆332Updated 2 years ago
- Simple Training and Deployment of Fast End-to-End Binary Networks☆159Updated 3 years ago
- Code for paper "SWALP: Stochastic Weight Averaging forLow-Precision Training".☆62Updated 6 years ago
- Discovering Neural Wirings (https://arxiv.org/abs/1906.00586)☆136Updated 3 weeks ago
- Implemented here a Binary Neural Network (BNN) achieving nearly state-of-art results but recorded a significant reduction in memory usage…☆75Updated 4 years ago