apple / ml-quantLinks
Research publication code for "Least Squares Binary Quantization of Neural Networks"
☆83Updated 2 years ago
Alternatives and similar repositories for ml-quant
Users that are interested in ml-quant are comparing it to the libraries listed below
Sorting:
- ☆57Updated 3 years ago
- ☆19Updated 3 years ago
- Reference implementations of popular Binarized Neural Networks☆108Updated last month
- Butterfly matrix multiplication in PyTorch☆174Updated last year
- Using ideas from product quantization for state-of-the-art neural network compression.☆146Updated 3 years ago
- Programmable Neural Network Compression☆148Updated 3 years ago
- A research library for pytorch-based neural network pruning, compression, and more.☆162Updated 2 years ago
- A highly modular PyTorch framework with a focus on Neural Architecture Search (NAS).☆23Updated 3 years ago
- Official implementation of "UNAS: Differentiable Architecture Search Meets Reinforcement Learning", CVPR 2020 Oral☆61Updated last year
- Code for High-Capacity Expert Binary Networks (ICLR 2021).☆27Updated 3 years ago
- Customized matrix multiplication kernels☆56Updated 3 years ago
- End-to-end training of sparse deep neural networks with little-to-no performance loss.☆324Updated 2 years ago
- Export utility for unconstrained channel pruned models☆71Updated 2 years ago
- Factorized Neural Layers☆29Updated 2 years ago
- ☆46Updated last year
- ☆22Updated 7 years ago
- [JMLR'20] NeurIPS 2019 MicroNet Challenge Efficient Language Modeling, Champion☆40Updated 4 years ago
- A curated list of binary neural network research papers and software packages.☆27Updated 5 years ago
- PyTorch interface for the IPU☆180Updated last year
- A collection of metrics to profile a single deep learning model or compare two different deep learning models☆26Updated last year
- [ECCV 2022] SuperTickets: Drawing Task-Agnostic Lottery Tickets from Supernets via Jointly Architecture Searching and Parameter Pruning☆21Updated 3 years ago
- ☆10Updated 3 years ago
- PyProf2: PyTorch Profiling tool☆82Updated 5 years ago
- Implementation for the paper "Latent Weights Do Not Exist: Rethinking Binarized Neural Network Optimization"☆74Updated 5 years ago
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆236Updated 2 years ago
- Simple Training and Deployment of Fast End-to-End Binary Networks☆157Updated 3 years ago
- Online Normalization for Training Neural Networks (Companion Repository)☆83Updated 4 years ago
- Code for paper "SWALP: Stochastic Weight Averaging forLow-Precision Training".☆62Updated 6 years ago
- Binarize convolutional neural networks using pytorch☆147Updated 3 years ago
- Repository containing pruned models and related information☆37Updated 4 years ago