BayesWatch / pytorch-moonshineLinks
Cheap distillation for convolutional neural networks.
☆33Updated 6 years ago
Alternatives and similar repositories for pytorch-moonshine
Users that are interested in pytorch-moonshine are comparing it to the libraries listed below
Sorting:
- ☆51Updated 6 years ago
- [ICML 2018] "Deep k-Means: Re-Training and Parameter Sharing with Harder Cluster Assignments for Compressing Deep Convolutions"☆152Updated 3 years ago
- PyTorch implementation for GAL.☆56Updated 5 years ago
- Code for BlockSwap (ICLR 2020).☆33Updated 4 years ago
- ☆70Updated 5 years ago
- Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons (AAAI 2019)☆105Updated 6 years ago
- Pytorch implementation of TRP☆45Updated 5 years ago
- ☆47Updated 5 years ago
- [NeurIPS '18] "Can We Gain More from Orthogonality Regularizations in Training Deep CNNs?" Official Implementation.☆129Updated 3 years ago
- Implementation of soft parameter sharing for neural networks☆69Updated 4 years ago
- PyTorch implementation of shake-drop regularization☆55Updated 5 years ago
- A Re-implementation of Fixed-update Initialization☆155Updated 6 years ago
- A PyTorch implementation of shake-shake☆111Updated 5 years ago
- Code for paper "SWALP: Stochastic Weight Averaging forLow-Precision Training".☆62Updated 6 years ago
- ☆55Updated 6 years ago
- ☆23Updated 6 years ago
- Scripts for Imagenet 32 dataset☆153Updated 4 years ago
- PyTorch implementation for Convolutional Networks with Adaptive Inference Graphs☆185Updated 6 years ago
- An official collection of code in different frameworks that reproduces experiments in "Group Normalization"☆117Updated 4 years ago
- Unofficial pytorch implementation of Born-Again Neural Networks.☆55Updated 4 years ago
- Code for https://arxiv.org/abs/1810.04622☆140Updated 6 years ago
- [NeurIPS 2020 Oral] Is normalization indispensable for training deep neural networks?☆34Updated 3 years ago
- BlockDrop: Dynamic Inference Paths in Residual Networks☆142Updated 2 years ago
- Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression☆49Updated 2 years ago
- Source code accompanying our CVPR 2019 paper: "NetTailor: Tuning the architecture, not just the weights."☆53Updated 4 years ago
- SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY☆114Updated 6 years ago
- Zero-Shot Knowledge Distillation in Deep Networks☆67Updated 3 years ago
- "Layer-wise Adaptive Rate Scaling" in PyTorch☆87Updated 4 years ago
- Code for "EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis" https://arxiv.org/abs/1905.05934☆113Updated 5 years ago
- DeepHoyer: Learning Sparser Neural Network with Differentiable Scale-Invariant Sparsity Measures☆33Updated 5 years ago