BayesWatch / pytorch-moonshineLinks
Cheap distillation for convolutional neural networks.
☆35Updated 7 years ago
Alternatives and similar repositories for pytorch-moonshine
Users that are interested in pytorch-moonshine are comparing it to the libraries listed below
Sorting:
- ☆51Updated 6 years ago
- PyTorch implementation for GAL.☆56Updated 5 years ago
- ☆47Updated 6 years ago
- [NeurIPS '18] "Can We Gain More from Orthogonality Regularizations in Training Deep CNNs?" Official Implementation.☆130Updated 3 years ago
- ☆69Updated 5 years ago
- PyTorch implementation of shake-drop regularization☆55Updated 5 years ago
- Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons (AAAI 2019)☆105Updated 6 years ago
- Pytorch implementation of TRP☆45Updated 5 years ago
- Unofficial pytorch implementation of Born-Again Neural Networks.☆56Updated 4 years ago
- Implementation of soft parameter sharing for neural networks☆70Updated 5 years ago
- "Layer-wise Adaptive Rate Scaling" in PyTorch☆87Updated 4 years ago
- Code for BlockSwap (ICLR 2020).☆33Updated 4 years ago
- Lookahead: A Far-sighted Alternative of Magnitude-based Pruning (ICLR 2020)☆32Updated 5 years ago
- Zero-Shot Knowledge Distillation in Deep Networks☆67Updated 3 years ago
- ☆23Updated 7 years ago
- BlockDrop: Dynamic Inference Paths in Residual Networks☆143Updated 3 years ago
- Codes for Understanding Architectures Learnt by Cell-based Neural Architecture Search☆28Updated 5 years ago
- Code for the paper "Training CNNs with Selective Allocation of Channels" (ICML 2019)☆25Updated 6 years ago
- [ICML 2018] "Deep k-Means: Re-Training and Parameter Sharing with Harder Cluster Assignments for Compressing Deep Convolutions"☆152Updated 3 years ago
- Pytorch Implementation of Neural Architecture Optimization☆113Updated 4 years ago
- SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY☆116Updated 6 years ago
- Delta Orthogonal Initialization for PyTorch☆18Updated 7 years ago
- Code for "Picking Winning Tickets Before Training by Preserving Gradient Flow" https://openreview.net/pdf?id=SkgsACVKPH☆105Updated 5 years ago
- Knowledge Distillation with Adversarial Samples Supporting Decision Boundary (AAAI 2019)☆71Updated 6 years ago
- An official collection of code in different frameworks that reproduces experiments in "Group Normalization"☆118Updated 4 years ago
- Source code accompanying our CVPR 2019 paper: "NetTailor: Tuning the architecture, not just the weights."☆53Updated 4 years ago
- DeepHoyer: Learning Sparser Neural Network with Differentiable Scale-Invariant Sparsity Measures☆32Updated 5 years ago
- PyTorch Code for "Evaluating the search phase of Neural Architecture Search" @ ICLR 2020☆49Updated 6 years ago
- Zero-Shot Knowledge Distillation in Deep Networks in ICML2019☆49Updated 6 years ago
- ☆34Updated 7 years ago