Cheap distillation for convolutional neural networks.
☆35Oct 22, 2018Updated 7 years ago
Alternatives and similar repositories for pytorch-moonshine
Users that are interested in pytorch-moonshine are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Code for BlockSwap (ICLR 2020).☆33Mar 25, 2021Updated 5 years ago
- Successfully training approximations to full-rank matrices for efficiency in deep learning.☆16Jan 5, 2021Updated 5 years ago
- Hybrid Binary Networks: Optimizing for Accuracy, Efficiency and Memory - WACV18☆12Jul 5, 2019Updated 6 years ago
- Temporary Discriminator GAN☆14Jul 21, 2020Updated 5 years ago
- ☆15Jan 8, 2020Updated 6 years ago
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- Contrastive Object-level Pre-training with Spatial Noise Curriculum Learning☆20Feb 4, 2022Updated 4 years ago
- Fast Image Restoration with Multi-bin Trainable Linear Units.☆11Dec 23, 2019Updated 6 years ago
- This is the official implementation of "DHP: Differentiable Meta Pruning via HyperNetworks".☆58Jan 17, 2026Updated 2 months ago
- Mode normalization (ICLR 2019).☆18Feb 27, 2019Updated 7 years ago
- Accompanying code for the paper "Zero-shot Knowledge Transfer via Adversarial Belief Matching"☆143Apr 29, 2020Updated 5 years ago
- Combined Group and Exclusive Sparsity for Deep Neural Networks, ICML 2017☆32Mar 21, 2018Updated 8 years ago
- Group Sparsity: The Hinge Between Filter Pruning and Decomposition for Network Compression. CVPR2020.☆64Jan 17, 2026Updated 2 months ago
- ☆47Dec 26, 2019Updated 6 years ago
- DeepHoyer: Learning Sparser Neural Network with Differentiable Scale-Invariant Sparsity Measures☆32Aug 13, 2020Updated 5 years ago
- Serverless GPU API endpoints on Runpod - Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- ☆137Oct 22, 2018Updated 7 years ago
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Mar 24, 2023Updated 3 years ago
- A drop-in replacement for CIFAR-10.☆246Mar 7, 2021Updated 5 years ago
- ☆17Mar 28, 2022Updated 4 years ago
- ☆19May 28, 2020Updated 5 years ago
- [ECCV18] Constraint-Aware Deep Neural Network Compression☆12Sep 11, 2018Updated 7 years ago
- My note-books for Kaggle Competition☆18Feb 7, 2018Updated 8 years ago
- Fast Autoaugment implementation for PyTorch☆10Jul 24, 2019Updated 6 years ago
- Try to train a Triplet-Siamese-Netwrok with the constrained Triplet Loss for few shot image classification.☆14Dec 26, 2019Updated 6 years ago
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- Implementation for the paper "Latent Weights Do Not Exist: Rethinking Binarized Neural Network Optimization"☆75Dec 8, 2019Updated 6 years ago
- [ECCV 2022] Prediction-Guided Distillation for Dense Object Detection☆62Oct 31, 2022Updated 3 years ago
- My notes and assessments from the 2021 "Complete Python Bootcamp From Zero to Hero in Python" course on Udemy. 🐍☆15Feb 5, 2021Updated 5 years ago
- Code for https://arxiv.org/abs/1810.04622☆140Aug 28, 2019Updated 6 years ago
- Implemented One Shot Face Recognition Using Siamese Neural Network using Pytorch on AT&T dataset☆13Jan 31, 2020Updated 6 years ago
- An implementation of Squared Earth-Mover's Distance loss for Neural Networks.☆14Mar 25, 2023Updated 3 years ago
- PyTorch implementation of "SNAPSHOT ENSEMBLES: TRAIN 1, GET M FOR FREE" [WIP]☆36May 20, 2017Updated 8 years ago
- Implementation of several knowledge distillation techniques on PyTorch☆15Feb 25, 2019Updated 7 years ago
- A PyTorch implementation of [VCT](https://github.com/google-research/google-research/tree/master/vct)☆10Nov 25, 2022Updated 3 years ago
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Using Teacher Assistants to Improve Knowledge Distillation: https://arxiv.org/pdf/1902.03393.pdf☆264Oct 3, 2019Updated 6 years ago
- Triplet Loss for Knowledge Distillation☆18Sep 4, 2022Updated 3 years ago
- Implements quantized distillation. Code for our paper "Model compression via distillation and quantization"☆336Jul 25, 2024Updated last year
- ☆13Mar 8, 2020Updated 6 years ago
- Low-rank convolutional neural networks☆97Mar 24, 2016Updated 10 years ago
- A pytorch implement of scalable neural netowrks.☆23Jun 9, 2020Updated 5 years ago
- Arch-Net: Model Distillation for Architecture Agnostic Model Deployment☆23Nov 4, 2021Updated 4 years ago