MingSun-Tse / Good-DA-in-KD
[NeurIPS'22] What Makes a "Good" Data Augmentation in Knowledge Distillation -- A Statistical Perspective
☆36Updated 2 years ago
Alternatives and similar repositories for Good-DA-in-KD:
Users that are interested in Good-DA-in-KD are comparing it to the libraries listed below
- ☆19Updated 2 years ago
- [ICLR'23] Trainability Preserving Neural Pruning (PyTorch)☆32Updated last year
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated last year
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 2 years ago
- A generic code base for neural network pruning, especially for pruning at initialization.☆30Updated 2 years ago
- Official pytorch implementation for CVPR2022 paper "Bootstrapping ViTs: Towards Liberating Vision Transformers from Pre-training"☆17Updated 2 years ago
- [IJCAI-2021] Contrastive Model Inversion for Data-Free Knowledge Distillation☆69Updated 2 years ago
- This repo is the official megengine implementation of the ECCV2022 paper: Efficient One Pass Self-distillation with Zipf's Label Smoothin…☆26Updated 2 years ago
- ☆27Updated 2 years ago
- PyTorch implementation of paper "Dataset Distillation via Factorization" in NeurIPS 2022.☆63Updated 2 years ago
- Repo for the paper "Extrapolating from a Single Image to a Thousand Classes using Distillation"☆36Updated 7 months ago
- Benchmarking Attention Mechanism in Vision Transformers.☆17Updated 2 years ago
- [ICLR 2022]: Fast AdvProp☆34Updated 2 years ago
- Official Codes and Pretrained Models for RecursiveMix☆22Updated last year
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Updated last year
- Official PyTorch Code for "Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation?" (https://arxiv.org/abs/2305.12954)☆45Updated last year
- ☆16Updated 2 years ago
- The code of the paper "Minimizing the Accumulated Trajectory Error to Improve Dataset Distillation" (CVPR2023)☆40Updated last year
- Code for ViTAS_Vision Transformer Architecture Search☆52Updated 3 years ago
- ☆57Updated 3 years ago
- [NeurIPS 2022] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach -- Official Implementation☆44Updated last year
- [NeurIPS-2021] Mosaicking to Distill: Knowledge Distillation from Out-of-Domain Data☆44Updated 2 years ago
- [NeurIPS 2024] Search for Efficient LLMs☆12Updated 3 weeks ago
- Metrics for "Beyond neural scaling laws: beating power law scaling via data pruning " (NeurIPS 2022 Outstanding Paper Award)☆55Updated last year
- [ICML2023] Revisiting Data-Free Knowledge Distillation with Poisoned Teachers☆23Updated 7 months ago
- Bag of Instances Aggregation Boosts Self-supervised Distillation (ICLR 2022)☆33Updated 2 years ago
- Code for ECCV 2022 paper “Learning with Recoverable Forgetting”☆20Updated 2 years ago
- ReSSL: Relational Self-Supervised Learning with Weak Augmentation☆57Updated 3 years ago
- [ICLR 2022] The Unreasonable Effectiveness of Random Pruning: Return of the Most Naive Baseline for Sparse Training by Shiwei Liu, Tianlo…☆73Updated 2 years ago
- ☆30Updated 4 years ago