bellymonster / Weighted-Soft-Label-Distillation
☆57Updated 3 years ago
Alternatives and similar repositories for Weighted-Soft-Label-Distillation:
Users that are interested in Weighted-Soft-Label-Distillation are comparing it to the libraries listed below
- Bag of Instances Aggregation Boosts Self-supervised Distillation (ICLR 2022)☆33Updated 2 years ago
- UniMoCo: Unsupervised, Semi-Supervised and Full-Supervised Visual Representation Learning☆55Updated 3 years ago
- Improving Contrastive Learning by Visualizing Feature Transformation, ICCV 2021 Oral☆90Updated 3 years ago
- Knowledge Transfer via Dense Cross-layer Mutual-distillation (ECCV'2020)☆30Updated 4 years ago
- Self-distillation with Batch Knowledge Ensembling Improves ImageNet Classification☆81Updated 3 years ago
- Code for ViTAS_Vision Transformer Architecture Search☆52Updated 3 years ago
- Pytorch implementation of CVPR2021 paper: SuperMix: Supervising the Mixing Data Augmentation☆92Updated 3 years ago
- [ICLR 2021 Spotlight Oral] "Undistillable: Making A Nasty Teacher That CANNOT teach students", Haoyu Ma, Tianlong Chen, Ting-Kuei Hu, Che…☆81Updated 3 years ago
- ReSSL: Relational Self-Supervised Learning with Weak Augmentation☆58Updated 3 years ago
- This repo is the official megengine implementation of the ECCV2022 paper: Efficient One Pass Self-distillation with Zipf's Label Smoothin…☆26Updated 2 years ago
- Un-Mix: Rethinking Image Mixtures for Unsupervised Visual Representation Learning.☆151Updated 2 years ago
- [NeurIPS'21] "Chasing Sparsity in Vision Transformers: An End-to-End Exploration" by Tianlong Chen, Yu Cheng, Zhe Gan, Lu Yuan, Lei Zhang…☆89Updated last year
- [ICLR 2022] "As-ViT: Auto-scaling Vision Transformers without Training" by Wuyang Chen, Wei Huang, Xianzhi Du, Xiaodan Song, Zhangyang Wa…☆76Updated 3 years ago
- ☆36Updated 3 years ago
- Code release for NeurIPS 2020 paper "Co-Tuning for Transfer Learning"☆40Updated 2 years ago
- Beyond Masking: Demystifying Token-Based Pre-Training for Vision Transformers☆26Updated 2 years ago
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 3 weeks ago
- Implementation of the Heterogeneous Knowledge Distillation using Information Flow Modeling method☆24Updated 4 years ago
- Code implementation for paper "On the Efficacy of Small Self-Supervised Contrastive Models without Distillation Signals".☆16Updated 3 years ago
- [NeurIPS'22] What Makes a "Good" Data Augmentation in Knowledge Distillation -- A Statistical Perspective☆36Updated 2 years ago
- [ICLR 2022]: Fast AdvProp☆34Updated 3 years ago
- Official PyTorch implementation of "Co-Mixup: Saliency Guided Joint Mixup with Supermodular Diversity" (ICLR'21 Oral)☆103Updated 3 years ago
- Official implementation of the paper "Function-Consistent Feature Distillation" (ICLR 2023)☆28Updated last year
- [CVPR 2021] "The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models" Tianlong Chen, Jon…☆69Updated 2 years ago
- [TMLR] "Adversarial Feature Augmentation and Normalization for Visual Recognition", Tianlong Chen, Yu Cheng, Zhe Gan, Jianfeng Wang, Liju…☆20Updated 2 years ago
- [ICML 2021] “ Self-Damaging Contrastive Learning”, Ziyu Jiang, Tianlong Chen, Bobak Mortazavi, Zhangyang Wang☆63Updated 3 years ago
- Compressing Representations for Self-Supervised Learning☆78Updated 4 years ago
- ☆42Updated 4 years ago
- Batch Normalization with Enhanced Linear Transformation☆53Updated last year
- Complementary Relation Contrastive Distillation☆15Updated 3 years ago