sanagno / neurips_2022_statistics
☆17Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for neurips_2022_statistics
- Official PyTorch implementation of "Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets" (ICLR 2023 notable top 25%)☆22Updated 8 months ago
- Mixture of Attention Heads☆39Updated 2 years ago
- [CVPR 2022] "The Principle of Diversity: Training Stronger Vision Transformers Calls for Reducing All Levels of Redundancy" by Tianlong C…☆25Updated 2 years ago
- Paper List for In-context Learning 🌷☆20Updated last year
- (CVPR 2022) Automated Progressive Learning for Efficient Training of Vision Transformers☆25Updated 2 years ago
- ☆21Updated last year
- Benchmarking Attention Mechanism in Vision Transformers.☆16Updated 2 years ago
- Pi-NAS: Improving Neural Architecture Search by Reducing Supernet Training Consistency Shift (ICCV 2021)☆19Updated 2 years ago
- Paper and Code for "Curriculum Learning by Optimizing Learning Dynamics" (AISTATS 2021)☆19Updated 3 years ago
- ☆26Updated 2 years ago
- [ICLR 2021] "Long Live the Lottery: The Existence of Winning Tickets in Lifelong Learning" by Tianlong Chen*, Zhenyu Zhang*, Sijia Liu, S…☆23Updated 2 years ago
- [NeurIPS'22] What Makes a "Good" Data Augmentation in Knowledge Distillation -- A Statistical Perspective☆36Updated last year
- ☆12Updated 2 years ago
- [ACL 2023] Code for paper “Tailoring Instructions to Student’s Learning Levels Boosts Knowledge Distillation”(https://arxiv.org/abs/2305.…☆37Updated last year
- A pytorch implementation of the ICCV2021 workshop paper SimDis: Simple Distillation Baselines for Improving Small Self-supervised Models☆14Updated 3 years ago
- Anytime Dense Prediction with Confidence Adaptivity (ICLR 2022)☆49Updated 3 months ago
- Beyond Masking: Demystifying Token-Based Pre-Training for Vision Transformers☆26Updated 2 years ago
- [ICLR2024] (EvALign-ICL Benchmark) Beyond Task Performance: Evaluating and Reducing the Flaws of Large Multimodal Models with In-Context …☆20Updated 8 months ago
- Self-Distillation with weighted ground-truth targets; ResNet and Kernel Ridge Regression☆17Updated 3 years ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- EfficientVLM: Fast and Accurate Vision-Language Models via Knowledge Distillation and Modal-adaptive Pruning (ACL 2023)☆22Updated last year
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆19Updated last year
- Repository containing code for blockwise SSL training☆27Updated last month
- ☆11Updated 8 months ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated last year
- ☆20Updated last year
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last year
- Official implementation for paper "Relational Surrogate Loss Learning", ICLR 2022☆37Updated 2 years ago
- ICML'20: SIGUA: Forgetting May Make Learning with Noisy Labels More Robust☆13Updated 3 years ago
- ☆13Updated 2 years ago