zju-SWJ / RLDLinks
Official implementation for "Knowledge Distillation with Refined Logits".
☆14Updated last year
Alternatives and similar repositories for RLD
Users that are interested in RLD are comparing it to the libraries listed below
Sorting:
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆68Updated 11 months ago
- ☆26Updated last year
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆60Updated last month
- [BMVC 2022] Information Theoretic Representation Distillation☆18Updated last year
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- ☆27Updated 2 years ago
- [ICML2024] DetKDS: Knowledge Distillation Search for Object Detectors☆15Updated last year
- Training ImageNet / CIFAR models with sota strategies and fancy techniques such as ViT, KD, Rep, etc.☆84Updated last year
- [NeurIPS'22] Projector Ensemble Feature Distillation☆29Updated last year
- BESA is a differentiable weight pruning technique for large language models.☆17Updated last year
- ☆47Updated 2 years ago
- This repo contains the source code for VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks (NeurIPS 2024).☆40Updated 10 months ago
- CVPR 2023, Class Attention Transfer Based Knowledge Distillation☆45Updated 2 years ago
- PELA: Learning Parameter-Efficient Models with Low-Rank Approximation [CVPR 2024]☆18Updated last year
- [2025] Efficient Vision Language Models: A Survey☆26Updated last month
- The official (TMLR) implementation of LumiNet: Perception-Driven Knowledge Distillation via Statistical Logit Calibration☆18Updated 2 weeks ago
- The offical implementation of [NeurIPS2024] Wasserstein Distance Rivals Kullback-Leibler Divergence for Knowledge Distillation https://ar…☆42Updated 8 months ago
- This is the official code for paper: Token Summarisation for Efficient Vision Transformers via Graph-based Token Propagation☆30Updated last year
- Official code for Scale Decoupled Distillation☆41Updated last year
- [AAAI 2023] Official PyTorch Code for "Curriculum Temperature for Knowledge Distillation"☆179Updated 8 months ago
- PyTorch code and checkpoints release for VanillaKD: https://arxiv.org/abs/2305.15781☆75Updated last year
- You Only Condense Once: Two Rules for Pruning Condensed Datasets (NeurIPS 2023)☆15Updated last year
- [CVPR-2022] Official implementation for "Knowledge Distillation with the Reused Teacher Classifier".☆100Updated 3 years ago
- Official pytorch implementation of NeurIPS 2022 paper, TokenMixup☆48Updated 2 years ago
- [TPAMI-2023] Official implementations of L-MCL: Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition☆26Updated 2 years ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆25Updated 9 months ago
- 🔥 🔥 [WACV2024] Mini but Mighty: Finetuning ViTs with Mini Adapters☆20Updated last year
- Official repository of our work "Finding Lottery Tickets in Vision Models via Data-driven Spectral Foresight Pruning" accepted at CVPR 20…☆24Updated 6 months ago