zju-SWJ / RLD
Official implementation for "Knowledge Distillation with Refined Logits".
☆13Updated 6 months ago
Alternatives and similar repositories for RLD:
Users that are interested in RLD are comparing it to the libraries listed below
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆60Updated 6 months ago
- This is the official code for paper: Token Summarisation for Efficient Vision Transformers via Graph-based Token Propagation☆27Updated last year
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆44Updated 2 weeks ago
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆21Updated 4 months ago
- [NeurIPS'22] Projector Ensemble Feature Distillation☆29Updated last year
- This repo contains the source code for VB-LoRA: Extreme Parameter Efficient Fine-Tuning with Vector Banks (NeurIPS 2024).☆37Updated 5 months ago
- [ICCV 2023 oral] This is the official repository for our paper: ''Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning''.☆66Updated last year
- The official implementation of LumiNet: The Bright Side of Perceptual Knowledge Distillation https://arxiv.org/abs/2310.03669☆19Updated last year
- ☆26Updated 2 years ago
- [ACL 2023] PuMer: Pruning and Merging Tokens for Efficient Vision Language Models☆29Updated 5 months ago
- ☆25Updated last year
- The official project website of "Small Scale Data-Free Knowledge Distillation" (SSD-KD for short, published in CVPR 2024).☆17Updated 9 months ago
- Official implementation of NeurIPS 2024 "Visual Fourier Prompt Tuning"☆27Updated 2 months ago
- BESA is a differentiable weight pruning technique for large language models.☆14Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆29Updated last year
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- [BMVC 2022] Information Theoretic Representation Distillation☆18Updated last year
- [CVPR 2024] VkD : Improving Knowledge Distillation using Orthogonal Projections☆50Updated 5 months ago
- (ICLR 2025) BinaryDM: Accurate Weight Binarization for Efficient Diffusion Models☆18Updated 5 months ago
- ☆43Updated last year
- CVPR 2023, Class Attention Transfer Based Knowledge Distillation☆41Updated last year
- [IJCV2025] https://arxiv.org/abs/2304.04521☆11Updated 2 months ago
- ☆12Updated last year
- 🔥 🔥 [WACV2024] Mini but Mighty: Finetuning ViTs with Mini Adapters☆18Updated 8 months ago
- Official code for Scale Decoupled Distillation☆40Updated 11 months ago
- Structured Pruning Adapters in PyTorch☆16Updated last year
- Official implementation of "SViT: Revisiting Token Pruning for Object Detection and Instance Segmentation"☆29Updated last year
- 🔥MixPro: Data Augmentation with MaskMix and Progressive Attention Labeling for Vision Transformer [Official, ICLR 2023]☆22Updated last year