zju-SWJ / RLD
Official implementation for "Knowledge Distillation with Refined Logits".
☆13Updated 5 months ago
Alternatives and similar repositories for RLD:
Users that are interested in RLD are comparing it to the libraries listed below
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆19Updated last year
- (ICLR 2025) BinaryDM: Accurate Weight Binarization for Efficient Diffusion Models☆17Updated 4 months ago
- ☆25Updated last year
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆57Updated 4 months ago
- The official implementation of LumiNet: The Bright Side of Perceptual Knowledge Distillation https://arxiv.org/abs/2310.03669☆19Updated 11 months ago
- BESA is a differentiable weight pruning technique for large language models.☆14Updated 11 months ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- Official implementation of NeurIPS 2024 "Visual Fourier Prompt Tuning"☆27Updated last month
- CVPR 2023, Class Attention Transfer Based Knowledge Distillation☆37Updated last year
- ☆12Updated last year
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning☆41Updated 6 months ago
- [BMVC 2022] Information Theoretic Representation Distillation☆18Updated last year
- ☆42Updated last year
- [NeurIPS 2024] Search for Efficient LLMs☆12Updated last month
- ☆26Updated 2 years ago
- The offical implementation of [NeurIPS2024] Wasserstein Distance Rivals Kullback-Leibler Divergence for Knowledge Distillation https://ar…☆27Updated 2 months ago
- Official code for Scale Decoupled Distillation☆38Updated 10 months ago
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆28Updated 11 months ago
- The official project website of "Small Scale Data-Free Knowledge Distillation" (SSD-KD for short, published in CVPR 2024).☆15Updated 8 months ago
- [NeurIPS2023]Lightweight Vision Transformer with Bidirectional Interaction☆23Updated last year
- [ICCV 2023 oral] This is the official repository for our paper: ''Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning''.☆65Updated last year
- [NeurIPS'22] Projector Ensemble Feature Distillation☆29Updated last year
- 🔥MixPro: Data Augmentation with MaskMix and Progressive Attention Labeling for Vision Transformer [Official, ICLR 2023]☆21Updated last year
- [CVPR 2024] VkD : Improving Knowledge Distillation using Orthogonal Projections☆50Updated 4 months ago
- Code for "ECoFLaP: Efficient Coarse-to-Fine Layer-Wise Pruning for Vision-Language Models" (ICLR 2024)☆18Updated last year
- Official PyTorch implementation of Which Tokens to Use? Investigating Token Reduction in Vision Transformers presented at ICCV 2023 NIVT …☆33Updated last year
- [ACL 2023] PuMer: Pruning and Merging Tokens for Efficient Vision Language Models☆29Updated 4 months ago
- [CVPR '24] Official implementation of the paper "Multiflow: Shifting Towards Task-Agnostic Vision-Language Pruning".☆18Updated 2 months ago
- GIFT: Generative Interpretable Fine-Tuning☆20Updated 4 months ago