zju-SWJ / RLDLinks
Official implementation for "Knowledge Distillation with Refined Logits".
☆14Updated 9 months ago
Alternatives and similar repositories for RLD
Users that are interested in RLD are comparing it to the libraries listed below
Sorting:
- ☆26Updated last year
- Code for 'Multi-level Logit Distillation' (CVPR2023)☆64Updated 8 months ago
- CVPR 2023, Class Attention Transfer Based Knowledge Distillation☆44Updated last year
- The codebase for paper "PPT: Token Pruning and Pooling for Efficient Vision Transformer"☆23Updated 6 months ago
- Official code for Scale Decoupled Distillation☆41Updated last year
- The official project website of "NORM: Knowledge Distillation via N-to-One Representation Matching" (The paper of NORM is published in IC…☆20Updated last year
- [NeurIPS 2024] Search for Efficient LLMs☆14Updated 4 months ago
- The official implementation of LumiNet: The Bright Side of Perceptual Knowledge Distillation https://arxiv.org/abs/2310.03669☆19Updated last year
- Learning Efficient Vision Transformers via Fine-Grained Manifold Distillation. NeurIPS 2022.☆32Updated 2 years ago
- The official project website of "Small Scale Data-Free Knowledge Distillation" (SSD-KD for short, published in CVPR 2024).☆17Updated 11 months ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆49Updated 2 months ago
- ☆27Updated 2 years ago
- [Preprint] Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Prunin…☆40Updated 2 years ago
- This is the official code for paper: Token Summarisation for Efficient Vision Transformers via Graph-based Token Propagation☆28Updated last year
- [CVPR 2023] Castling-ViT: Compressing Self-Attention via Switching Towards Linear-Angular Attention During Vision Transformer Inference☆30Updated last year
- [ECCV 2024] Isomorphic Pruning for Vision Models☆68Updated 10 months ago
- BESA is a differentiable weight pruning technique for large language models.☆16Updated last year
- [BMVC 2022] Information Theoretic Representation Distillation☆18Updated last year
- ☆19Updated 3 years ago
- Official implementation of NeurIPS 2024 "Visual Fourier Prompt Tuning"☆28Updated 4 months ago
- [CVPR 2024] VkD : Improving Knowledge Distillation using Orthogonal Projections☆53Updated 7 months ago
- Pytorch implementation of our paper accepted by IEEE TNNLS, 2022 — Carrying out CNN Channel Pruning in a White Box☆18Updated 3 years ago
- [ACM MM'23] Official implementation of paper "Avatar Knowledge Distillation: Self-ensemble Teacher Paradigm with Uncertainty".☆13Updated last year
- a training-free approach to accelerate ViTs and VLMs by pruning redundant tokens based on similarity☆24Updated last week
- [NeurIPS2023]Lightweight Vision Transformer with Bidirectional Interaction☆24Updated last year
- A token pruning method that accelerates ViTs for various tasks while maintaining high performance.☆12Updated 4 months ago
- ☆12Updated last year
- PyTorch code and checkpoints release for VanillaKD: https://arxiv.org/abs/2305.15781☆75Updated last year
- [NeurIPS'22] Projector Ensemble Feature Distillation☆29Updated last year
- Low-Rank Rescaled Vision Transformer Fine-Tuning: A Residual Design Approach, CVPR 2024☆22Updated 10 months ago