vorobeevich / distillation-in-dg
Implementation of "Weight Averaging Improves Knowledge Distillation under Domain Shift" (ICCV 2023 OOD-CV Workshop)
☆19Updated last year
Alternatives and similar repositories for distillation-in-dg:
Users that are interested in distillation-in-dg are comparing it to the libraries listed below
- Moved to https://github.com/NUS-HPC-AI-Lab/InfoBatch☆6Updated last year
- Official Pytorch Implementation of Self-emerging Token Labeling☆32Updated 10 months ago
- Official implementation for "Knowledge Distillation with Refined Logits".☆13Updated 5 months ago
- Official PyTorch implementation of "Meta-prediction Model for Distillation-Aware NAS on Unseen Datasets" (ICLR 2023 notable top 25%)☆23Updated 11 months ago
- The official implementation of LumiNet: The Bright Side of Perceptual Knowledge Distillation https://arxiv.org/abs/2310.03669☆19Updated 11 months ago
- This is a PyTorch implementation of the paperViP A Differentially Private Foundation Model for Computer Vision☆36Updated last year
- [ICLR24] AutoVP: An Automated Visual Prompting Framework and Benchmark☆17Updated last year
- [WACV2023] This is the official PyTorch impelementation of our paper "[Rethinking Rotation in Self-Supervised Contrastive Learning: Adapt…☆12Updated last year
- [ICML 2024] "Envisioning Outlier Exposure by Large Language Models for Out-of-Distribution Detection"☆12Updated last week
- [ICLR 2024] This is the official PyTorch implementation of "QLLM: Accurate and Efficient Low-Bitwidth Quantization for Large Language Mod…☆25Updated 11 months ago
- Code for "Don't trust your eyes: on the (un)reliability of feature visualizations"☆31Updated last year
- [ACL 2023] PuMer: Pruning and Merging Tokens for Efficient Vision Language Models☆29Updated 4 months ago
- (ICLR 2025) BinaryDM: Accurate Weight Binarization for Efficient Diffusion Models☆17Updated 4 months ago
- ☆37Updated last year
- This repository compiles a list of papers related to Video LLM.☆19Updated 7 months ago
- [ICLR 2024] "Data Distillation Can Be Like Vodka: Distilling More Times For Better Quality" by Xuxi Chen*, Yu Yang*, Zhangyang Wang, Baha…☆11Updated 9 months ago
- [NeurIPS 2024] The official implementation of "Image Copy Detection for Diffusion Models"☆16Updated 4 months ago
- 🔥MixPro: Data Augmentation with MaskMix and Progressive Attention Labeling for Vision Transformer [Official, ICLR 2023]☆21Updated last year
- Official PyTorch Code for "Is Synthetic Data From Diffusion Models Ready for Knowledge Distillation?" (https://arxiv.org/abs/2305.12954)☆45Updated last year
- [NeurIPS'24] "NeuralFuse: Learning to Recover the Accuracy of Access-Limited Neural Network Inference in Low-Voltage Regimes"☆10Updated last year
- survery of small language models☆14Updated 7 months ago
- ☆32Updated last month
- Data-free knowledge distillation using Gaussian noise (NeurIPS paper)☆15Updated last year
- [ICCV 2023] DataDAM: Efficient Dataset Distillation with Attention Matching☆32Updated 8 months ago
- [ICML2023] Revisiting Data-Free Knowledge Distillation with Poisoned Teachers☆23Updated 7 months ago
- This repository is the implementation of the paper Training Free Pretrained Model Merging (CVPR2024).☆27Updated 11 months ago
- [CVPR 2024] DiffAgent: Fast and Accurate Text-to-Image API Selection with Large Language Model☆17Updated 10 months ago
- Official code for the paper "Examining Post-Training Quantization for Mixture-of-Experts: A Benchmark"☆10Updated 7 months ago
- BESA is a differentiable weight pruning technique for large language models.☆14Updated 11 months ago