Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆25Updated 2 months ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆76Updated 3 months ago
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆55Updated last year
- ☆49Updated 3 months ago
- [CVPR-2024] Official implementations of CLIP-KD: An Empirical Study of CLIP Model Distillation☆123Updated last year
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆322Updated 2 weeks ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆58Updated last month
- Low-Rank Rescaled Vision Transformer Fine-Tuning: A Residual Design Approach, CVPR 2024☆22Updated last year
- [ICLR 2025] - Cross the Gap: Exposing the Intra-modal Misalignment in CLIP via Modality Inversion☆51Updated 3 months ago
- ☆41Updated 9 months ago
- [ICCV 2023 & AAAI 2023] Binary Adapters & FacT, [Tech report] Convpass☆192Updated 2 years ago
- [ECCV 2024] Soft Prompt Generation for Domain Generalization☆26Updated 10 months ago
- [ICCV'23 Main Track, WECIA'23 Oral] Official repository of paper titled "Self-regulating Prompts: Foundational Model Adaptation without F…☆272Updated last year
- [AAAI'25, CVPRW 2024] Official repository of paper titled "Learning to Prompt with Text Only Supervision for Vision-Language Models".☆111Updated 8 months ago
- Pytorch implementation of "Test-time Adaption against Multi-modal Reliability Bias".☆37Updated 7 months ago
- [CVPR 2024] FairCLIP: Harnessing Fairness in Vision-Language Learning☆86Updated 3 weeks ago
- ☆18Updated 3 months ago
- Twin Contrastive Learning with Noisy Labels (CVPR 2023)☆70Updated 2 years ago
- Official PyTorch implementation of DiffuseMix : Label-Preserving Data Augmentation with Diffusion Models (CVPR'2024)☆120Updated 5 months ago
- ☆45Updated 5 months ago
- [Survey] Awesome List of Mixup Augmentation and Beyond (https://arxiv.org/abs/2409.05202)☆153Updated 10 months ago
- Awesome List of Vision Language Prompt Papers☆46Updated last year
- [ECCV 2024] Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models☆52Updated last year
- [NeurIPS2023] LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning☆97Updated last month
- [CVPR 2024] On the Diversity and Realism of Distilled Dataset: An Efficient Dataset Distillation Paradigm☆73Updated 5 months ago
- ☆24Updated last year
- Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models☆114Updated this week
- Source code for the paper "Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts" (ICML 2024)☆87Updated 9 months ago
- [CVPR 2024] Code for our Paper "DeiT-LT: Distillation Strikes Back for Vision Transformer training on Long-Tailed Datasets"☆41Updated 7 months ago
- Official code for ICLR 2024 paper, "A Hard-to-Beat Baseline for Training-free CLIP-based Adaptation"☆81Updated last year
- Official Repository for ICML 2024 Paper "OT-CLIP: Understanding and Generalizing CLIP via Optimal Transport"☆16Updated last year