Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆35Updated 8 months ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆64Updated last year
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆95Updated 9 months ago
- Adaptation of vision-language models (CLIP) to downstream tasks using local and global prompts.☆50Updated 6 months ago
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆347Updated last month
- [MICCAI 2023][Early Accept] Official code repository of paper titled "Cross-modulated Few-shot Image Generation for Colorectal Tissue Cla…☆47Updated 2 years ago
- Code for the paper "Compositional Entailment Learning for Hyperbolic Vision-Language Models".☆98Updated 7 months ago
- [ICML 2024] Official implementation for "Predictive Dynamic Fusion."☆70Updated last year
- [ICML 2023] Provable Dynamic Fusion for Low-Quality Multimodal Data☆116Updated 7 months ago
- ☆46Updated last year
- [TPAMI 2026] Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models☆168Updated this week
- An easy way to apply LoRA to CLIP. Implementation of the paper "Low-Rank Few-Shot Adaptation of Vision-Language Models" (CLIP-LoRA) [CVPR…☆282Updated 7 months ago
- The code of "Logits DeConfusion with CLIP for Few-Shot Learning" (CVPR 2025)☆67Updated 7 months ago
- [AAAI'25, CVPRW 2024] Official repository of paper titled "Learning to Prompt with Text Only Supervision for Vision-Language Models".☆120Updated last year
- The official code repository of ShaSpec model from CVPR 2023 [paper](https://arxiv.org/pdf/2307.14126) "Multi-modal Learning with Missing…☆88Updated 9 months ago
- Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024☆269Updated 4 months ago
- [ICCV'23 Main Track, WECIA'23 Oral] Official repository of paper titled "Self-regulating Prompts: Foundational Model Adaptation without F…☆283Updated 2 years ago
- [CVPR 2025] Official implementation of the paper "Show and Tell: Visually Explainable Deep Neural Nets via Spatially-Aware Concept Bottle…☆16Updated 7 months ago
- [ICLR 2025] - Cross the Gap: Exposing the Intra-modal Misalignment in CLIP via Modality Inversion☆59Updated 2 months ago
- Pytorch implementation of "Test-time Adaption against Multi-modal Reliability Bias".☆44Updated last year
- ☆64Updated 3 months ago
- [ECCV 2024] Soft Prompt Generation for Domain Generalization☆30Updated last year
- ☆28Updated last year
- A Comprehensive Survey on Knowledge Distillation☆60Updated last month
- [ICLR 2025] Multi-modal representation learning of shared, unique and synergistic features between modalities☆57Updated 8 months ago
- Official implementation of the "Multimodal Parameter-Efficient Few-Shot Class Incremental Learning" paper☆24Updated last year
- [CVPR 2024] TEA: Test-time Energy Adaptation☆71Updated last year
- [ICLR'24] Consistency-guided Prompt Learning for Vision-Language Models☆85Updated last year
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆69Updated 6 months ago
- [CVPR 2024] Leveraging Vision-Language Models for Improving Domain Generalization in Image Classification☆38Updated last year
- The repo for "Enhancing Multi-modal Cooperation via Sample-level Modality Valuation", CVPR 2024☆59Updated last year