Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆34Updated 5 months ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- Computation-Efficient Era: A Comprehensive Survey of State Space Models in Medical Image Analysis☆257Updated 3 months ago
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆84Updated 6 months ago
- [ICML 2024] Official implementation for "Predictive Dynamic Fusion."☆66Updated 9 months ago
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆61Updated last year
- Code for the paper "Compositional Entailment Learning for Hyperbolic Vision-Language Models".☆86Updated 4 months ago
- The official code repository of ShaSpec model from CVPR 2023 [paper](https://arxiv.org/pdf/2307.14126) "Multi-modal Learning with Missing…☆77Updated 6 months ago
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆335Updated 2 months ago
- The repo for "Enhancing Multi-modal Cooperation via Sample-level Modality Valuation", CVPR 2024☆57Updated last year
- Pytorch implementation of "Test-time Adaption against Multi-modal Reliability Bias".☆44Updated 10 months ago
- A Comprehensive Survey on Knowledge Distillation☆50Updated last month
- Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models☆136Updated last week
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆68Updated 4 months ago
- [ICML 2023] Provable Dynamic Fusion for Low-Quality Multimodal Data☆111Updated 4 months ago
- The code of "Logits DeConfusion with CLIP for Few-Shot Learning" (CVPR 2025)☆60Updated 5 months ago
- Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model☆21Updated last year
- Official PyTorch Implementation for Active Prompt Learning in Vision Language Models☆37Updated last year
- ☆52Updated 10 months ago
- ☆43Updated last year
- ☆77Updated 8 months ago
- A novel cross-modal decoupling and alignment framework for multimodal representation learning.☆36Updated 7 months ago
- Official implementation of ResCLIP: Residual Attention for Training-free Dense Vision-language Inference☆51Updated 2 weeks ago
- Adaptation of vision-language models (CLIP) to downstream tasks using local and global prompts.☆49Updated 3 months ago
- Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024☆254Updated last month
- An easy way to apply LoRA to CLIP. Implementation of the paper "Low-Rank Few-Shot Adaptation of Vision-Language Models" (CLIP-LoRA) [CVPR…☆262Updated 5 months ago
- [MICCAI 2023][Early Accept] Official code repository of paper titled "Cross-modulated Few-shot Image Generation for Colorectal Tissue Cla…☆47Updated 2 years ago
- The official code for "TextRefiner: Internal Visual Feature as Efficient Refiner for Vision-Language Models Prompt Tuning" | [AAAI2025]☆45Updated 7 months ago
- (CVPR2024 Highlight) Novel Class Discovery for Ultra-Fine-Grained Visual Categorization (UFG-NCD)☆22Updated last year
- Med-DANet Series (ECCV 2022 & WACV 2024)☆13Updated last year
- [CVPR 2024] Code for our Paper "DeiT-LT: Distillation Strikes Back for Vision Transformer training on Long-Tailed Datasets"☆45Updated 10 months ago
- [CVPR 2024] TEA: Test-time Energy Adaptation☆71Updated last year