Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆35Updated 7 months ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆94Updated 7 months ago
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆63Updated last year
- Code for the paper "Compositional Entailment Learning for Hyperbolic Vision-Language Models".☆91Updated 6 months ago
- Adaptation of vision-language models (CLIP) to downstream tasks using local and global prompts.☆49Updated 5 months ago
- Pytorch implementation of "Test-time Adaption against Multi-modal Reliability Bias".☆44Updated 11 months ago
- ☆63Updated 2 months ago
- [ICML 2024] Official implementation for "Predictive Dynamic Fusion."☆68Updated 11 months ago
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆345Updated last week
- ☆46Updated last year
- [CVPR 2024] TEA: Test-time Energy Adaptation☆72Updated last year
- Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models☆147Updated last week
- A Comprehensive Survey on Knowledge Distillation☆55Updated last week
- ☆78Updated 9 months ago
- Official implementation of the "Multimodal Parameter-Efficient Few-Shot Class Incremental Learning" paper☆23Updated last year
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆69Updated 5 months ago
- [ICML 2023] Provable Dynamic Fusion for Low-Quality Multimodal Data☆113Updated 5 months ago
- (CVPR2024 Highlight) Novel Class Discovery for Ultra-Fine-Grained Visual Categorization (UFG-NCD)☆23Updated last year
- The code of "Logits DeConfusion with CLIP for Few-Shot Learning" (CVPR 2025)☆64Updated 6 months ago
- Code for paper "Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts Adapters" CVPR2024☆260Updated 3 months ago
- [CVPR 2025] Official implementation of the paper "Show and Tell: Visually Explainable Deep Neural Nets via Spatially-Aware Concept Bottle…☆15Updated 5 months ago
- [NeurIPS2023] LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning☆102Updated 5 months ago
- ☆28Updated last year
- ☆22Updated 7 months ago
- [CVPR 2024] Open-Set Domain Adaptation for Semantic Segmentation☆53Updated last year
- Decoupling common and unique representations for multimodal self-supervised learning☆69Updated last year
- [NeurIPS 2024 Spotlight] Code for the paper "Flex-MoE: Modeling Arbitrary Modality Combination via the Flexible Mixture-of-Experts"☆68Updated 6 months ago
- The repo for "Enhancing Multi-modal Cooperation via Sample-level Modality Valuation", CVPR 2024☆58Updated last year
- The official code for "TextRefiner: Internal Visual Feature as Efficient Refiner for Vision-Language Models Prompt Tuning" | [AAAI2025]☆47Updated 9 months ago
- [ECCV' 24 Oral] CLIFF: Continual Latent Diffusion for Open-Vocabulary Object Detection☆29Updated last year
- ☆22Updated last year