Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆35Updated 7 months ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆64Updated last year
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆95Updated 8 months ago
- The official code repository of ShaSpec model from CVPR 2023 [paper](https://arxiv.org/pdf/2307.14126) "Multi-modal Learning with Missing…☆86Updated 8 months ago
- ☆64Updated 3 months ago
- [ICML 2024] Official implementation for "Predictive Dynamic Fusion."☆69Updated last year
- Pytorch implementation of "Test-time Adaption against Multi-modal Reliability Bias".☆44Updated last year
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆69Updated 6 months ago
- [CVPR 2024] Code for our Paper "DeiT-LT: Distillation Strikes Back for Vision Transformer training on Long-Tailed Datasets"☆47Updated last year
- [CVPR 2024] TEA: Test-time Energy Adaptation☆71Updated last year
- Adaptation of vision-language models (CLIP) to downstream tasks using local and global prompts.☆50Updated 6 months ago
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆344Updated 3 weeks ago
- Code for the paper "Compositional Entailment Learning for Hyperbolic Vision-Language Models".☆96Updated 6 months ago
- ☆46Updated last year
- Masked Autoencoder meets GANs☆29Updated 2 years ago
- ☆79Updated 10 months ago
- Computation-Efficient Era: A Comprehensive Survey of State Space Models in Medical Image Analysis☆262Updated 5 months ago
- [TPAMI 2025] Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models☆156Updated last week
- [ICML 2023] Provable Dynamic Fusion for Low-Quality Multimodal Data☆115Updated 6 months ago
- [MICCAI 2023][Early Accept] Official code repository of paper titled "Cross-modulated Few-shot Image Generation for Colorectal Tissue Cla…☆47Updated 2 years ago
- ☆54Updated last year
- [NeurIPS 2024 Spotlight] Code for the paper "Flex-MoE: Modeling Arbitrary Modality Combination via the Flexible Mixture-of-Experts"☆69Updated 7 months ago
- The repo for "Enhancing Multi-modal Cooperation via Sample-level Modality Valuation", CVPR 2024☆59Updated last year
- [ECCV 2024] Soft Prompt Generation for Domain Generalization☆28Updated last year
- (CVPR2024 Highlight) Novel Class Discovery for Ultra-Fine-Grained Visual Categorization (UFG-NCD)☆23Updated last year
- [ICLR 2025] - Cross the Gap: Exposing the Intra-modal Misalignment in CLIP via Modality Inversion☆58Updated last month
- [CVPR 2024] FairCLIP: Harnessing Fairness in Vision-Language Learning☆96Updated 5 months ago
- Code and Dataset for the paper "LAMM: Label Alignment for Multi-Modal Prompt Learning" AAAI 2024☆33Updated 2 years ago
- Low-Rank Rescaled Vision Transformer Fine-Tuning: A Residual Design Approach, CVPR 2024☆22Updated last year
- ☆28Updated last year
- [ICLR 2025] Multi-modal representation learning of shared, unique and synergistic features between modalities☆55Updated 8 months ago