Henrymachiyu / ProtoViTLinks
This code implements ProtoViT, a novel approach that combines Vision Transformers with prototype-based learning to create interpretable image classification models. Our implementation provides both high accuracy and explainability through learned prototypes.
☆22Updated 2 weeks ago
Alternatives and similar repositories for ProtoViT
Users that are interested in ProtoViT are comparing it to the libraries listed below
Sorting:
- [CVPR 2024] PriViLege: Pre-trained Vision and Language Transformers Are Few-Shot Incremental Learners☆52Updated 9 months ago
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆69Updated last month
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆50Updated last year
- [NeurIPS2023] LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning☆96Updated 10 months ago
- The official implementation of CVPR 24' Paper "Learning Transferable Negative Prompts for Out-of-Distribution Detection"☆55Updated last year
- [ICLR 2025] Official Implementation of Local-Prompt: Extensible Local Prompts for Few-Shot Out-of-Distribution Detection☆30Updated last month
- ☆41Updated 3 weeks ago
- Official code for ICCV 2023 paper, "Improving Zero-Shot Generalization for CLIP with Synthesized Prompts"☆101Updated last year
- (CVPR2024 Highlight) Novel Class Discovery for Ultra-Fine-Grained Visual Categorization (UFG-NCD)☆18Updated 11 months ago
- ☆36Updated 7 months ago
- [NeurIPS 2024 Spotlight] Code for the paper "Flex-MoE: Modeling Arbitrary Modality Combination via the Flexible Mixture-of-Experts"☆52Updated 7 months ago
- Official code for ICLR 2024 paper, "A Hard-to-Beat Baseline for Training-free CLIP-based Adaptation"☆81Updated last year
- PyTorch Implementation for InMaP☆11Updated last year
- Adaptation of vision-language models (CLIP) to downstream tasks using local and global prompts.☆44Updated 8 months ago
- Everything to the Synthetic: Diffusion-driven Test-time Adaptation via Synthetic-Domain Alignment, arXiv 2024 / CVPR 2025☆29Updated 3 months ago
- The official implementation for MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning (CVPR '24)☆49Updated 2 months ago
- ☆17Updated 7 months ago
- Exploring prompt tuning with pseudolabels for multiple modalities, learning settings, and training strategies.☆50Updated 6 months ago
- [ICLR 2024 Spotlight] "Negative Label Guided OOD Detection with Pretrained Vision-Language Models"☆21Updated 7 months ago
- PyTorch implementation of our CVPR 2024 paper "Unified Entropy Optimization for Open-Set Test-Time Adaptation"☆23Updated 9 months ago
- [COLING'25] HGCLIP: Exploring Vision-Language Models with Graph Representations for Hierarchical Understanding☆40Updated 6 months ago
- [CVPR'24] Validation-free few-shot adaptation of CLIP, using a well-initialized Linear Probe (ZSLP) and class-adaptive constraints (CLAP)…☆72Updated last month
- CVPR 2023: Language in a Bottle: Language Model Guided Concept Bottlenecks for Interpretable Image Classification☆91Updated last year
- Low-Rank Rescaled Vision Transformer Fine-Tuning: A Residual Design Approach, CVPR 2024☆22Updated 10 months ago
- [ECCV 2024] Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models☆48Updated 10 months ago
- [CVPR 2024] Code for our Paper "DeiT-LT: Distillation Strikes Back for Vision Transformer training on Long-Tailed Datasets"☆42Updated 5 months ago
- The PyTorch implementation for "DEAL: Disentangle and Localize Concept-level Explanations for VLMs" (ECCV 2024 Strong Double Blind)☆20Updated 7 months ago
- ☆21Updated last year
- [CVPR-25🔥] Test-time Counterattacks (TTC) towards adversarial robustness of CLIP☆25Updated this week
- Test-time adaptation via Nearest neighbor information (TAST), submitted to ICLR'23☆22Updated last year