ZjjConan / VLM-MultiModalAdapterLinks
The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".
☆68Updated 2 months ago
Alternatives and similar repositories for VLM-MultiModalAdapter
Users that are interested in VLM-MultiModalAdapter are comparing it to the libraries listed below
Sorting:
- [ECCV 2024] Mind the Interference: Retaining Pre-trained Knowledge in Parameter Efficient Continual Learning of Vision-Language Models☆49Updated 11 months ago
- [CVPR 2024] Official Repository for "Efficient Test-Time Adaptation of Vision-Language Models"☆98Updated 11 months ago
- CVPR2024: Dual Memory Networks: A Versatile Adaptation Approach for Vision-Language Models☆73Updated 11 months ago
- Official code for ICCV 2023 paper, "Improving Zero-Shot Generalization for CLIP with Synthesized Prompts"☆101Updated last year
- cliptrase☆35Updated 9 months ago
- [AAAI'25, CVPRW 2024] Official repository of paper titled "Learning to Prompt with Text Only Supervision for Vision-Language Models".☆108Updated 6 months ago
- Easy wrapper for inserting LoRA layers in CLIP.☆33Updated last year
- [ICLR'24] Consistency-guided Prompt Learning for Vision-Language Models☆75Updated last year
- [NeurIPS 2023] Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization☆105Updated last year
- [ICLR 2024] Test-Time Adaptation with CLIP Reward for Zero-Shot Generalization in Vision-Language Models.☆80Updated 10 months ago
- ☆47Updated last year
- [NeurIPS 2024] Code for Dual Prototype Evolving for Test-Time Generalization of Vision-Language Models☆41Updated 3 months ago
- Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models (AAAI 2024)☆74Updated 4 months ago
- ☆96Updated last year
- ☆42Updated 4 months ago
- The official code for "TextRefiner: Internal Visual Feature as Efficient Refiner for Vision-Language Models Prompt Tuning" | [AAAI2025]☆41Updated 3 months ago
- ☆24Updated last year
- ☆21Updated last year
- [ICLR 2025] See What You Are Told: Visual Attention Sink in Large Multimodal Models☆30Updated 4 months ago
- [NeurIPS2023] LoCoOp: Few-Shot Out-of-Distribution Detection via Prompt Learning☆96Updated 10 months ago
- [ICLR 2025] - Cross the Gap: Exposing the Intra-modal Misalignment in CLIP via Modality Inversion☆47Updated 2 months ago
- [CVPR 2024] Offical implemention of the paper "DePT: Decoupled Prompt Tuning"☆104Updated 3 weeks ago
- [CVPR'24] Validation-free few-shot adaptation of CLIP, using a well-initialized Linear Probe (ZSLP) and class-adaptive constraints (CLAP)…☆74Updated 2 weeks ago
- [ICLR2023] PLOT: Prompt Learning with Optimal Transport for Vision-Language Models☆167Updated last year
- Official implementation of ResCLIP: Residual Attention for Training-free Dense Vision-language Inference☆37Updated 3 months ago
- ☆18Updated 8 months ago
- Official Pytorch implementation of "E2VPT: An Effective and Efficient Approach for Visual Prompt Tuning". (ICCV2023)☆69Updated last year
- FineCLIP: Self-distilled Region-based CLIP for Better Fine-grained Understanding (NIPS24)☆23Updated 6 months ago
- 【ICCV 2023】Diverse Data Augmentation with Diffusions for Effective Test-time Prompt Tuning & 【IJCV 2025】Diffusion-Enhanced Test-time Adap…☆64Updated 5 months ago
- ☆36Updated 2 weeks ago