JiuTian-VL / MoME
☆25Updated last month
Related projects ⓘ
Alternatives and complementary repositories for MoME
- [ICCV 2023 oral] This is the official repository for our paper: ''Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning''.☆64Updated last year
- [ECCV 2024] API: Attention Prompting on Image for Large Vision-Language Models☆48Updated last month
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Prompt…☆32Updated 4 months ago
- Task Residual for Tuning Vision-Language Models (CVPR 2023)☆66Updated last year
- [ICML 2024] Memory-Space Visual Prompting for Efficient Vision-Language Fine-Tuning☆44Updated 6 months ago
- Official code for ICCV 2023 paper, "Improving Zero-Shot Generalization for CLIP with Synthesized Prompts"☆94Updated 8 months ago
- Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning☆17Updated 2 months ago
- This is the official repo for Debiasing Large Visual Language Models, including a Post-Hoc debias method and Visual Debias Decoding strat…☆72Updated 7 months ago
- Distribution-Aware Prompt Tuning for Vision-Language Models (ICCV 2023)☆37Updated 11 months ago
- Domain Generalization through Distilling CLIP with Language Guidance☆24Updated last year
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- ☆32Updated 7 months ago
- This repository houses the code for the paper - "The Neglected of VLMs"☆23Updated 3 months ago
- Adapting LLaMA Decoder to Vision Transformer☆27Updated 6 months ago
- [CVPR'24] Validation-free few-shot adaptation of CLIP, using a well-initialized Linear Probe (ZSLP) and class-adaptive constraints (CLAP)…☆59Updated 5 months ago
- Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆49Updated 3 months ago
- Towards Unified and Effective Domain Generalization☆29Updated 11 months ago
- PyTorch code for "Contrastive Region Guidance: Improving Grounding in Vision-Language Models without Training"☆28Updated 8 months ago
- 【ICCV 2023】Diverse Data Augmentation with Diffusions for Effective Test-time Prompt Tuning☆56Updated 3 months ago
- Instruction Tuning in Continual Learning paradigm☆26Updated 4 months ago
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆40Updated 3 months ago
- ☆21Updated 5 months ago
- [CVPR2024] The code of "UniPT: Universal Parallel Tuning for Transfer Learning with Efficient Parameter and Memory"☆64Updated last month
- CVPR2024: Dual Memory Networks: A Versatile Adaptation Approach for Vision-Language Models☆60Updated 4 months ago
- Implementation of "VL-Mamba: Exploring State Space Models for Multimodal Learning"☆78Updated 8 months ago
- CLIP-MoE: Mixture of Experts for CLIP☆17Updated last month
- Compress conventional Vision-Language Pre-training data☆49Updated last year
- [ICLR'24] Consistency-guided Prompt Learning for Vision-Language Models☆58Updated 5 months ago
- [NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"☆166Updated 8 months ago
- Official code for "Understanding and Mitigating Overfitting in Prompt Tuning for Vision-Language Models" (TCSVT'2023)☆27Updated 10 months ago