ssfgunner / IISLinks
[ICLR 2025 Spotlight] This is the official repository for our paper: ''Enhancing Pre-trained Representation Classifiability can Boost its Interpretability''.
☆25Updated 8 months ago
Alternatives and similar repositories for IIS
Users that are interested in IIS are comparing it to the libraries listed below
Sorting:
- A curated list of awesome prompt/adapter learning methods for vision-language models like CLIP.☆737Updated last month
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆344Updated 3 weeks ago
- ☆79Updated 2 months ago
- This repo lists relevant papers summarized in our survey paper: A Systematic Survey of Prompt Engineering on Vision-Language Foundation …☆505Updated 9 months ago
- [CVPR 2024 Highlight] Mitigating Object Hallucinations in Large Vision-Language Models through Visual Contrastive Decoding☆365Updated last year
- [WACV 2025] Code for Enhancing Vision-Language Few-Shot Adaptation with Negative Learning☆11Updated 10 months ago
- [CVPR 2023] Official repository of paper titled "MaPLe: Multi-modal Prompt Learning".☆794Updated 2 years ago
- This is the first released survey paper on hallucinations of large vision-language models (LVLMs). To keep track of this field and contin…☆90Updated last year
- [ICCV'23 Main Track, WECIA'23 Oral] Official repository of paper titled "Self-regulating Prompts: Foundational Model Adaptation without F…☆281Updated 2 years ago
- up-to-date curated list of state-of-the-art Large vision language models hallucinations research work, papers & resources☆241Updated 3 months ago
- Collection of Composed Image Retrieval (CIR) papers.☆291Updated 2 weeks ago
- ☆55Updated last year
- [ICLR'25] Official code for the paper 'MLLMs Know Where to Look: Training-free Perception of Small Visual Details with Multimodal LLMs'☆311Updated 8 months ago
- FineCLIP: Self-distilled Region-based CLIP for Better Fine-grained Understanding (NIPS24)☆33Updated 2 months ago
- Awesome papers & datasets specifically focused on long-term videos.☆338Updated 3 months ago
- ☆18Updated last year
- Visualizing the attention of vision-language models☆272Updated 10 months ago
- ☆154Updated 10 months ago
- [CVPR25 Highlight] A ChatGPT-Prompted Visual hallucination Evaluation Dataset, featuring over 100,000 data samples and four advanced eval…☆30Updated 8 months ago
- Papers about Hallucination in Multi-Modal Large Language Models (MLLMs)☆99Updated last year
- [NeurIPS2023] Exploring Diverse In-Context Configurations for Image Captioning☆42Updated last year
- A comprehensive survey of Composed Multi-modal Retrieval (CMR), including Composed Image Retrieval (CIR) and Composed Video Retrieval (CV…☆77Updated 4 months ago
- This is a summary of research on noisy correspondence. There may be omissions. If anything is missing please get in touch with us. Our em…☆76Updated 2 months ago
- Test-time Prompt Tuning (TPT) for zero-shot generalization in vision-language models (NeurIPS 2022))☆203Updated 3 years ago
- [ICML 2024] Official implementation for "HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding"☆106Updated last year
- Codes of the Fine-grained Textual Inversion network for Zero-Shot Composed Image Retrieval☆27Updated 9 months ago
- Survey: https://arxiv.org/pdf/2507.20198☆269Updated 3 weeks ago
- 📖 A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).☆946Updated 3 months ago
- official implementation of "Interpreting CLIP's Image Representation via Text-Based Decomposition"☆233Updated 7 months ago
- Awsome of VLM-CL. Continual Learning for VLMs: A Survey and Taxonomy Beyond Forgetting☆134Updated 2 weeks ago