LHL3341 / ContextBLIP
☆10Updated 6 months ago
Related projects ⓘ
Alternatives and complementary repositories for ContextBLIP
- Code for paper "AGLA: Mitigating Object Hallucinations in Large Vision-Language Models with Assembly of Global and Local Attention"☆16Updated 4 months ago
- [TPAMI 2024] This is the Pytorch code for our paper "Context Disentangling and Prototype Inheriting for Robust Visual Grounding".☆14Updated last month
- Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models (AAAI 2024)☆67Updated 9 months ago
- ☆22Updated last year
- [CVPR2024 Highlight] Official implementation for Transferable Visual Prompting. The paper "Exploring the Transferability of Visual Prompt…☆32Updated 4 months ago
- [ECCV2024] Learning Video Context as Interleaved Multimodal Sequences☆29Updated last month
- [NeurIPS 2024] Official PyTorch implementation of LoTLIP: Improving Language-Image Pre-training for Long Text Understanding☆30Updated this week
- FreeVA: Offline MLLM as Training-Free Video Assistant☆48Updated 5 months ago
- [CVPR' 2024] Contrasting Intra-Modal and Ranking Cross-Modal Hard Negatives to Enhance Visio-Linguistic Fine-grained Understanding☆41Updated 3 months ago
- This repo holds the official code and data for "Beyond Literal Descriptions: Understanding and Locating Open-World Objects Aligned with H…☆17Updated 6 months ago
- [CVPR 2024] Retrieval-Augmented Image Captioning with External Visual-Name Memory for Open-World Comprehension☆34Updated 7 months ago
- ☆33Updated 2 years ago
- The official implementation of paper "Prototype-based Aleatoric Uncertainty Quantification for Cross-modal Retrieval" accepted by NeurIPS…☆21Updated 6 months ago
- NegCLIP.☆26Updated last year
- [TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”☆29Updated 7 months ago
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)☆41Updated 4 months ago
- Code for Label Propagation for Zero-shot Classification with Vision-Language Models (CVPR2024)☆33Updated 3 months ago
- ☆16Updated 6 months ago
- 【ICLR 2024, Spotlight】Sentence-level Prompts Benefit Composed Image Retrieval☆68Updated 7 months ago
- A Large Multimodal Model for Pixel-Level Visual Grounding in Videos☆34Updated last week
- CLIP-MoE: Mixture of Experts for CLIP☆17Updated last month
- [CVPR 2024] The official implementation of paper "synthesize, diagnose, and optimize: towards fine-grained vision-language understanding"☆30Updated last week
- ☆25Updated last year
- [NeurlPS 2024] One Token to Seg Them All: Language Instructed Reasoning Segmentation in Videos☆33Updated 2 weeks ago
- Look, Compare, Decide: Alleviating Hallucination in Large Vision-Language Models via Multi-View Multi-Path Reasoning☆17Updated 2 months ago
- PyTorch code for "Contrastive Region Guidance: Improving Grounding in Vision-Language Models without Training"☆28Updated 8 months ago
- ☆24Updated 4 months ago
- The First to Know: How Token Distributions Reveal Hidden Knowledge in Large Vision-Language Models?☆19Updated 2 weeks ago
- ☆13Updated 3 months ago
- Task Residual for Tuning Vision-Language Models (CVPR 2023)☆66Updated last year