HAWLYQ / InfoMetIC
☆12Updated last year
Alternatives and similar repositories for InfoMetIC:
Users that are interested in InfoMetIC are comparing it to the libraries listed below
- Repository for the paper: dense and aligned captions (dac) promote compositional reasoning in vl models☆26Updated last year
- ICCV 2023 (Oral) Open-domain Visual Entity Recognition Towards Recognizing Millions of Wikipedia Entities☆37Updated 5 months ago
- The SVO-Probes Dataset for Verb Understanding☆31Updated 3 years ago
- Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation. CVPR 2023☆59Updated 3 months ago
- ☆58Updated last year
- Source code for EMNLP 2022 paper “PEVL: Position-enhanced Pre-training and Prompt Tuning for Vision-language Models”☆48Updated 2 years ago
- Code and data for ImageCoDe, a contextual vison-and-language benchmark☆39Updated 11 months ago
- ☆28Updated 3 months ago
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 2 months ago
- Official repository for the A-OKVQA dataset☆75Updated 9 months ago
- Benchmark data for "Rethinking Benchmarks for Cross-modal Image-text Retrieval" (SIGIR 2023)☆26Updated last year
- ☆11Updated last month
- [CVPR' 2024] Contrasting Intra-Modal and Ranking Cross-Modal Hard Negatives to Enhance Visio-Linguistic Fine-grained Understanding☆45Updated 6 months ago
- (ACL'2023) MultiCapCLIP: Auto-Encoding Prompts for Zero-Shot Multilingual Visual Captioning☆35Updated 6 months ago
- 【ICLR 2024, Spotlight】Sentence-level Prompts Benefit Composed Image Retrieval☆75Updated 10 months ago
- ☆24Updated last year
- [CVPR 2024] Retrieval-Augmented Image Captioning with External Visual-Name Memory for Open-World Comprehension☆45Updated 10 months ago
- Colorful Prompt Tuning for Pre-trained Vision-Language Models☆48Updated 2 years ago
- HalluciDoctor: Mitigating Hallucinatory Toxicity in Visual Instruction Data (Accepted by CVPR 2024)☆43Updated 7 months ago
- NegCLIP.☆30Updated 2 years ago
- Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"☆35Updated 6 months ago
- natual language guided image captioning☆78Updated last year
- [ICLR 23] Contrastive Aligned of Vision to Language Through Parameter-Efficient Transfer Learning☆38Updated last year
- Implementation of our paper, Your Negative May not Be True Negative: Boosting Image-Text Matching with False Negative Elimination..☆17Updated last year
- MultiInstruct: Improving Multi-Modal Zero-Shot Learning via Instruction Tuning☆135Updated last year
- Data repository for the VALSE benchmark.☆37Updated last year
- Python 3 support for the MS COCO caption evaluation tools☆14Updated 8 months ago
- [ACM MM 2024] Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives☆26Updated 3 months ago
- Source code and data used in the papers ViQuAE (Lerner et al., SIGIR'22), Multimodal ICT (Lerner et al., ECIR'23) and Cross-modal Retriev…☆31Updated 2 months ago
- Preference Learning for LLaVA☆37Updated 3 months ago