g-luo / geolocation_via_guidebook_groundingLinks
G^3: Geolocation via Guidebook Grounding, Findings of EMNLP 2022
☆17Updated last year
Alternatives and similar repositories for geolocation_via_guidebook_grounding
Users that are interested in geolocation_via_guidebook_grounding are comparing it to the libraries listed below
Sorting:
- [CVPR24 Highlights] Polos: Multimodal Metric Learning from Human Feedback for Image Captioning☆32Updated 6 months ago
- Command-line tool for downloading and extending the RedCaps dataset.☆50Updated last year
- Code for 'Why is Winoground Hard? Investigating Failures in Visuolinguistic Compositionality', EMNLP 2022☆31Updated 2 years ago
- ICCV 2023 (Oral) Open-domain Visual Entity Recognition Towards Recognizing Millions of Wikipedia Entities☆43Updated 5 months ago
- NLX-GPT: A Model for Natural Language Explanations in Vision and Vision-Language Tasks, CVPR 2022 (Oral)☆49Updated last year
- [NeurIPS 2023] A faithful benchmark for vision-language compositionality☆88Updated last year
- Code and datasets for "What’s “up” with vision-language models? Investigating their struggle with spatial reasoning".☆66Updated last year
- Data and code for NeurIPS 2021 Paper "IconQA: A New Benchmark for Abstract Diagram Understanding and Visual Language Reasoning".☆53Updated last year
- Visual Language Transformer Interpreter - An interactive visualization tool for interpreting vision-language transformers☆98Updated 2 years ago
- ☆53Updated 3 years ago
- The SVO-Probes Dataset for Verb Understanding☆31Updated 3 years ago
- We present **FOCI**, a benchmark for Fine-grained Object ClassIfication for large vision language models (LVLMs).☆18Updated last year
- ☆37Updated 2 years ago
- ☆120Updated 2 years ago
- Situation With Groundings (SWiG) dataset and Joint Situation Localizer (JSL)☆69Updated 4 years ago
- Experiments and data for the paper "When and why vision-language models behave like bags-of-words, and what to do about it?" Oral @ ICLR …☆286Updated 2 years ago
- Code and data for EMNLP 2023 paper "Grounding Visual Illusions in Language: Do Vision-Language Models Perceive Illusions Like Humans?"☆14Updated last year
- Code, data, models for the Sherlock corpus☆58Updated 3 years ago
- [IJCAI 2025] Image Captioning Evaluation in the Age of Multimodal LLMs: Challenges and Future Perspectives☆22Updated last week
- ☆67Updated 2 years ago
- ☆22Updated 3 years ago
- Code for paper "Point and Ask: Incorporating Pointing into Visual Question Answering"☆19Updated 3 years ago
- Official Code Release for "Diagnosing and Rectifying Vision Models using Language" (ICLR 2023)☆34Updated 2 years ago
- ☆40Updated 2 years ago
- This repository provides data for the VAW dataset as described in the CVPR 2021 paper titled "Learning to Predict Visual Attributes in th…☆68Updated 3 years ago
- An ever-growing playground of notebooks showcasing CLIP's impressive zero-shot capabilities☆175Updated 3 years ago
- [ICML 2024] Fool Your (Vision and) Language Model With Embarrassingly Simple Permutations☆15Updated 2 years ago
- CapDec: SOTA Zero Shot Image Captioning Using CLIP and GPT2, EMNLP 2022 (findings)☆202Updated last year
- Using LLMs and pre-trained caption models for super-human performance on image captioning.☆42Updated 2 years ago
- A task-agnostic vision-language architecture as a step towards General Purpose Vision☆92Updated 4 years ago