inquire-benchmark / INQUIRELinks
This repo contains the evaluation code for the INQUIRE benchmark
☆51Updated 7 months ago
Alternatives and similar repositories for INQUIRE
Users that are interested in INQUIRE are comparing it to the libraries listed below
Sorting:
- This is the repository for the BioCLIP model and the TreeOfLife-10M dataset [CVPR'24 Oral, Best Student Paper].☆208Updated last month
- Official repository of paper "Subobject-level Image Tokenization" (ICML-25)☆78Updated 2 weeks ago
- Official implementation of "Describing Differences in Image Sets with Natural Language" (CVPR 2024 Oral)☆120Updated last year
- Sapsucker Woods 60 Audiovisual Dataset☆15Updated 2 years ago
- Code and data for the paper "Emergent Visual-Semantic Hierarchies in Image-Text Representations" (ECCV 2024)☆28Updated 11 months ago
- This repo contains the official implementation of ICLR 2024 paper "Is ImageNet worth 1 video? Learning strong image encoders from 1 long …☆90Updated last year
- Official implementation of "HowToCaption: Prompting LLMs to Transform Video Annotations at Scale." ECCV 2024☆54Updated 9 months ago
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 7 months ago
- NegCLIP.☆33Updated 2 years ago
- ☆51Updated 4 months ago
- Official This-Is-My Dataset published in CVPR 2023☆16Updated last year
- [NeurIPS 2023] A faithful benchmark for vision-language compositionality☆82Updated last year
- An open source implementation of CLIP (With TULIP Support)☆160Updated 2 months ago
- Code for the paper "Hyperbolic Image-Text Representations", Desai et al, ICML 2023☆172Updated last year
- Code and datasets for "What’s “up” with vision-language models? Investigating their struggle with spatial reasoning".☆54Updated last year
- Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)☆169Updated 3 weeks ago
- [CVPR23 Highlight] CREPE: Can Vision-Language Foundation Models Reason Compositionally?☆33Updated 2 years ago
- Code base of SynthCLIP: CLIP training with purely synthetic text-image pairs from LLMs and TTIs.☆100Updated 3 months ago
- Official implementation of "Connect, Collapse, Corrupt: Learning Cross-Modal Tasks with Uni-Modal Data" (ICLR 2024)☆32Updated 9 months ago
- Code for Finetune like you pretrain: Improved finetuning of zero-shot vision models☆100Updated last year
- [ECCV 2024] Official Release of SILC: Improving vision language pretraining with self-distillation☆44Updated 9 months ago
- [CVPR24] Official Implementation of GEM (Grounding Everything Module)☆126Updated 3 months ago
- [ACL2025] Unsolvable Problem Detection: Robust Understanding Evaluation for Large Multimodal Models☆77Updated last month
- ☆77Updated 9 months ago
- Learning to Count without Annotations☆23Updated last year
- Library implementation of "No Train, all Gain: Self-Supervised Gradients Improve Deep Frozen Representations"☆38Updated 8 months ago
- This is an official implementation for [ICLR'24] INTR: Interpretable Transformer for Fine-grained Image Classification.☆51Updated last year
- Code and Models for "GeneCIS A Benchmark for General Conditional Image Similarity"☆59Updated 2 years ago
- LLaVA-MORE: A Comparative Study of LLMs and Visual Backbones for Enhanced Visual Instruction Tuning☆140Updated 2 months ago
- Code release for "Improved baselines for vision-language pre-training"☆60Updated last year