Detail-Oriented CLIP for Fine-Grained Tasks (ICLR SSI-FM 2025)
☆57Mar 26, 2025Updated 11 months ago
Alternatives and similar repositories for DetailCLIP
Users that are interested in DetailCLIP are comparing it to the libraries listed below
Sorting:
- ☆37Apr 13, 2023Updated 2 years ago
- Awesome Vision-Language Compositionality, a comprehensive curation of research papers in literature.☆35Feb 13, 2025Updated last year
- Generalizing from SIMPLE to HARD Visual Reasoning: Can We Mitigate Modality Imbalance in VLMs?☆16Jun 3, 2025Updated 9 months ago
- Retrieval-augmented Image Captioning☆13Feb 16, 2023Updated 3 years ago
- Implementation of "DIME-FM: DIstilling Multimodal and Efficient Foundation Models"☆15Oct 12, 2023Updated 2 years ago
- ☆14Dec 31, 2024Updated last year
- ☆29Jun 10, 2024Updated last year
- [EMNLP 2024] IFCap: Image-like Retrieval and Frequency-based Entity Filtering for Zero-shot Captioning☆15May 13, 2025Updated 9 months ago
- Official Implementation of Attentive Mask CLIP (ICCV2023, https://arxiv.org/abs/2212.08653)☆35May 29, 2024Updated last year
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆43Mar 11, 2025Updated 11 months ago
- Using LLMs and pre-trained caption models for super-human performance on image captioning.☆42Oct 13, 2023Updated 2 years ago
- Knowledge Distillation using Contrastive Language-Image Pretraining (CLIP) without a teacher model.☆18Sep 6, 2024Updated last year
- [AAAI 2025] Explore In-Context Segmentation via Latent Diffusion Models☆22Mar 25, 2025Updated 11 months ago
- Code for "CLIP Behaves like a Bag-of-Words Model Cross-modally but not Uni-modally"☆20Updated this week
- Pytorch implementation for "Erasing the Bias: Fine-Tuning Foundation Models for Semi-Supervised Learning" (ICML 2024)☆24May 11, 2025Updated 9 months ago
- ☆20May 3, 2025Updated 10 months ago
- [ICCV 2023] With a Little Help from your own Past: Prototypical Memory Networks for Image Captioning.☆19Jun 7, 2024Updated last year
- MAGNet: Multi-scale Awareness and Global Fusion Network for RGB-D Salient Object Detection☆25Aug 10, 2024Updated last year
- ☆22Apr 27, 2024Updated last year
- ULPatch is open source user space live patch tool.☆13Jan 11, 2026Updated last month
- ☆49Feb 18, 2025Updated last year
- SVL-Adapter: Self-Supervised Adapter for Vision-Language Pretrained Models☆21Jan 11, 2024Updated 2 years ago
- 📍 Official repository of paper "ProtoCLIP: Prototypical Contrastive Language Image Pretraining" (IEEE TNNLS 2023)☆55Nov 8, 2023Updated 2 years ago
- Code for ECCV 2022 Workshop paper "See Finer, See More: Implicit Modality Alignment for Text-based Person Retrieval"☆21Nov 16, 2025Updated 3 months ago
- CVPR2023:Uncertainty-Aware Optimal Transport for Semantically Coherent Out-of-Distribution Detection☆26Mar 27, 2023Updated 2 years ago
- Official PyTorch implementation of the paper "CoVR: Learning Composed Video Retrieval from Web Video Captions".☆118Oct 9, 2025Updated 4 months ago
- [CVPR 2023 & IJCV 2025] Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation☆65Jul 29, 2025Updated 7 months ago
- The official implementation of paper: "Multi-Grained Contrast for Data-Efficient Unsupervised Representation Learning"☆30Jul 17, 2024Updated last year
- This repository houses the code for the paper - "The Neglected of VLMs"☆30Dec 31, 2025Updated 2 months ago
- The official implementation of 《MLLMs-Augmented Visual-Language Representation Learning》☆31Mar 12, 2024Updated last year
- Margin-based Vision Transformer☆67Nov 28, 2025Updated 3 months ago
- Code for "AVG-LLaVA: A Multimodal Large Model with Adaptive Visual Granularity"☆33Oct 12, 2024Updated last year
- BenchX: A Unified Benchmark Framework for Medical Vision-Language Pretraining on Chest X-Rays☆47Dec 27, 2025Updated 2 months ago
- LLM2CLIP significantly improves already state-of-the-art CLIP models.☆630Feb 1, 2026Updated last month
- An easy to use, user-friendly and efficient code for extracting OpenAI CLIP (Global/Grid) features from image and text respectively.☆136Jan 1, 2025Updated last year
- A collection of multimodal datasets, and visual features for VQA and captionning in pytorch. Just run "pip install multimodal"☆84Feb 25, 2022Updated 4 years ago
- CLIP-Mamba: CLIP Pretrained Mamba Models with OOD and Hessian Evaluation☆78Aug 15, 2024Updated last year
- Belief Revision based Caption Re-ranker with Visual Semantic Information. COLING 2022☆11Apr 13, 2025Updated 10 months ago
- This repo contains evaluation code for the paper "BLINK: Multimodal Large Language Models Can See but Not Perceive". https://arxiv.or…☆163Sep 27, 2025Updated 5 months ago