[TMM 2023] Self-paced Curriculum Adapting of CLIP for Visual Grounding.
☆132Nov 10, 2025Updated 3 months ago
Alternatives and similar repositories for CLIP-VG
Users that are interested in CLIP-VG are comparing it to the libraries listed below
Sorting:
- [ACM MM 2024] Hierarchical Multimodal Fine-grained Modulation for Visual Grounding.☆59Nov 10, 2025Updated 3 months ago
- ☆13Mar 14, 2025Updated 11 months ago
- ☆23Aug 20, 2024Updated last year
- [NeurIPS 2024] OneRef: Unified One-tower Expression Grounding and Segmentation with Mask Referring Modeling.☆30Nov 13, 2025Updated 3 months ago
- ☆28Nov 27, 2025Updated 3 months ago
- [NeurIPS2024] - SimVG: A Simple Framework for Visual Grounding with Decoupled Multi-modal Fusion☆100Oct 29, 2025Updated 4 months ago
- ☆195Feb 27, 2024Updated 2 years ago
- Pytorch code for paper From CLIP to DINO: Visual Encoders Shout in Multi-modal Large Language Models☆207Jan 8, 2025Updated last year
- ☆39Jun 28, 2023Updated 2 years ago
- ☆32Mar 25, 2024Updated last year
- ☆41Jun 3, 2022Updated 3 years ago
- ☆20Apr 2, 2024Updated last year
- [TPAMI 2025] Towards Visual Grounding: A Survey☆294Nov 18, 2025Updated 3 months ago
- ☆10Jun 21, 2024Updated last year
- ☆58Aug 7, 2023Updated 2 years ago
- [CVPR 2022] Pseudo-Q: Generating Pseudo Language Queries for Visual Grounding☆153Jul 13, 2024Updated last year
- Video Feature Enhancement with PyTorch☆32Nov 28, 2024Updated last year
- [CVPR 2023] Cascade Evidential Learning for Open-world Weakly-supervised Temporal Action Localization☆12Jul 9, 2024Updated last year
- [CVPR 2024] Retrieval-Augmented Image Captioning with External Visual-Name Memory for Open-World Comprehension☆61Apr 8, 2024Updated last year
- The source code of the CVPR22 paper titled "Multi-Modal Dynamic Graph Transformer for Visual Grounding".☆22Mar 26, 2022Updated 3 years ago
- [CVPR 2025] DynRefer: Delving into Region-level Multimodal Tasks via Dynamic Resolution☆58Mar 4, 2025Updated last year
- Improving One-stage Visual Grounding by Recursive Sub-query Construction, ECCV 2020☆89Sep 30, 2021Updated 4 years ago
- Improving Visual Grounding with Visual-Linguistic Verification and Iterative Reasoning, CVPR 2022☆96Dec 2, 2022Updated 3 years ago
- Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone☆131Oct 10, 2023Updated 2 years ago
- [ECCV 2024] ControlCap: Controllable Region-level Captioning☆80Oct 25, 2024Updated last year
- awesome grounding: A curated list of research papers in visual grounding☆1,125Sep 21, 2025Updated 5 months ago
- ☆61May 2, 2025Updated 10 months ago
- Official repository of "Interactive Text-to-Image Retrieval with Large Language Models: A Plug-and-Play Approach" (ACL 2024 Oral)☆34Mar 24, 2025Updated 11 months ago
- Transactions on Multimedia (TMM25)☆19Apr 8, 2025Updated 10 months ago
- [CVPR2024] ViP-LLaVA: Making Large Multimodal Models Understand Arbitrary Visual Prompts☆336Jul 17, 2024Updated last year
- (NeurIPS 2023) Open-set visual object query search & localization in long-form videos☆26Feb 1, 2024Updated 2 years ago
- [TPAMI 2024] Dynamic MDETR: A Dynamic Multimodal Transformer Decoder for Visual Grounding☆29Sep 11, 2024Updated last year
- [NeurIPS'24] MemVLT: Vision-Language Tracking with Adaptive Memory-based Prompts☆18Oct 7, 2024Updated last year
- [CVPR 2023] Code for "Improving Visual Grounding by Encouraging Consistent Gradient-based Explanations"☆19Oct 10, 2023Updated 2 years ago
- RSVG: Exploring Data and Model for Visual Grounding on Remote Sensing Data, 2022☆170Dec 10, 2025Updated 2 months ago
- The official implementation of RAR☆92Dec 9, 2025Updated 2 months ago
- A detection/segmentation dataset with labels characterized by intricate and flexible expressions. "Described Object Detection: Liberating…☆137Mar 20, 2024Updated last year
- Official Implementation of Attentive Mask CLIP (ICCV2023, https://arxiv.org/abs/2212.08653)☆35May 29, 2024Updated last year
- New starting point, keep trying. 平时学习时记录的一些东西,对他人参考价值有限。最近在重新规划,建一个知识库,会涵盖更多的知识面。链接 👉 https://github.com/Angus-Liu/mtbox☆17Jul 7, 2023Updated 2 years ago