sjy0727 / CLIP-Text-Image-RetrievalLinks
该项目旨在通过输入文本描述来检索与之相匹配的图片。
☆41Updated 2 years ago
Alternatives and similar repositories for CLIP-Text-Image-Retrieval
Users that are interested in CLIP-Text-Image-Retrieval are comparing it to the libraries listed below
Sorting:
- ☆49Updated 2 years ago
- 毕业设计:《基于CLIP模型的视频文本检索设计与实现》☆11Updated last year
- USER: Unified Semantic Enhancement with Momentum Contrast for Image-Text Retrieval, TIP 2024☆33Updated 3 months ago
- 基于ClipCap的看图说话Image Caption模型☆313Updated 3 years ago
- Implementation of our paper, 'Unifying Two-Stream Encoders with Transformers for Cross-Modal Retrieval.'☆26Updated last year
- Efficient Token-Guided Image-Text Retrieval with Consistent Multimodal Contrastive Training☆28Updated 2 years ago
- Official Code for the ICCV23 Paper: "LexLIP: Lexicon-Bottlenecked Language-Image Pre-Training for Large-Scale Image-Text Sparse Retrieval…☆40Updated last year
- Summary of Related Research on Image-Text Matching☆71Updated 2 years ago
- [SIGIR 2024] - Simple but Effective Raw-Data Level Multimodal Fusion for Composed Image Retrieval☆43Updated last year
- ☆19Updated last year
- [ACM MM 2024] Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives☆39Updated last week
- ☆28Updated 2 years ago
- This is the official repository for Retrieval Augmented Visual Question Answering☆237Updated 9 months ago
- [Paper][AAAI2024]Structure-CLIP: Towards Scene Graph Knowledge to Enhance Multi-modal Structured Representations☆148Updated last year
- 计算机视觉课程设计-基于Chinese-CLIP的图文检索系统☆94Updated 2 years ago
- [CVPR 2024] LION: Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge☆151Updated 2 weeks ago
- Source code of our AAAI 2024 paper "Cross-Modal and Uni-Modal Soft-Label Alignment for Image-Text Retrieval"☆47Updated last year
- A comprehensive survey of Composed Multi-modal Retrieval (CMR), including Composed Image Retrieval (CIR) and Composed Video Retrieval (CV…☆55Updated last month
- Reason-before-Retrieve: One-Stage Reflective Chain-of-Thoughts for Training-Free Zero-Shot Composed Image Retrieval [CVPR 2025 Highlight]☆58Updated 2 months ago
- Research Code for Multimodal-Cognition Team in Ant Group☆165Updated 2 months ago
- 自己阅读的多模态对话系统论文(及部分笔记)汇总☆23Updated 2 years ago
- GroundVLP: Harnessing Zero-shot Visual Grounding from Vision-Language Pre-training and Open-Vocabulary Object Detection (AAAI 2024)☆71Updated last year
- [TIP2023] The code of “Plug-and-Play Regulators for Image-Text Matching”☆33Updated last year
- Multimodal-Composite-Editing-and-Retrieval-update☆33Updated 10 months ago
- [AAAI 2024] TagCLIP: A Local-to-Global Framework to Enhance Open-Vocabulary Multi-Label Classification of CLIP Without Training☆101Updated last year
- Cross-Modal-Real-valuded-Retrieval☆84Updated 2 years ago
- The official implementation for BLIP4CIR with bi-directional training | Bi-directional Training for Composed Image Retrieval via Text Pro…☆32Updated last year
- Implementation of our CVPR2022 paper, Negative-Aware Attention Framework for Image-Text Matching.☆119Updated 2 years ago
- A project that can generate ancient poems based on pictures, including CLIP, T5, GPT2 models☆22Updated 7 months ago
- Towards Efficient and Effective Text-to-Video Retrieval with Coarse-to-Fine Visual Representation Learning☆19Updated 7 months ago