Official PyTorch Implementation of MLLM Is a Strong Reranker: Advancing Multimodal Retrieval-augmented Generation via Knowledge-enhanced Reranking and Noise-injected Training.
☆91Nov 15, 2024Updated last year
Alternatives and similar repositories for RagVL
Users that are interested in RagVL are comparing it to the libraries listed below
Sorting:
- [CVPR 2025] Docopilot: Improving Multimodal Models for Document-Level Understanding☆36Jul 22, 2025Updated 7 months ago
- The official repository of MM-R5☆29Jun 22, 2025Updated 8 months ago
- [EMNLP 2024] SURf: Teaching Large Vision-Language Models to Selectively Utilize Retrieved Information☆12Oct 11, 2024Updated last year
- We introduce new approach, Token Reduction using CLIP Metric (TRIM), aimed at improving the efficiency of MLLMs without sacrificing their…☆21Jan 11, 2026Updated 2 months ago
- ☆68Oct 27, 2023Updated 2 years ago
- [ECCV'24] Official Implementation of Autoregressive Visual Entity Recognizer.☆14Mar 2, 2024Updated 2 years ago
- [EMNLP25 Main]The official code of "Gradient-Attention Guided Dual-Masking Synergetic Framework for Robust Text-based Person Retrieval"☆22Mar 11, 2026Updated last week
- Parsing-free RAG supported by VLMs☆935Dec 7, 2025Updated 3 months ago
- [ACM TOMM] Official implementation of "TextCoT: Zoom-In for Enhanced Multimodal Text-Rich Image Understanding"☆44Feb 27, 2026Updated 3 weeks ago
- (ICLR 2025) AgentRefine: Enhancing Agent Generalization through Refinement Tuning☆19Nov 22, 2025Updated 3 months ago
- [EMNLP 2025] Official codebase for Rearank: Reasoning Re-ranking Agent☆33Aug 20, 2025Updated 7 months ago
- For the rlhf learning environment of Koreans☆25Sep 25, 2023Updated 2 years ago
- ☆10Dec 16, 2023Updated 2 years ago
- [CVPR 2025] Augmenting Multimodal LLMs with Self-Reflective Tokens for Knowledge-based Visual Question Answering☆54Jul 14, 2025Updated 8 months ago
- ☆58Feb 27, 2025Updated last year
- RuleRAG: Rule Meets Retrieval-Augmented Generation for Question Answering☆32Oct 8, 2025Updated 5 months ago
- ICCV 2023 (Oral) Open-domain Visual Entity Recognition Towards Recognizing Millions of Wikipedia Entities☆43Jun 7, 2025Updated 9 months ago
- [EMNLP 2024 Findings] The official PyTorch implementation of EchoSight: Advancing Visual-Language Models with Wiki Knowledge.☆81Jan 19, 2026Updated 2 months ago
- Stereo Depth Estimation with Echoes at ECCV 2022☆10Sep 20, 2022Updated 3 years ago
- [CVPR 2025] LamRA: Large Multimodal Model as Your Advanced Retrieval Assistant☆179Jul 7, 2025Updated 8 months ago
- [ICLR 2025] Vision-Centric Evaluation for Retrieval-Augmented Multimodal Models☆61Jan 22, 2025Updated last year
- [EMNLP 2025] The official implementation of "Zero-shot Multimodal Document Retrieval via Cross-Modal Question Generation"☆15Aug 26, 2025Updated 6 months ago
- [ACM MM2025] The official repository for the RealSyn dataset☆40Dec 14, 2025Updated 3 months ago
- ☆34Oct 9, 2025Updated 5 months ago
- Both Text and Images Leaked! A Systematic Analysis of Data Contamination in Multimodal LLM | EMNLP 2025 Findings☆18Oct 17, 2025Updated 5 months ago
- EMNLP2023 - InfoSeek: A New VQA Benchmark focus on Visual Info-Seeking Questions☆25May 30, 2024Updated last year
- ☆15May 15, 2025Updated 10 months ago
- This repo contains the code for "VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks" [ICLR 2025]☆597Updated this week
- E5-V: Universal Embeddings with Multimodal Large Language Models☆275Dec 10, 2025Updated 3 months ago
- [CVPR 2025] OmniMMI: A Comprehensive Multi-modal Interaction Benchmark in Streaming Video Contexts☆17Apr 2, 2025Updated 11 months ago
- The huggingface implementation of Fine-grained Late-interaction Multi-modal Retriever.☆104May 30, 2025Updated 9 months ago
- ☆13Apr 23, 2025Updated 10 months ago
- Code and Data for "FaithfulRAG: Fact-Level Conflict Modeling for Context-Faithful Retrieval-Augmented Generation" (ACL25)☆29Oct 26, 2025Updated 4 months ago
- OmniGAIA: Towards Native Omni-Modal AI Agents☆66Feb 28, 2026Updated 2 weeks ago
- This is the code repo for our paper "Benchmarking Retrieval-Augmented Generation in Multi-Modal Contexts".☆43Sep 27, 2025Updated 5 months ago
- LLM2CLIP significantly improves already state-of-the-art CLIP models.☆640Feb 1, 2026Updated last month
- Code release for "SPIQA: A Dataset for Multimodal Question Answering on Scientific Papers" [NeurIPS D&B, 2024]☆74Jan 13, 2025Updated last year
- Repo for Benchmarking Multimodal Retrieval Augmented Generation with Dynamic VQA Dataset and Self-adaptive Planning Agent☆417Apr 22, 2025Updated 10 months ago
- ☆63Jan 3, 2025Updated last year