assafbk / mocha_codeLinks
Mitigating Open-Vocabulary Caption Hallucinations (EMNLP 2024)
☆19Updated last year
Alternatives and similar repositories for mocha_code
Users that are interested in mocha_code are comparing it to the libraries listed below
Sorting:
- UnifiedMLLM: Enabling Unified Representation for Multi-modal Multi-tasks With Large Language Model☆22Updated last year
- ViCToR: Improving Visual Comprehension via Token Reconstruction for Pretraining LMMs☆27Updated 5 months ago
- text-only training or language-free training for multimodal tasks (image/audio/video caption, retrieval, text2image)☆12Updated last year
- Question-Aware Gaussian Experts for Audio-Visual Question Answering -- Official Pytorch Implementation (CVPR'25, Highlight)☆26Updated 7 months ago
- ☆24Updated 2 years ago
- ✨✨The Curse of Multi-Modalities (CMM): Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audio☆53Updated 6 months ago
- Official repository of "Interactive Text-to-Image Retrieval with Large Language Models: A Plug-and-Play Approach" (ACL 2024 Oral)☆34Updated 10 months ago
- ☆19Updated last year
- [MM2024, oral] "Self-Supervised Visual Preference Alignment" https://arxiv.org/abs/2404.10501☆61Updated last year
- ☆11Updated last year
- LMM solved catastrophic forgetting, AAAI2025☆45Updated 9 months ago
- Repository for the paper: dense and aligned captions (dac) promote compositional reasoning in vl models☆27Updated 2 years ago
- [ICLR2024] The official implementation of paper "UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling", by …☆77Updated 2 years ago
- [ECCV’24] Official Implementation for CAT: Enhancing Multimodal Large Language Model to Answer Questions in Dynamic Audio-Visual Scenario…☆58Updated last year
- An automatic MLLM hallucination detection framework☆19Updated 2 years ago
- Official implementation of "Connect, Collapse, Corrupt: Learning Cross-Modal Tasks with Uni-Modal Data" (ICLR 2024)☆34Updated last year
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆39Updated this week
- A project for tri-modal LLM benchmarking and instruction tuning.☆56Updated 10 months ago
- [ECCV'24] Official Implementation of Autoregressive Visual Entity Recognizer.☆14Updated last year
- ☆11Updated last year
- LAVIS - A One-stop Library for Language-Vision Intelligence☆48Updated last year
- [NAACL 2024] LaDiC: Are Diffusion Models Really Inferior to Autoregressive Counterparts for Image-to-text Generation?☆43Updated last year
- ICCV 2023 (Oral) Open-domain Visual Entity Recognition Towards Recognizing Millions of Wikipedia Entities☆43Updated 7 months ago
- Official Implementation of CODE☆16Updated last year
- HallE-Control: Controlling Object Hallucination in LMMs☆31Updated last year
- VPEval Codebase from Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆45Updated 2 years ago
- VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)☆42Updated last month
- ☆23Updated last year
- This repo contains evaluation code for the paper "AV-Odyssey: Can Your Multimodal LLMs Really Understand Audio-Visual Information?"☆31Updated last year
- ☆58Updated last year