FuxiaoLiu / MMCView external linksLinks
[NAACL 2024] MMC: Advancing Multimodal Chart Understanding with LLM Instruction Tuning
☆95Jan 7, 2025Updated last year
Alternatives and similar repositories for MMC
Users that are interested in MMC are comparing it to the libraries listed below
Sorting:
- [EACL'23] COVID-VTS: Fact Extraction and Verification on Short Video Platforms☆11Sep 26, 2023Updated 2 years ago
- ☆85Aug 18, 2024Updated last year
- [ICPRAI 2024] DocumentCLIP: Linking Figures and Main Body Text in Reflowed Documents☆16Apr 4, 2024Updated last year
- A curated list of recent and past chart understanding work based on our IEEE TKDE survey paper: From Pixels to Insights: A Survey on Auto…☆231Dec 17, 2025Updated last month
- ☆255Dec 7, 2023Updated 2 years ago
- The proposed simulated dataset consisting of 9,536 charts and associated data annotations in CSV format.☆26Feb 22, 2024Updated last year
- [CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(…☆325Oct 14, 2025Updated 3 months ago
- ☆238Apr 18, 2025Updated 9 months ago
- Official repository of MMDU dataset☆103Sep 29, 2024Updated last year
- MTVQA: Benchmarking Multilingual Text-Centric Visual Question Answering. A comprehensive evaluation of multimodal large model multilingua…☆63May 15, 2025Updated 8 months ago
- TAT-DQA: Towards Complex Document Understanding By Discrete Reasoning☆23Sep 17, 2024Updated last year
- Dataset and annotations for ASSETS 2022 publication☆12Oct 6, 2022Updated 3 years ago
- ☆17Jun 12, 2024Updated last year
- Code and data for the ACL 2024 Findings paper "Do LVLMs Understand Charts? Analyzing and Correcting Factual Errors in Chart Captioning"☆27Jun 5, 2024Updated last year
- ☆72Jul 14, 2024Updated last year
- Dataset introduced in PlotQA: Reasoning over Scientific Plots☆82Jun 20, 2023Updated 2 years ago
- ☆88Jul 4, 2024Updated last year
- VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs☆54Mar 9, 2025Updated 11 months ago
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆79Jun 17, 2024Updated last year
- [ACL 2024] ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.☆132Sep 7, 2024Updated last year
- [IEEE VIS 2024] LLaVA-Chart: Advancing Multimodal Large Language Models in Chart Question Answering with Visualization-Referenced Instruc…☆75Jan 22, 2025Updated last year
- On the Hidden Mystery of OCR in Large Multimodal Models (OCRBench)☆791Jul 5, 2025Updated 7 months ago
- The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM", IJCV2025☆276May 26, 2025Updated 8 months ago
- ☆19Mar 12, 2025Updated 11 months ago
- [ICLR2025] LLaVA-HR: High-Resolution Large Language-Vision Assistant☆246Aug 14, 2024Updated last year
- Lion: Kindling Vision Intelligence within Large Language Models☆51Jan 25, 2024Updated 2 years ago
- ☆124Jul 14, 2024Updated last year
- CVPR 2022 (Oral) Pytorch Code for Unsupervised Vision-and-Language Pre-training via Retrieval-based Multi-Granular Alignment☆22Apr 15, 2022Updated 3 years ago
- ☆23Aug 17, 2024Updated last year
- A template for a Djinni library that can be used in Java/Kotlin, ObjC/Swift and C#☆11Oct 6, 2022Updated 3 years ago
- ☆69Jan 9, 2024Updated 2 years ago
- Code for Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense? [COLM 2024]☆24Aug 13, 2024Updated last year
- ☆20Apr 24, 2024Updated last year
- Official Repository of MMLONGBENCH-DOC: Benchmarking Long-context Document Understanding with Visualizations☆120Sep 28, 2025Updated 4 months ago
- [EMNLP'21] Visual News: Benchmark and Challenges in News Image Captioning☆104Jul 18, 2024Updated last year
- ☆27Jul 20, 2024Updated last year
- [ICLR 2025 Spotlight] OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text☆412May 5, 2025Updated 9 months ago
- LLaVA-UHD v3: Progressive Visual Compression for Efficient Native-Resolution Encoding in MLLMs☆413Dec 20, 2025Updated last month
- ☆12Jun 20, 2023Updated 2 years ago