nttmdlab-nlp / InstructDocLinks
InstructDoc: A Dataset for Zero-Shot Generalization of Visual Document Understanding with Instructions (AAAI2024)
☆158Updated last year
Alternatives and similar repositories for InstructDoc
Users that are interested in InstructDoc are comparing it to the libraries listed below
Sorting:
- My implementation of Kosmos2.5 from the paper: "KOSMOS-2.5: A Multimodal Literate Model"☆72Updated 2 weeks ago
- The WordScape repository contains code for the WordScape pipeline to create datasets to train document understanding models.☆37Updated last year
- ☆141Updated last year
- official code for "Fox: Focus Anywhere for Fine-grained Multi-page Document Understanding"☆171Updated last year
- ☆98Updated 10 months ago
- Vary-tiny codebase upon LAVIS (for training from scratch)and a PDF image-text pairs data (about 600k including English/Chinese)☆86Updated last year
- ☆67Updated last year
- Datasets and Evaluation Scripts for CompHRDoc☆51Updated 8 months ago
- Visually-Situated Natural Language Understanding with Contrastive Reading Model and Frozen Large Language Models, EMNLP 2023☆46Updated last year
- We identify the desiderata for a comprehensive benchmark and propose Visually Rich Document Understanding (VRDU). VRDU contains two datas…☆80Updated 2 years ago
- (ICCV 2025) OCR Hinders RAG: Evaluating the Cascading Impact of OCR on Retrieval-Augmented Generation☆88Updated 3 months ago
- E5-V: Universal Embeddings with Multimodal Large Language Models☆272Updated 10 months ago
- Document Artifical Intelligence☆189Updated 3 weeks ago
- Document Haystacks: Vision-Language Reasoning Over Piles of 1000+ Documents, CVPR 2025☆25Updated 9 months ago
- Doc2Graph transforms documents into graphs and exploit a GNN to solve several tasks.☆133Updated last week
- Official Repository of MMLONGBENCH-DOC: Benchmarking Long-context Document Understanding with Visualizations☆100Updated 3 weeks ago
- SlideVQA: A Dataset for Document Visual Question Answering on Multiple Images (AAAI2023)☆98Updated 6 months ago
- Dataset and Code for our ACL 2024 paper: "Multimodal Table Understanding". We propose the first large-scale Multimodal IFT and Pre-Train …☆218Updated 4 months ago
- MTVQA: Benchmarking Multilingual Text-Centric Visual Question Answering. A comprehensive evaluation of multimodal large model multilingua…☆63Updated 5 months ago
- This project is a collection of fine-tuning scripts to help researchers fine-tune Qwen 2 VL on HuggingFace datasets.☆74Updated 3 months ago
- [NAACL 2024] MMC: Advancing Multimodal Chart Understanding with LLM Instruction Tuning☆96Updated 9 months ago
- Parameter-efficient finetuning script for Phi-3-vision, the strong multimodal language model by Microsoft.☆58Updated last year
- Code/Data for the paper: "LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image Understanding"☆268Updated last year
- [ACL 2024] ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.☆130Updated last year
- Algorithms, papers, datasets, performance comparisons for Document AI. Continuously updating.☆202Updated 7 months ago
- The huggingface implementation of Fine-grained Late-interaction Multi-modal Retriever.☆100Updated 4 months ago
- Evaluation of the Optical Character Recognition (OCR) capabilities of GPT-4V(ision)☆125Updated last year
- Exploring Efficient Fine-Grained Perception of Multimodal Large Language Models☆63Updated 11 months ago
- Dataset and scripts for HRDoc☆39Updated 2 years ago
- [EMNLP 2025] Distill Visual Chart Reasoning Ability from LLMs to MLLMs☆56Updated 2 months ago