zengxingchen / ChartQA-MLLM
[IEEE VIS 2024] LLaVA-Chart: Advancing Multimodal Large Language Models in Chart Question Answering with Visualization-Referenced Instruction Tuning
☆63Updated 3 weeks ago
Alternatives and similar repositories for ChartQA-MLLM:
Users that are interested in ChartQA-MLLM are comparing it to the libraries listed below
- Code & Dataset for Paper: "Distill Visual Chart Reasoning Ability from LLMs to MLLMs"☆50Updated 3 months ago
- The official repository for "2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining"☆141Updated last month
- Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Models☆131Updated 2 months ago
- [NeurIPS 2024] A task generation and model evaluation system for multimodal language models.☆64Updated 2 months ago
- [ICLR 2025] Video-STaR: Self-Training Enables Video Instruction Tuning with Any Supervision☆59Updated 7 months ago
- Code for Paper: Harnessing Webpage Uis For Text Rich Visual Understanding☆46Updated 2 months ago
- official code for "BoostStep: Boosting mathematical capability of Large Language Models via improved single-step reasoning"☆28Updated 3 weeks ago
- MM-Instruct: Generated Visual Instructions for Large Multimodal Model Alignment☆33Updated 7 months ago
- Codes for Visual Sketchpad: Sketching as a Visual Chain of Thought for Multimodal Language Models☆161Updated 3 months ago
- ☆47Updated last month
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆89Updated last month
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs☆74Updated 3 months ago
- Code and data for OS-Genesis: Automating GUI Agent Trajectory Construction via Reverse Task Synthesis☆103Updated 3 weeks ago
- ☆16Updated 3 weeks ago
- Official repository for Montessori-Instruct: Generate Influential Training Data Tailored for Student Learning [ICLR 2025]☆41Updated 3 weeks ago
- Code for the paper "AutoPresent: Designing Structured Visuals From Scratch"☆48Updated last month
- The codebase for our EMNLP24 paper: Multimodal Self-Instruct: Synthetic Abstract Image and Visual Reasoning Instruction Using Language Mo…☆71Updated 3 weeks ago
- This repository contains the code for the paper: SirLLM: Streaming Infinite Retentive LLM☆57Updated 8 months ago
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"☆81Updated 4 months ago
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models☆77Updated 7 months ago
- [NeurIPS 2024] Official PyTorch implementation code for realizing the technical part of Mamba-based traversal of rationale (Meteor) to im…☆108Updated 8 months ago
- [ACL 2024] ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.☆111Updated 5 months ago
- [ICLR2025] MMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models☆61Updated 5 months ago
- [NeurIPS 2024] Official Implementation for Optimus-1: Hybrid Multimodal Memory Empowered Agents Excel in Long-Horizon Tasks☆61Updated last month
- OCR Hinders RAG: Evaluating the Cascading Impact of OCR on Retrieval-Augmented Generation☆65Updated 2 weeks ago
- Ruler: A Model-Agnostic Method to Control Generated Length for Large Language Models☆34Updated 4 months ago
- Large Language Models Can Self-Improve in Long-context Reasoning☆62Updated 2 months ago
- ☆73Updated 11 months ago
- ☆13Updated 2 months ago