MMMU-Benchmark / MMMU
This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI"
☆412Updated last month
Alternatives and similar repositories for MMMU:
Users that are interested in MMMU are comparing it to the libraries listed below
- (CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.☆337Updated 3 months ago
- Aligning LMMs with Factually Augmented RLHF☆362Updated last year
- [CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback☆275Updated 7 months ago
- Official Repo of "MMBench: Is Your Multi-modal Model an All-around Player?"☆198Updated 7 months ago
- MM-Vet: Evaluating Large Multimodal Models for Integrated Capabilities (ICML 2024)☆294Updated 2 months ago
- [CVPR'25 highlight] RLAIF-V: Open-Source AI Feedback Leads to Super GPT-4V Trustworthiness