uni-medical / GMAI-MMBenchLinks
GMAI-MMBench: A Comprehensive Multimodal Evaluation Benchmark Towards General Medical AI.
☆68Updated 5 months ago
Alternatives and similar repositories for GMAI-MMBench
Users that are interested in GMAI-MMBench are comparing it to the libraries listed below
Sorting:
- A generalist foundation model for healthcare capable of handling diverse medical data modalities.☆74Updated last year
- GMAI-VL & GMAI-VL-5.5M: A Large Vision-Language Model and A Comprehensive Multimodal Dataset Towards General Medical AI.☆72Updated last month
- Official repository of paper titled "UniMed-CLIP: Towards a Unified Image-Text Pretraining Paradigm for Diverse Medical Imaging Modalitie…☆110Updated last month
- [ACL 2025] Exploring Compositional Generalization of Multimodal LLMs for Medical Imaging☆33Updated this week
- Med-R1: Reinforcement Learning for Generalizable Medical Reasoning in Vision-Language Models☆39Updated last week
- [ICLR 2025] MedRegA: Interpretable Bilingual Multimodal Large Language Model for Diverse Biomedical Tasks☆32Updated 2 months ago
- The first Chinese medical large vision-language model designed to integrate the analysis of textual and visual data☆61Updated last year
- The official repository of paper named 'A Refer-and-Ground Multimodal Large Language Model for Biomedicine'☆24Updated 7 months ago
- ☆48Updated 3 months ago
- [ICLR 2025] This is the official repository of our paper "MedTrinity-25M: A Large-scale Multimodal Dataset with Multigranular Annotations…☆340Updated 3 months ago
- The official repository of the paper 'Towards a Multimodal Large Language Model with Pixel-Level Insight for Biomedicine'☆60Updated 4 months ago
- [npj digital medicine] The official codes for "Towards Evaluating and Building Versatile Large Language Models for Medicine"☆65Updated last month
- ☆33Updated 4 months ago
- PMC-VQA is a large-scale medical visual question-answering dataset, which contains 227k VQA pairs of 149k images that cover various modal…☆203Updated 6 months ago
- Learning to Use Medical Tools with Multi-modal Agent☆157Updated 3 months ago
- The official codes for "Can Modern LLMs Act as Agent Cores in Radiology Environments?"☆25Updated 4 months ago
- ☆64Updated 4 months ago
- [CVPR 2025] BIOMEDICA: An Open Biomedical Image-Caption Archive, Dataset, and Vision-Language Models Derived from Scientific Literature☆61Updated 2 months ago
- [ICML'25] MMedPO: Aligning Medical Vision-Language Models with Clinical-Aware Multimodal Preference Optimization☆36Updated 3 months ago
- ☆122Updated 2 months ago
- Code implementation of RP3D-Diag☆15Updated 6 months ago
- [ICML 2025] MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding☆72Updated last month
- ☆46Updated last month
- The code for paper: PeFoM-Med: Parameter Efficient Fine-tuning on Multi-modal Large Language Models for Medical Visual Question Answering☆49Updated 2 weeks ago
- Multi-Aspect Vision Language Pretraining - CVPR2024☆78Updated 9 months ago
- MedAgentsBench: Benchmarking Thinking Models and Agent Frameworks for Complex Medical Reasoning☆40Updated last month
- OphNet: A Large-Scale Video Benchmark for Ophthalmic Surgical Workflow Understanding☆48Updated 2 months ago
- Codebase for Quilt-LLaVA☆55Updated 11 months ago
- ☆20Updated this week
- ☆78Updated last year