FreedomIntelligence / HuatuoGPT-Vision
Medical Multimodal LLMs
☆239Updated last month
Alternatives and similar repositories for HuatuoGPT-Vision:
Users that are interested in HuatuoGPT-Vision are comparing it to the libraries listed below
- The official codes for "PMC-CLIP: Contrastive Language-Image Pre-training using Biomedical Documents"☆204Updated 5 months ago
- An interpretable large language model (LLM) for medical diagnosis.☆117Updated 5 months ago
- ClinicalLab: Aligning Agents for Multi-Departmental Clinical Diagnostics in the Real World☆76Updated 6 months ago
- Learning to Use Medical Tools with Multi-modal Agent☆113Updated this week
- 中文医学多模态大模型 Large Chinese Language-and-Vision Assistant for BioMedicine☆69Updated 8 months ago
- HuatuoGPT2, One-stage Training for Medical Adaption of LLMs. (An Open Medical GPT)☆356Updated 5 months ago
- [ICLR 2024] FairSeg: A Large-Scale Medical Image Segmentation Dataset for Fairness Learning Using Segment Anything Model with Fair Error-…☆86Updated last month
- u-LLaVA: Unifying Multi-Modal Tasks via Large Language Model☆129Updated 7 months ago
- [ICLR‘25 Spotlight] LeFusion: Controllable Pathology Synthesis via Lesion-Focused Diffusion Models☆73Updated last week
- The first Chinese medical large vision-language model designed to integrate the analysis of textual and visual data☆59Updated last year
- [NAACL 2025] From redundancy to relevance: Enhancing explainability in multimodal large language models☆82Updated last week
- [EMNLP'24] Code and data for paper "Med-MoE: Mixture of Domain-Specific Experts for Lightweight Medical Vision-Language Models"☆87Updated 2 months ago
- [ICLR'25] MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models☆109Updated 3 weeks ago
- Reverse Chain-of-Thought Problem Generation for Geometric Reasoning in Large Multimodal Models☆170Updated 3 months ago
- (AAAI 2024) BLIVA: A Simple Multimodal LLM for Better Handling of Text-rich Visual Questions☆253Updated 10 months ago
- Code for AAAl 2024 paper: Relax Image-Specific Prompt Requirement in SAM: A Single Generic Prompt for Segmenting Camouflaged Objects☆137Updated 4 months ago
- [ICLR 2025] This is the official repository of our paper "MedTrinity-25M: A Large-scale Multimodal Dataset with Multigranular Annotations…☆256Updated last month
- PMC-VQA is a large-scale medical visual question-answering dataset, which contains 227k VQA pairs of 149k images that cover various modal…☆190Updated 2 months ago
- A generalist foundation model for healthcare capable of handling diverse medical data modalities.☆63Updated 9 months ago
- [NeurIPS'24] Leveraging Hallucinations to Reduce Manual Prompt Dependency in Promptable Segmentation☆55Updated 2 months ago
- Chain-of-Spot: Interactive Reasoning Improves Large Vision-language Models☆90Updated 10 months ago
- (ECCV 2024) Empowering Multimodal Large Language Model as a Powerful Data Generator☆104Updated 4 months ago
- MedLSAM: Localize and Segment Anything Model for 3D Medical Images☆482Updated 9 months ago
- [Arxiv-2024] CheXagent: Towards a Foundation Model for Chest X-Ray Interpretation☆143Updated last month
- Dataset of paper: On the Compositional Generalization of Multimodal LLMs for Medical Imaging☆28Updated last month
- A simple-to-use yet function-rich medical image processing toolbox☆161Updated 2 years ago
- ☆60Updated 2 months ago
- The official code for "Towards Generalist Foundation Model for Radiology by Leveraging Web-scale 2D&3D Medical Data".☆372Updated 3 months ago
- [npj digital medicine] The official codes for "Towards Evaluating and Building Versatile Large Language Models for Medicine"☆54Updated 2 weeks ago