rajesh-lab / symileLinks
Symile is a flexible, architecture-agnostic contrastive loss that enables training modality-specific representations for any number of modalities.
☆46Updated 9 months ago
Alternatives and similar repositories for symile
Users that are interested in symile are comparing it to the libraries listed below
Sorting:
- Expert-level AI radiology report evaluator☆35Updated 8 months ago
- [ICLR 2023] MultiViz: Towards Visualizing and Understanding Multimodal Models☆98Updated last year
- I2M2: Jointly Modeling Inter- & Intra-Modality Dependencies for Multi-modal Learning (NeurIPS 2024)☆22Updated last year
- [CVPR 2025] BIOMEDICA: An Open Biomedical Image-Caption Archive, Dataset, and Vision-Language Models Derived from Scientific Literature☆87Updated 9 months ago
- [EMNLP 2025] Med-PRM: Medical Reasoning Models with Stepwise, Guideline-verified Process Rewards☆54Updated 3 months ago
- [CVPR 2025] MicroVQA eval and 🤖RefineBot code for "MicroVQA: A Multimodal Reasoning Benchmark for Microscopy-Based Scientific Research"…☆31Updated last month
- [NeurIPS 2023, ICMI 2023] Quantifying & Modeling Multimodal Interactions☆84Updated last year
- ☆70Updated 5 months ago
- INSPECT dataset/benchmark paper, accepted by NeurIPS 2023☆42Updated 7 months ago
- [CVPR 2025] CheXWorld: Exploring Image World Modeling for Radiograph Representation Learning☆33Updated 8 months ago
- BiomedCLIP data pipeline☆97Updated 11 months ago
- [Arxiv-2024] CheXagent: Towards a Foundation Model for Chest X-Ray Interpretation☆208Updated 11 months ago
- Official repository of paper titled "UniMed-CLIP: Towards a Unified Image-Text Pretraining Paradigm for Diverse Medical Imaging Modalitie…☆149Updated 8 months ago
- LLaVa Version of RaDialog☆24Updated 7 months ago
- [NeurIPS 2023] Factorized Contrastive Learning: Going Beyond Multi-view Redundancy☆74Updated 2 years ago
- MedMax: Mixed-Modal Instruction Tuning for Training Biomedical Assistants☆41Updated 3 months ago
- ☆31Updated last year
- The dataset and evaluation code for MediConfusion: Can you trust your AI radiologist? Probing the reliability of multimodal medical found…☆23Updated last month
- ☆43Updated last year
- A new collection of medical VQA dataset based on MIMIC-CXR. Part of the work 'EHRXQA: A Multi-Modal Question Answering Dataset for Electr…☆94Updated last year
- Code for CheXlocalize☆37Updated 2 years ago
- ☆64Updated 4 months ago
- [NeurIPS'24] CARES: A Comprehensive Benchmark of Trustworthiness in Medical Vision Language Models☆77Updated last year
- [ICML'25] MMedPO: Aligning Medical Vision-Language Models with Clinical-Aware Multimodal Preference Optimization☆64Updated 6 months ago
- ☆98Updated last year
- [ACL 2025 Findings] "Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"☆24Updated 10 months ago
- EHRXQA: A Multi-Modal Question Answering Dataset for Electronic Health Records with Chest X-ray Images, NeurIPS 2023 D&B☆87Updated last year
- The official code to build up dataset PMC-OA☆33Updated last year
- ☆25Updated 2 years ago
- BenchX: A Unified Benchmark Framework for Medical Vision-Language Pretraining on Chest X-Rays☆44Updated 7 months ago