rajesh-lab / symileLinks
Symile is a flexible, architecture-agnostic contrastive loss that enables training modality-specific representations for any number of modalities.
☆37Updated 4 months ago
Alternatives and similar repositories for symile
Users that are interested in symile are comparing it to the libraries listed below
Sorting:
- [CVPR 2025] BIOMEDICA: An Open Biomedical Image-Caption Archive, Dataset, and Vision-Language Models Derived from Scientific Literature☆75Updated 4 months ago
- [NeurIPS 2023, ICMI 2023] Quantifying & Modeling Multimodal Interactions☆77Updated 9 months ago
- Expert-level AI radiology report evaluator☆32Updated 4 months ago
- [NeurIPS 2023] Factorized Contrastive Learning: Going Beyond Multi-view Redundancy☆70Updated last year
- BiomedCLIP data pipeline☆60Updated 6 months ago
- [CVPR 2025] MicroVQA eval and 🤖RefineBot code for "MicroVQA: A Multimodal Reasoning Benchmark for Microscopy-Based Scientific Research"…☆21Updated 3 weeks ago
- [ICLR 2023] MultiViz: Towards Visualizing and Understanding Multimodal Models☆96Updated 11 months ago
- I2M2: Jointly Modeling Inter- & Intra-Modality Dependencies for Multi-modal Learning (NeurIPS 2024)☆21Updated 9 months ago
- Med-PRM: Medical Reasoning Models with Stepwise, Guideline-verified Process Rewards☆39Updated 3 weeks ago
- [ACL 2025 Findings] "Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"☆22Updated 5 months ago
- ☆30Updated 9 months ago
- MedMax: Mixed-Modal Instruction Tuning for Training Biomedical Assistants☆36Updated 2 months ago
- The official code to build up dataset PMC-OA☆32Updated last year
- A new collection of medical VQA dataset based on MIMIC-CXR. Part of the work 'EHRXQA: A Multi-Modal Question Answering Dataset for Electr…☆87Updated 11 months ago
- LLaVa Version of RaDialog☆21Updated 2 months ago
- MultiModN – Multimodal, Multi-Task, Interpretable Modular Networks (NeurIPS 2023)☆33Updated last year
- Code for the paper "Explain Any Concept: Segment Anything Meets Concept-Based Explanation". Poster @ NeurIPS 2023☆44Updated last year
- [NeurIPS'24] CARES: A Comprehensive Benchmark of Trustworthiness in Medical Vision Language Models☆74Updated 8 months ago
- More dimensions = More fun☆22Updated last year
- [Arxiv-2024] CheXagent: Towards a Foundation Model for Chest X-Ray Interpretation☆184Updated 6 months ago
- INSPECT dataset/benchmark paper, accepted by NeurIPS 2023☆36Updated 2 months ago
- A multi-modal CLIP model trained on the medical dataset ROCO☆141Updated 2 months ago
- EHRXQA: A Multi-Modal Question Answering Dataset for Electronic Health Records with Chest X-ray Images, NeurIPS 2023 D&B☆82Updated last year
- Official repository of paper titled "UniMed-CLIP: Towards a Unified Image-Text Pretraining Paradigm for Diverse Medical Imaging Modalitie…☆120Updated 3 months ago
- ☆22Updated 2 months ago
- MedAgentsBench: Benchmarking Thinking Models and Agent Frameworks for Complex Medical Reasoning https://arxiv.org/abs/2503.07459☆50Updated last month
- ☆37Updated last year
- [IPCAI'24 Best Paper] Advancing Surgical VQA with Scene Graph Knowledge☆44Updated 2 months ago
- [CVPR 2025] CheXWorld: Exploring Image World Modeling for Radiograph Representation Learning☆19Updated 3 months ago
- ☆66Updated last month