rajesh-lab / symileLinks
Symile is a flexible, architecture-agnostic contrastive loss that enables training modality-specific representations for any number of modalities.
β33Updated 2 months ago
Alternatives and similar repositories for symile
Users that are interested in symile are comparing it to the libraries listed below
Sorting:
- Expert-level AI radiology report evaluatorβ30Updated 2 months ago
- [CVPR 2025] MicroVQA eval and π€RefineBot code for "MicroVQA: A Multimodal Reasoning Benchmark for Microscopy-Based Scientific Research"β¦β20Updated 2 months ago
- I2M2: Jointly Modeling Inter- & Intra-Modality Dependencies for Multi-modal Learning (NeurIPS 2024)β19Updated 7 months ago
- Holistic evaluation of multimodal foundation modelsβ47Updated 9 months ago
- MedMax: Mixed-Modal Instruction Tuning for Training Biomedical Assistantsβ35Updated 3 weeks ago
- m1: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Modelsβ28Updated last month
- "Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"β17Updated 3 months ago
- β43Updated 8 months ago
- More dimensions = More funβ22Updated 10 months ago
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domainsβ43Updated 3 weeks ago
- MedAgentsBench: Benchmarking Thinking Models and Agent Frameworks for Complex Medical Reasoningβ39Updated last month
- BiomedCLIP data pipelineβ46Updated 4 months ago
- [NeurIPS 2023, ICMI 2023] Quantifying & Modeling Multimodal Interactionsβ75Updated 7 months ago
- β11Updated 2 months ago
- [NeurIPS 2023] Factorized Contrastive Learning: Going Beyond Multi-view Redundancyβ67Updated last year
- Code for Principal Masked Autoencodersβ27Updated 2 months ago
- Official Implementation of DiffCLIP: Differential Attention Meets CLIPβ30Updated 2 months ago
- [NeurIPS 2024] Official implementation of the paper "MambaLRP: Explaining Selective State Space Sequence Models".β38Updated 7 months ago
- β50Updated 4 months ago
- Repo for the pape Benchmarking Large Language Models on Answering and Explaining Challenging Medical Questionsβ38Updated 9 months ago
- [ICLR 2023] MultiViz: Towards Visualizing and Understanding Multimodal Modelsβ96Updated 9 months ago
- Custom Open CLIP repo to train biomedical CLIP modelsβ17Updated 2 months ago
- β48Updated 3 months ago
- Code and benchmark for the paper: "A Practitioner's Guide to Continual Multimodal Pretraining" [NeurIPS'24]β56Updated 5 months ago
- Distributed Optimization Infra for learning CLIP modelsβ26Updated 8 months ago
- Official implementation of LLaVa-Rad, a small multimodal model for chest X-ray findings generation.β26Updated last week
- Evaluation and dataset construction code for the CVPR 2025 paper "Vision-Language Models Do Not Understand Negation"β23Updated last month
- [CVPR 2025] BIOMEDICA: An Open Biomedical Image-Caption Archive, Dataset, and Vision-Language Models Derived from Scientific Literatureβ59Updated 2 months ago
- A Contrastive Learning Boost from Intermediate Pre-Trained Representationsβ41Updated 8 months ago
- β41Updated 4 months ago