rajesh-lab / symile
Symile is a flexible, architecture-agnostic contrastive loss that enables training modality-specific representations for any number of modalities.
☆33Updated last month
Alternatives and similar repositories for symile
Users that are interested in symile are comparing it to the libraries listed below
Sorting:
- I2M2: Jointly Modeling Inter- & Intra-Modality Dependencies for Multi-modal Learning (NeurIPS 2024)☆19Updated 6 months ago
- [NeurIPS 2023] Factorized Contrastive Learning: Going Beyond Multi-view Redundancy☆66Updated last year
- m1: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Models☆26Updated last month
- Code and benchmark for the paper: "A Practitioner's Guide to Continual Multimodal Pretraining" [NeurIPS'24]☆54Updated 5 months ago
- Expert-level AI radiology report evaluator☆29Updated last month
- ☆45Updated 4 months ago
- [CVPR 2025] MicroVQA eval and 🤖RefineBot code for "MicroVQA: A Multimodal Reasoning Benchmark for Microscopy-Based Scientific Research"…☆20Updated last month
- ☆11Updated last month
- Implementation of the general framework for AMIE, from the paper "Towards Conversational Diagnostic AI", out of Google Deepmind☆62Updated 8 months ago
- Holistic evaluation of multimodal foundation models☆47Updated 9 months ago
- [ICLR 2023] MultiViz: Towards Visualizing and Understanding Multimodal Models☆97Updated 8 months ago
- ☆43Updated 7 months ago
- Unsolvable Problem Detection: Evaluating Trustworthiness of Vision Language Models☆76Updated 8 months ago
- More dimensions = More fun☆22Updated 9 months ago
- "Worse than Random? An Embarrassingly Simple Probing Evaluation of Large Multimodal Models in Medical VQA"☆17Updated 2 months ago
- X-Reasoner: Towards Generalizable Reasoning Across Modalities and Domains☆40Updated last week
- [NeurIPS 2023, ICMI 2023] Quantifying & Modeling Multimodal Interactions☆74Updated 6 months ago
- ☆43Updated 3 weeks ago
- ☆48Updated 2 months ago
- MedMax: Mixed-Modal Instruction Tuning for Training Biomedical Assistants☆32Updated this week
- [ECCV 2024] Official Release of SILC: Improving vision language pretraining with self-distillation☆43Updated 7 months ago
- Repo for the pape Benchmarking Large Language Models on Answering and Explaining Challenging Medical Questions☆35Updated 8 months ago
- ☆41Updated 9 months ago
- Evaluation and dataset construction code for the CVPR 2025 paper "Vision-Language Models Do Not Understand Negation"☆21Updated 3 weeks ago
- Official implementation of Scaling Laws in Patchification: An Image Is Worth 50,176 Tokens And More☆20Updated 2 months ago
- Official Implementation of DiffCLIP: Differential Attention Meets CLIP☆26Updated 2 months ago
- [NeurIPS'24] CARES: A Comprehensive Benchmark of Trustworthiness in Medical Vision Language Models☆68Updated 5 months ago
- MultiModN – Multimodal, Multi-Task, Interpretable Modular Networks (NeurIPS 2023)☆33Updated last year
- Official implementation for NeurIPS'23 paper "Geodesic Multi-Modal Mixup for Robust Fine-Tuning"☆33Updated 7 months ago
- [ICCV 2023] ViLLA: Fine-grained vision-language representation learning from real-world data☆44Updated last year