☆57Aug 16, 2025Updated 6 months ago
Alternatives and similar repositories for prolip
Users that are interested in prolip are comparing it to the libraries listed below
Sorting:
- [ECCV 2024] Official PyTorch implementation of "HYPE: Hyperbolic Entailment Filtering for Underspecified Images and Texts"☆20Nov 22, 2024Updated last year
- Official Pytorch implementation of "Improved Probabilistic Image-Text Representations" (ICLR 2024)☆60May 26, 2024Updated last year
- [NeurIPS 2024] Official PyTorch implementation of "Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives"☆46Dec 1, 2024Updated last year
- Code and data for the paper "Emergent Visual-Semantic Hierarchies in Image-Text Representations" (ECCV 2024)☆34Aug 12, 2024Updated last year
- [ECCV 2024] Official PyTorch implementation of LUT "Learning with Unmasked Tokens Drives Stronger Vision Learners"☆13Dec 1, 2024Updated last year
- Model Stock: All we need is just a few fine-tuned models☆129Aug 9, 2025Updated 6 months ago
- Official Implementation of FedRCL (CVPR 2024)☆24Jun 6, 2024Updated last year
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆21Oct 8, 2024Updated last year
- [ECCV2024][ICCV2023] Official PyTorch implementation of SeiT++ and SeiT☆56Aug 12, 2024Updated last year
- This repository is related to 'Intriguing Properties of Hyperbolic Embeddings in Vision-Language Models', published at TMLR (2024), https…☆22Jul 5, 2024Updated last year
- Code for "CLIP Behaves like a Bag-of-Words Model Cross-modally but not Uni-modally"☆20Feb 27, 2026Updated last week
- ☆20Aug 19, 2024Updated last year
- [ECCV2024] Official implementation of paper, "DenseNets Reloaded: Paradigm Shift Beyond ResNets and ViTs".☆156Aug 8, 2024Updated last year
- Official PyTorch implementation of ReWaS (AAAI'25) "Read, Watch and Scream! Sound Generation from Text and Video"☆43Dec 13, 2024Updated last year
- ☆27Jun 4, 2024Updated last year
- [ICML 2025] Official PyTorch implementation of "NegMerge: Sign-Consensual Weight Merging for Machine Unlearning"☆14Nov 25, 2025Updated 3 months ago
- [ICLR'25] Official repository for "AVHBench: A Cross-Modal Hallucination Evaluation for Audio-Visual Large Language Models"☆20Feb 25, 2026Updated last week
- ☆11May 17, 2024Updated last year
- [ICLR 2025] Official PyTorch implementation of "DaWin: Training-free Dynamic Weight Interpolation for Robust Adaptation"☆26Jul 11, 2025Updated 7 months ago
- Official Pytorch implementation of "Probabilistic Cross-Modal Embedding" (CVPR 2021)☆138Mar 1, 2024Updated 2 years ago
- (NeurIPS 2024) What Makes CLIP More Robust to Long-Tailed Pre-Training Data? A Controlled Study for Transferable Insights☆27Oct 28, 2024Updated last year
- Official implementation of BPA (CVPR 2022)☆13Jun 17, 2022Updated 3 years ago
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"☆18Mar 15, 2024Updated last year
- Official PyTorch implementation of “MaskRIS: Semantic Distortion-aware Data Augmentation for Referring Image Segmentation”☆18Dec 5, 2024Updated last year
- ☆11Nov 5, 2024Updated last year
- [ICCV 2023] Going Beyond Nouns With Vision & Language Models Using Synthetic Data☆14Sep 30, 2023Updated 2 years ago
- ☆11Oct 2, 2024Updated last year
- [ICLR 2024 Spotlight] "Negative Label Guided OOD Detection with Pretrained Vision-Language Models"☆30Oct 23, 2024Updated last year
- Training code for CLIP-FlanT5☆30Jul 29, 2024Updated last year
- Code and benchmark for the paper: "A Practitioner's Guide to Continual Multimodal Pretraining" [NeurIPS'24]☆62Dec 10, 2024Updated last year
- Learning Debiased and Disentangled Representations for Semantic Segmentation (NeurIPS 2021)☆13Jan 23, 2022Updated 4 years ago
- ☆54Jan 17, 2025Updated last year
- ☆16Jan 20, 2026Updated last month
- ☆13Feb 25, 2025Updated last year
- ☆17Nov 29, 2024Updated last year
- Learning Features with Parameter-Free Layers, ICLR 2022☆84May 3, 2023Updated 2 years ago
- This is the official implementation of the paper "MM-SHAP: A Performance-agnostic Metric for Measuring Multimodal Contributions in Vision…☆32Mar 12, 2024Updated last year
- [ICLR 2025] Official code repository for "TULIP: Token-length Upgraded CLIP"☆33Jan 26, 2026Updated last month
- Code and data for the paper: Learning Action and Reasoning-Centric Image Editing from Videos and Simulation☆35Jun 30, 2025Updated 8 months ago