3cology / dinov2_with_attention_extractionLinks
PyTorch code and models for the DINOv2 self-supervised learning method.
☆12Updated last year
Alternatives and similar repositories for dinov2_with_attention_extraction
Users that are interested in dinov2_with_attention_extraction are comparing it to the libraries listed below
Sorting:
- ICCV 2023: CLIPN for Zero-Shot OOD Detection: Teaching CLIP to Say No☆139Updated last year
- [ICLR 2024] Test-Time Adaptation with CLIP Reward for Zero-Shot Generalization in Vision-Language Models.☆83Updated 11 months ago
- official implementation of "Interpreting CLIP's Image Representation via Text-Based Decomposition"☆216Updated last month
- [ICCV'23 Main Track, WECIA'23 Oral] Official repository of paper titled "Self-regulating Prompts: Foundational Model Adaptation without F…☆268Updated last year
- Awesome List of Vision Language Prompt Papers☆46Updated last year
- [NeurIPS 2023] Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization☆106Updated last year
- [CVPR 2023] CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation☆198Updated 10 months ago
- PyTorch Implementation of NACLIP in "Pay Attention to Your Neighbours: Training-Free Open-Vocabulary Semantic Segmentation"☆58Updated 9 months ago
- [ICLR2023] PLOT: Prompt Learning with Optimal Transport for Vision-Language Models☆167Updated last year
- Official implementation of "Why are Visually-Grounded Language Models Bad at Image Classification?" (NeurIPS 2024)☆86Updated 9 months ago
- ☆34Updated last year
- Implementation of "VL-Mamba: Exploring State Space Models for Multimodal Learning"☆81Updated last year
- [ICLR 2023] Official code repository for "Meta Learning to Bridge Vision and Language Models for Multimodal Few-Shot Learning"☆59Updated 2 years ago
- Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey [Miyai+, TMLR2025]☆91Updated last month
- [Pattern Recognition 25] CLIP Surgery for Better Explainability with Enhancement in Open-Vocabulary Tasks☆432Updated 4 months ago
- Pytorch Implementation for CVPR 2024 paper: Learn to Rectify the Bias of CLIP for Unsupervised Semantic Segmentation☆48Updated this week
- PyTorch implementation of ICML 2023 paper "SegCLIP: Patch Aggregation with Learnable Centers for Open-Vocabulary Semantic Segmentation"☆93Updated 2 years ago
- ☆51Updated last year
- Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference☆158Updated 9 months ago
- An easy way to apply LoRA to CLIP. Implementation of the paper "Low-Rank Few-Shot Adaptation of Vision-Language Models" (CLIP-LoRA) [CVPR…☆221Updated last month
- [AAAI 2024] TagCLIP: A Local-to-Global Framework to Enhance Open-Vocabulary Multi-Label Classification of CLIP Without Training☆98Updated last year
- [AAAI'25, CVPRW 2024] Official repository of paper titled "Learning to Prompt with Text Only Supervision for Vision-Language Models".☆108Updated 7 months ago
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆73Updated 2 months ago
- LLM-Seg: Bridging Image Segmentation and Large Language Model Reasoning☆164Updated last year
- [CVPR2024] GSVA: Generalized Segmentation via Multimodal Large Language Models☆137Updated 10 months ago
- [CVPR24] Official Implementation of GEM (Grounding Everything Module)☆126Updated 3 months ago
- [ICLR 2025] VL-ICL Bench: The Devil in the Details of Multimodal In-Context Learning☆61Updated 5 months ago
- PyTorch implementation of MCM (Delving into out-of-distribution detection with vision-language representations), NeurIPS 2022☆85Updated last year
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆54Updated last year
- [CVPR 2024] Improving language-visual pretraining efficiency by perform cluster-based masking on images.☆28Updated last year