wysoczanska / clip_dinoiserView external linksLinks
Official implementation of 'CLIP-DINOiser: Teaching CLIP a few DINO tricks' paper.
☆274Oct 26, 2024Updated last year
Alternatives and similar repositories for clip_dinoiser
Users that are interested in clip_dinoiser are comparing it to the libraries listed below
Sorting:
- [CVPR24] Official Implementation of GEM (Grounding Everything Module)☆136Apr 10, 2025Updated 10 months ago
- Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference☆180Oct 10, 2024Updated last year
- Official implementation of the WACV 2024 paper CLIP-DIY☆34Dec 20, 2023Updated 2 years ago
- [ECCV'24] Official PyTorch implementation of In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic Segmentation☆49Sep 24, 2024Updated last year
- [NeurIPS 2023] This repo contains the code for our paper Convolutions Die Hard: Open-Vocabulary Segmentation with Single Frozen Convoluti…☆337Feb 5, 2024Updated 2 years ago
- Code and datasets for "Text encoders are performance bottlenecks in contrastive vision-language models". Coming soon!☆11May 24, 2023Updated 2 years ago
- PyTorch Implementation of NACLIP in "Pay Attention to Your Neighbours: Training-Free Open-Vocabulary Semantic Segmentation"☆72Sep 23, 2024Updated last year
- Official PyTorch implementation of "Extract Free Dense Labels from CLIP" (ECCV 22 Oral)☆470Sep 19, 2022Updated 3 years ago
- [CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want☆866Jul 20, 2025Updated 6 months ago
- Official Implementation of "CAT-Seg🐱: Cost Aggregation for Open-Vocabulary Semantic Segmentation"☆362Apr 11, 2024Updated last year
- Pytorch code for paper From CLIP to DINO: Visual Encoders Shout in Multi-modal Large Language Models☆206Jan 8, 2025Updated last year
- A curated publication list on open vocabulary semantic segmentation and related area (e.g. zero-shot semantic segmentation) resources..☆828Jan 20, 2026Updated 3 weeks ago
- This is the official code release for our work, Denoising Vision Transformers.☆393Nov 13, 2024Updated last year
- [ECCV 2024] The official code of paper "Open-Vocabulary SAM".☆1,028Aug 4, 2025Updated 6 months ago
- [ECCV2024] ProxyCLIP: Proxy Attention Improves CLIP for Open-Vocabulary Segmentation☆111Mar 26, 2025Updated 10 months ago
- [CVPR 2024 🔥] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses tha…☆945Aug 5, 2025Updated 6 months ago
- Code for experiments for "ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet Accuracy"☆102Sep 11, 2024Updated last year
- Official PyTorch Implementation of Efficient and Versatile Robust Fine-Tuning of Zero-shot Models, ECCV 2024☆17Oct 3, 2024Updated last year
- [CVPR 2024] Official implementation of the paper "Visual In-context Learning"☆529Apr 8, 2024Updated last year
- Code and datasets for "What’s “up” with vision-language models? Investigating their struggle with spatial reasoning".☆70Feb 28, 2024Updated last year
- Code for "Are “Hierarchical” Visual Representations Hierarchical?" in NeurIPS Workshop for Symmetry and Geometry in Neural Representation…☆22Nov 8, 2023Updated 2 years ago
- [NeurIPS 2024] Official implementation of the paper "Interfacing Foundation Models' Embeddings"☆129Aug 21, 2024Updated last year
- When do we not need larger vision models?☆412Feb 8, 2025Updated last year
- ☆24Apr 17, 2024Updated last year
- Official Repo for PosSAM: Panoptic Open-vocabulary Segment Anything☆70Apr 7, 2024Updated last year
- (TPAMI 2024) A Survey on Open Vocabulary Learning☆987Dec 24, 2025Updated last month
- [ICLR2024 Spotlight] Code Release of CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction☆201Feb 5, 2024Updated 2 years ago
- (ICCV 2023) MasQCLIP for Open-Vocabulary Universal Image Segmentation☆37Oct 18, 2023Updated 2 years ago
- Open-vocabulary Semantic Segmentation☆372Oct 16, 2024Updated last year
- Official Implementation for CVPR 2024 paper: CLIP as RNN: Segment Countless Visual Concepts without Training Endeavor☆110Jun 23, 2024Updated last year
- Official Pytorch Implementation of Paper "A Semantic Space is Worth 256 Language Descriptions: Make Stronger Segmentation Models with Des…☆55Aug 27, 2025Updated 5 months ago
- [ECCV2024] ClearCLIP: Decomposing CLIP Representations for Dense Vision-Language Inference☆97Mar 26, 2025Updated 10 months ago
- [Pattern Recognition 25] CLIP Surgery for Better Explainability with Enhancement in Open-Vocabulary Tasks☆463Mar 1, 2025Updated 11 months ago
- PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models☆261Aug 5, 2025Updated 6 months ago
- [ECCV 2024] official code for "Long-CLIP: Unlocking the Long-Text Capability of CLIP"☆889Aug 13, 2024Updated last year
- [ICCV2025] Harnessing CLIP, DINO and SAM for Open Vocabulary Segmentation☆106Nov 22, 2025Updated 2 months ago
- Code Release for MaskCLIP (ICML 2023)☆76Nov 29, 2023Updated 2 years ago
- Official implementation of TagAlign☆35Dec 11, 2024Updated last year
- LaFTer: Label-Free Tuning of Zero-shot Classifier using Language and Unlabeled Image Collections (NeurIPS 2023)☆29Dec 27, 2023Updated 2 years ago