sheng-eatamath / S3A
repo for paper titled: Towards Realistic Zero-Shot Classification via Self Structural Semantic Alignment (AAAI'24 Oral)
☆25Updated 8 months ago
Alternatives and similar repositories for S3A:
Users that are interested in S3A are comparing it to the libraries listed below
- Compress conventional Vision-Language Pre-training data☆49Updated last year
- Official code for the paper, "TaCA: Upgrading Your Visual Foundation Model with Task-agnostic Compatible Adapter".☆16Updated last year
- Code release for "Understanding Bias in Large-Scale Visual Datasets"☆15Updated last month
- ☆29Updated last month
- ☆22Updated 7 months ago
- LaFTer: Label-Free Tuning of Zero-shot Classifier using Language and Unlabeled Image Collections (NeurIPS 2023)☆27Updated last year
- Generating Image Specific Text☆26Updated last year
- ☆25Updated last year
- [NeurIPS 2023] Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization☆103Updated 11 months ago
- SVL-Adapter: Self-Supervised Adapter for Vision-Language Pretrained Models☆20Updated last year
- Official Pytorch codebase for Open-Vocabulary Instance Segmentation without Manual Mask Annotations [CVPR 2023]☆49Updated last week
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated last month
- [ECCV-2022]Grounding Visual Representations with Texts for Domain Generalization☆31Updated last year
- Code for Label Propagation for Zero-shot Classification with Vision-Language Models (CVPR2024)☆35Updated 5 months ago
- Distribution-Aware Prompt Tuning for Vision-Language Models (ICCV 2023)☆38Updated last year
- Official code for "Understanding and Mitigating Overfitting in Prompt Tuning for Vision-Language Models" (TCSVT'2023)☆27Updated last year
- Task Residual for Tuning Vision-Language Models (CVPR 2023)☆68Updated last year
- The Curious Layperson: Fine-Grained Image Recognition without Expert Labels (BMVC 2021 best student paper)☆23Updated 2 years ago
- Official code for "Disentangling Visual Embeddings for Attributes and Objects" Published at CVPR 2022☆35Updated last year
- [ICLR 2024] Test-Time Adaptation with CLIP Reward for Zero-Shot Generalization in Vision-Language Models.☆65Updated 5 months ago
- Official Pytorch implementation of 'Facing the Elephant in the Room: Visual Prompt Tuning or Full Finetuning'? (ICLR2024)☆11Updated 10 months ago
- ☆59Updated 2 years ago
- Repository for the paper: Teaching VLMs to Localize Specific Objects from In-context Examples☆19Updated last month
- Perceptual Grouping in Contrastive Vision-Language Models (ICCV'23)☆37Updated last year
- 📍 Official pytorch implementation of paper "ProtoCLIP: Prototypical Contrastive Language Image Pretraining" (IEEE TNNLS)☆52Updated last year
- Code Release of F-LMM: Grounding Frozen Large Multimodal Models☆60Updated 5 months ago
- This repository contains the code and datasets for our ICCV-W paper 'Enhancing CLIP with GPT-4: Harnessing Visual Descriptions as Prompts…☆27Updated 10 months ago
- [CVPR'24] Validation-free few-shot adaptation of CLIP, using a well-initialized Linear Probe (ZSLP) and class-adaptive constraints (CLAP)…☆63Updated 7 months ago
- ☆22Updated last year