neil-ab / clip-zs-promptingLinks
Using CLIP for zero-shot learning and image classification with text & visual prompting.
☆16Updated 3 years ago
Alternatives and similar repositories for clip-zs-prompting
Users that are interested in clip-zs-prompting are comparing it to the libraries listed below
Sorting:
- code for studying OpenAI's CLIP explainability☆38Updated 4 years ago
- [NeurIPS2023] Parameter-efficient Tuning of Large-scale Multimodal Foundation Model☆89Updated 2 years ago
- [NeurIPS 2023] Align Your Prompts: Test-Time Prompting with Distribution Alignment for Zero-Shot Generalization☆110Updated last year
- [ICLR 2023] Official code repository for "Meta Learning to Bridge Vision and Language Models for Multimodal Few-Shot Learning"☆60Updated 2 years ago
- The official implementation of paper Dual Modality Prompt Tuning for Vision-Language Pre-Trained Model. If you find our code or paper use…☆50Updated 2 years ago
- Awesome List of Vision Language Prompt Papers☆46Updated 2 years ago
- [ICLR'24] Consistency-guided Prompt Learning for Vision-Language Models☆85Updated last year
- [NeurIPS 2023] Meta-Adapter☆48Updated 2 years ago
- The efficient tuning method for VLMs☆80Updated last year
- [ICLR2023] PLOT: Prompt Learning with Optimal Transport for Vision-Language Models☆174Updated 2 years ago
- [ACLW'24] LMPT: Prompt Tuning with Class-Specific Embedding Loss for Long-tailed Multi-Label Visual Recognition☆57Updated last year
- Code and results accompanying our paper titled CHiLS: Zero-Shot Image Classification with Hierarchical Label Sets☆59Updated 2 years ago
- code for "Multitask Vision-Language Prompt Tuning" https://arxiv.org/abs/2211.11720☆56Updated last year
- [Paper][AAAI2024]Structure-CLIP: Towards Scene Graph Knowledge to Enhance Multi-modal Structured Representations☆152Updated last year
- This repo is the official implementation of UPL (Unsupervised Prompt Learning for Vision-Language Models).☆118Updated 3 years ago
- Task Residual for Tuning Vision-Language Models (CVPR 2023)☆75Updated 2 years ago
- MixGen: A New Multi-Modal Data Augmentation☆126Updated 3 years ago
- [NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"☆182Updated last year
- Visual self-questioning for large vision-language assistant.☆45Updated 6 months ago
- [NeurIPS 2023] Text data, code and pre-trained models for paper "Improving CLIP Training with Language Rewrites"☆287Updated 2 years ago
- ☆200Updated 2 years ago
- Distilling Large Vision-Language Model with Out-of-Distribution Generalizability (ICCV 2023)☆60Updated last year
- [ICCV2023] Official code for "VL-PET: Vision-and-Language Parameter-Efficient Tuning via Granularity Control"☆53Updated 2 years ago
- [CVPR 2024] Offical implemention of the paper "DePT: Decoupled Prompt Tuning"☆109Updated 2 months ago
- Generalized Out-of-Distribution Detection and Beyond in Vision Language Model Era: A Survey [Miyai+, TMLR2025]☆98Updated 7 months ago
- FInetuning CLIP for Few Shot Learning☆46Updated 4 years ago
- Distribution-Aware Prompt Tuning for Vision-Language Models (ICCV 2023)☆44Updated 2 years ago
- [TPAMI] Searching prompt modules for parameter-efficient transfer learning.☆238Updated 2 years ago
- [ICCV 2023] Prompt-aligned Gradient for Prompt Tuning☆167Updated 2 years ago
- [CVPR 2024] PriViLege: Pre-trained Vision and Language Transformers Are Few-Shot Incremental Learners☆55Updated last year