TalalWasim / Vita-CLIPLinks
Official repository for "Vita-CLIP: Video and text adaptive CLIP via Multimodal Prompting" [CVPR 2023]
☆125Updated 2 years ago
Alternatives and similar repositories for Vita-CLIP
Users that are interested in Vita-CLIP are comparing it to the libraries listed below
Sorting:
- ☆84Updated 2 years ago
- ☆41Updated last year
- ☆30Updated 2 years ago
- [ICLR 2024] FROSTER: Frozen CLIP is a Strong Teacher for Open-Vocabulary Action Recognition☆91Updated 9 months ago
- [CVPR 2024] TeachCLIP for Text-to-Video Retrieval☆40Updated 5 months ago
- [CVPR 2024] Do you remember? Dense Video Captioning with Cross-Modal Memory Retrieval☆60Updated last year
- [CVPR 2023 Highlight & TPAMI] Video-Text as Game Players: Hierarchical Banzhaf Interaction for Cross-Modal Representation Learning☆121Updated 9 months ago
- ICCV2023: Disentangling Spatial and Temporal Learning for Efficient Image-to-Video Transfer Learning☆41Updated 2 years ago
- [ECCV 2022] A pytorch implementation for TS2-Net: Token Shift and Selection Transformer for Text-Video Retrieval☆77Updated 2 years ago
- Pytorch Code for "Unified Coarse-to-Fine Alignment for Video-Text Retrieval" (ICCV 2023)☆66Updated last year
- ☆37Updated 3 years ago
- ☆119Updated last year
- CVPR 2023 Accepted Paper HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models☆69Updated last year
- ☆190Updated 3 years ago
- [CVPR 2024] Context-Guided Spatio-Temporal Video Grounding☆59Updated last year
- ☆48Updated 2 years ago
- Official implementation of "Test-Time Zero-Shot Temporal Action Localization", CVPR 2024☆66Updated last year
- Composed Video Retrieval☆61Updated last year
- [CVPR2023] The code for 《Position-guided Text Prompt for Vision-Language Pre-training》☆151Updated 2 years ago
- [NeurIPS 2022] Embracing Consistency: A One-Stage Approach for Spatio-Temporal Video Grounding☆52Updated last year
- ☆94Updated 2 years ago
- Official implementation of "Everything at Once - Multi-modal Fusion Transformer for Video Retrieval." CVPR 2022☆112Updated 3 years ago
- Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)☆82Updated last year
- Official implementation of "ConZIC: Controllable Zero-shot Image Captioning by Sampling-Based Polishing"☆74Updated 2 years ago
- ☆62Updated 2 years ago
- ICLR 2023 DeCap: Decoding CLIP Latents for Zero-shot Captioning☆137Updated 2 years ago
- Code for the paper: "SuS-X: Training-Free Name-Only Transfer of Vision-Language Models" [ICCV'23]☆105Updated 2 years ago
- (CVPR2024) MeaCap: Memory-Augmented Zero-shot Image Captioning☆51Updated last year
- 【ICLR 2024, Spotlight】Sentence-level Prompts Benefit Composed Image Retrieval☆89Updated last year
- https://layer6ai-labs.github.io/xpool/☆129Updated 2 years ago