Yutong-Zhou-cv / Awesome-Survey-PapersLinks
A curated list of Survey Papers on Deep Learning.
☆12Updated last year
Alternatives and similar repositories for Awesome-Survey-Papers
Users that are interested in Awesome-Survey-Papers are comparing it to the libraries listed below
Sorting:
- Masked Vision-Language Transformer in Fashion☆33Updated last year
- ☆26Updated last year
- Implementation of the model: "(MC-ViT)" from the paper: "Memory Consolidation Enables Long-Context Video Understanding"☆20Updated 2 months ago
- [PR 2024] A large Cross-Modal Video Retrieval Dataset with Reading Comprehension☆26Updated last year
- Official repository for the General Robust Image Task (GRIT) Benchmark☆54Updated 2 years ago
- Code for CVPR 2023 paper "SViTT: Temporal Learning of Sparse Video-Text Transformers"☆18Updated 2 years ago
- [WACV2025 Oral] DeepMIM: Deep Supervision for Masked Image Modeling☆53Updated last month
- Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"☆35Updated 10 months ago
- ECCV2024_Parrot Captions Teach CLIP to Spot Text☆66Updated 9 months ago
- Official Pytorch Implementation of Self-emerging Token Labeling☆34Updated last year
- OLA-VLM: Elevating Visual Perception in Multimodal LLMs with Auxiliary Embedding Distillation, arXiv 2024☆60Updated 4 months ago
- ☆29Updated 2 years ago
- ☆19Updated last month
- Code and Models for "GeneCIS A Benchmark for General Conditional Image Similarity"☆59Updated 2 years ago
- [ECCV 2024] This is the official implementation of "Stitched ViTs are Flexible Vision Backbones".☆27Updated last year
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆37Updated last year
- ☆50Updated 5 months ago
- ☆58Updated last year
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆39Updated 3 months ago
- [CVPR-2023] The official dataset of Advancing Visual Grounding with Scene Knowledge: Benchmark and Method.☆31Updated last year
- ☆14Updated last year
- How Good is Google Bard's Visual Understanding? An Empirical Study on Open Challenges☆30Updated last year
- A curated list of video-text datasets in a variety of languages. These datasets can be used for video captioning (video description) or v…☆36Updated last year
- ☆59Updated last year
- Code for paper: Unified Text-to-Image Generation and Retrieval☆15Updated 11 months ago
- Code for the paper titled "CiT Curation in Training for Effective Vision-Language Data".☆78Updated 2 years ago
- ☆34Updated last year
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 7 months ago
- LAVIS - A One-stop Library for Language-Vision Intelligence☆48Updated 10 months ago
- [ICLR 23] Contrastive Aligned of Vision to Language Through Parameter-Efficient Transfer Learning☆39Updated last year