guilk / VLCLinks
Research code for "Training Vision-Language Transformers from Captions Alone"
☆33Updated 3 years ago
Alternatives and similar repositories for VLC
Users that are interested in VLC are comparing it to the libraries listed below
Sorting:
- UniTAB: Unifying Text and Box Outputs for Grounded VL Modeling, ECCV 2022 (Oral Presentation)☆89Updated 2 years ago
- [ECCV2022] Contrastive Vision-Language Pre-training with Limited Resources☆46Updated 3 years ago
- Use CLIP to represent video for Retrieval Task☆70Updated 4 years ago
- A Unified Framework for Video-Language Understanding☆61Updated 2 years ago
- ☆73Updated 3 years ago
- Code for paper "Point and Ask: Incorporating Pointing into Visual Question Answering"☆19Updated 3 years ago
- [CVPR-2023] The official dataset of Advancing Visual Grounding with Scene Knowledge: Benchmark and Method.☆33Updated 2 years ago
- Official repository for the General Robust Image Task (GRIT) Benchmark☆54Updated 2 years ago
- ☆43Updated 4 years ago
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Updated 2 years ago
- Repository for the paper "Data Efficient Masked Language Modeling for Vision and Language".☆18Updated 4 years ago
- The Pytorch implementation for "Video-Text Pre-training with Learned Regions"☆42Updated 3 years ago
- Pytorch version of DeCEMBERT: Learning from Noisy Instructional Videos via Dense Captions and Entropy Minimization (NAACL 2021)☆17Updated 3 years ago
- MLPs for Vision and Langauge Modeling (Coming Soon)☆27Updated 4 years ago
- Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training☆141Updated last month
- source code and pre-trained/fine-tuned checkpoint for NAACL 2021 paper LightningDOT☆72Updated 3 years ago
- Pytorch code for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners☆116Updated 3 years ago
- ☆32Updated 3 years ago
- A task-agnostic vision-language architecture as a step towards General Purpose Vision☆92Updated 4 years ago
- Code for the paper titled "CiT Curation in Training for Effective Vision-Language Data".☆78Updated 3 years ago
- [ACL 2023] Official PyTorch code for Singularity model in "Revealing Single Frame Bias for Video-and-Language Learning"☆136Updated 2 years ago
- [ICLR2024] Codes and Models for COSA: Concatenated Sample Pretrained Vision-Language Foundation Model☆43Updated last year
- [BMVC22] Official Implementation of ViCHA: "Efficient Vision-Language Pretraining with Visual Concepts and Hierarchical Alignment"☆55Updated 3 years ago
- ☆110Updated 3 years ago
- [ACL 2021] mTVR: Multilingual Video Moment Retrieval☆27Updated 3 years ago
- [ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos☆126Updated 2 years ago
- Data and code for NeurIPS 2021 Paper "IconQA: A New Benchmark for Abstract Diagram Understanding and Visual Language Reasoning".☆55Updated 2 years ago
- PyTorch code for "Perceiver-VL: Efficient Vision-and-Language Modeling with Iterative Latent Attention" (WACV 2023)☆33Updated 3 years ago
- [Arxiv2022] Revitalize Region Feature for Democratizing Video-Language Pre-training☆21Updated 3 years ago
- CVPR 2022 (Oral) Pytorch Code for Unsupervised Vision-and-Language Pre-training via Retrieval-based Multi-Granular Alignment☆22Updated 3 years ago