facebookresearch / connect-caption-and-trace
A unified framework to jointly model images, text, and human attention traces.
☆78Updated 3 years ago
Alternatives and similar repositories for connect-caption-and-trace:
Users that are interested in connect-caption-and-trace are comparing it to the libraries listed below
- CLIP-It! Language-Guided Video Summarization☆73Updated 3 years ago
- Localized Narratives☆82Updated 3 years ago
- PyTorch code for: Learning to Generate Grounded Visual Captions without Localization Supervision☆44Updated 4 years ago
- Data and code for CVPR 2020 paper: "VIOLIN: A Large-Scale Dataset for Video-and-Language Inference"☆159Updated 4 years ago
- source code and pre-trained/fine-tuned checkpoint for NAACL 2021 paper LightningDOT☆73Updated 2 years ago
- [ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos☆118Updated last year
- Dense video captioning in PyTorch☆41Updated 5 years ago
- A length-controllable and non-autoregressive image captioning model.☆68Updated 3 years ago
- L-Verse: Bidirectional Generation Between Image and Text☆108Updated 2 years ago
- Learning phrase grounding from captioned images through InfoNCE bound on mutual information☆72Updated 4 years ago
- ☆73Updated 2 years ago
- [ACL 2020] PyTorch code for TVQA+: Spatio-Temporal Grounding for Video Question Answering☆127Updated 2 years ago
- Data Release for VALUE Benchmark☆31Updated 3 years ago
- ☆34Updated 5 years ago
- Code and Resources for the Transformer Encoder Reasoning Network (TERN) - https://arxiv.org/abs/2004.09144☆57Updated last year
- [CVPR 2020] Transform and Tell: Entity-Aware News Image Captioning☆90Updated 10 months ago
- [CVPR20] Video Object Grounding using Semantic Roles in Language Description (https://arxiv.org/abs/2003.10606)☆67Updated 4 years ago
- Code of Dense Relational Captioning☆69Updated last year
- A PyTorch implementation of VIOLET☆137Updated last year
- Starter Code for VALUE benchmark☆80Updated 2 years ago
- [ACL 2020] PyTorch code for MART: Memory-Augmented Recurrent Transformer for Coherent Video Paragraph Captioning☆169Updated 4 years ago
- We present a framework for training multi-modal deep learning models on unlabelled video data by forcing the network to learn invariances…☆46Updated 3 years ago
- Use CLIP to represent video for Retrieval Task☆69Updated 3 years ago
- A Dataset for Grounded Video Description☆160Updated 3 years ago
- A one-stop shop for YouCook2 info such as leaderboard and recent advances on (cooking) video retrieval and captioning.☆38Updated 2 years ago
- Implementation of STAM (Space Time Attention Model), a pure and simple attention model that reaches SOTA for video classification☆131Updated 3 years ago
- Project page for "Visual Grounding in Video for Unsupervised Word Translation" CVPR 2020☆42Updated 4 years ago
- ☆31Updated 6 years ago
- ☆26Updated 3 years ago
- ☆62Updated 3 years ago