aimagelab / speakseeLinks
PyTorch library for Visual-Semantic tasks
☆29Updated 2 years ago
Alternatives and similar repositories for speaksee
Users that are interested in speaksee are comparing it to the libraries listed below
Sorting:
- [EMNLP 2018] PyTorch code for TVQA: Localized, Compositional Video Question Answering☆178Updated 2 years ago
- Implementation for the AAAI2019 paper "Large-scale Visual Relationship Understanding"☆145Updated 5 years ago
- Pre-trained V+L Data Preparation☆46Updated 5 years ago
- Code for our paper: Learning Conditioned Graph Structures for Interpretable Visual Question Answering☆149Updated 6 years ago
- Baseline model for nocaps benchmark, ICCV 2019 paper "nocaps: novel object captioning at scale".☆76Updated last year
- Dense video captioning in PyTorch☆41Updated 5 years ago
- Code for CVPR'19 "Recursive Visual Attention in Visual Dialog"☆64Updated 2 years ago
- ☆54Updated 5 years ago
- A Dataset for Grounded Video Description☆162Updated 3 years ago
- This is the repo for Multi-level textual grounding☆33Updated 4 years ago
- [ACL 2020] PyTorch code for TVQA+: Spatio-Temporal Grounding for Video Question Answering☆129Updated 2 years ago
- PyTorch code for: Learning to Generate Grounded Visual Captions without Localization Supervision☆44Updated 4 years ago
- Evaluation code for Dense-Captioning Events in Videos☆128Updated 6 years ago
- MUREL (CVPR 2019), a multimodal relational reasoning module for VQA☆195Updated 5 years ago
- Implementation for "Multilevel Language and Vision Integration for Text-to-Clip Retrieval"☆50Updated 6 years ago
- Rethinking Diversified and Discriminative Proposal Generation for Visual Grounding☆23Updated 7 years ago
- Learning to Evaluate Image Captioning. CVPR 2018☆84Updated 7 years ago
- Torch Implementation of Speaker-Listener-Reinforcer for Referring Expression Generation and Comprehension☆34Updated 7 years ago
- Pytorch implementation of "Explainable and Explicit Visual Reasoning over Scene Graphs "☆93Updated 6 years ago
- Implementation for our paper "Conditional Image-Text Embedding Networks"☆39Updated 5 years ago
- Joint Embedding with Multimodal Cues for Cross-Modal Video-Text Retrieval☆67Updated 5 years ago
- Use transformer for captioning☆156Updated 6 years ago
- Code release for Hu et al., Language-Conditioned Graph Networks for Relational Reasoning. in ICCV, 2019☆92Updated 5 years ago
- Data and code for CVPR 2020 paper: "VIOLIN: A Large-Scale Dataset for Video-and-Language Inference"☆162Updated 5 years ago
- Mixture-of-Embeddings-Experts☆120Updated 4 years ago
- [CVPR20] Video Object Grounding using Semantic Roles in Language Description (https://arxiv.org/abs/2003.10606)☆67Updated 5 years ago
- Show, Edit and Tell: A Framework for Editing Image Captions, CVPR 2020☆80Updated 5 years ago
- Code for Discriminability objective for training descriptive captions(CVPR 2018)☆109Updated 5 years ago
- Stack-Captioning: Coarse-to-Fine Learning for Image Captioning☆62Updated 7 years ago
- Data of ACL 2019 Paper "Expressing Visual Relationships via Language".☆62Updated 4 years ago