RachanaJayaram / Cross-Attention-VizWiz-VQALinks
A self-evident application of the VQA task is to design systems that aid blind people with sight reliant queries. The VizWiz VQA dataset originates from images and questions compiled by members of the visually impaired community and as such, highlights some of the challenges presented by this particular use case.
☆15Updated 2 years ago
Alternatives and similar repositories for Cross-Attention-VizWiz-VQA
Users that are interested in Cross-Attention-VizWiz-VQA are comparing it to the libraries listed below
Sorting:
- Code of Dense Relational Captioning☆69Updated 2 years ago
- Code for "Aligning Visual Regions and Textual Concepts for Semantic-Grounded Image Representations" (NeurIPS 2019)☆65Updated 5 years ago
- The source code of ACL 2020 paper: "Cross-Modality Relevance for Reasoning on Language and Vision"☆27Updated 4 years ago
- [CVPR 2020] Transform and Tell: Entity-Aware News Image Captioning☆93Updated last year
- Implementation for MAF: Multimodal Alignment Framework☆46Updated 5 years ago
- Official code and dataset link for ''VMSMO: Learning to Generate Multimodal Summary for Video-based News Articles''☆36Updated 4 years ago
- Code for paper "Adaptively Aligned Image Captioning via Adaptive Attention Time". NeurIPS 2019☆51Updated 5 years ago
- Code for ACL 2020 paper "Dense-Caption Matching and Frame-Selection Gating for Temporal Localization in VideoQA." Hyounghun Kim, Zineng T…☆34Updated 5 years ago
- ☆68Updated 3 years ago
- [EMNLP 2018] Training for Diversity in Image Paragraph Captioning☆91Updated 6 years ago
- Show, Edit and Tell: A Framework for Editing Image Captions, CVPR 2020☆81Updated 5 years ago
- Implementation of paper "Improving Image Captioning with Better Use of Caption"☆33Updated 5 years ago
- ☆64Updated 3 years ago
- ROCK model for Knowledge-Based VQA in Videos☆31Updated 5 years ago
- Human-like Controllable Image Captioning with Verb-specific Semantic Roles.☆36Updated 3 years ago
- ☆38Updated 2 years ago
- A PyTorch implementation of the paper Multimodal Transformer with Multiview Visual Representation for Image Captioning☆25Updated 5 years ago
- ☆44Updated 5 months ago
- Implementation for the paper "Hierarchical Conditional Relation Networks for Video Question Answering" (Le et al., CVPR 2020, Oral)☆133Updated last year
- An implementation that downstreams pre-trained V+L models to VQA tasks. Now support: VisualBERT, LXMERT, and UNITER☆165Updated 3 years ago
- [ECCV 2020] Official code for "Comprehensive Image Captioning via Scene Graph Decomposition"☆99Updated last year
- Video captioning baseline models on Video2Commonsense Dataset.☆57Updated 4 years ago
- Code for our ACL2021 paper: "Check It Again: Progressive Visual Question Answering via Visual Entailment"☆31Updated 4 years ago
- A length-controllable and non-autoregressive image captioning model.☆68Updated 4 years ago
- An image-oriented evaluation tool for image captioning systems (EMNLP-IJCNLP 2019)☆37Updated 5 years ago
- Position Focused Attention Network for Image-Text Matching☆69Updated 6 years ago
- Microsoft COCO Caption Evaluation Tool - Python 3☆33Updated 6 years ago
- [ECCV 2020] PyTorch code of MMT (a multimodal transformer captioning model) on TVCaption dataset☆90Updated 2 years ago
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Updated 2 years ago
- CNN+LSTM, Attention based, and MUTAN-based models for Visual Question Answering☆76Updated 5 years ago