BierOne / bottom-up-attention-vqa
An updated PyTorch implementation of hengyuan-hu's version for 'Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering'
☆36Updated 3 years ago
Alternatives and similar repositories for bottom-up-attention-vqa:
Users that are interested in bottom-up-attention-vqa are comparing it to the libraries listed below
- A pytorch implementation of "Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering" for image captioning.☆47Updated 3 years ago
- Official Code for 'RSTNet: Captioning with Adaptive Attention on Visual and Non-Visual Words' (CVPR 2021)☆122Updated 2 years ago
- Implementation of 'End-to-End Transformer Based Model for Image Captioning' [AAAI 2022]☆67Updated 10 months ago
- Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for …☆60Updated 2 years ago
- ☆67Updated 2 years ago
- Official pytorch implementation of paper "Dual-Level Collaborative Transformer for Image Captioning" (AAAI 2021).☆198Updated 2 years ago
- Code for our IJCAI2020 paper: Overcoming Language Priors with Self-supervised Learning for Visual Question Answering☆50Updated 4 years ago
- Implementation for CVPR 2022 paper " Injecting Semantic Concepts into End-to-End Image Captionin".☆42Updated 2 years ago
- [ECCV 2020] Official code for "Comprehensive Image Captioning via Scene Graph Decomposition"☆97Updated 7 months ago
- Deep Multimodal Neural Architecture Search☆28Updated 4 years ago
- Dynamic Modality Interaction Modeling for Image-Text Retrieval. SIGIR'21☆68Updated 2 years ago
- Code for "Learning the Best Pooling Strategy for Visual Semantic Embedding", CVPR 2021 (Oral)☆161Updated 2 years ago
- Compact Trilinear Interaction for Visual Question Answering (ICCV 2019)☆38Updated 2 years ago
- Implementation of our ACMMM2019 paper, Focus Your Attention: A Bidirectional Focal Attention Network for Image-Text Matching☆38Updated last year
- A PyTorch reimplementation of bottom-up-attention models☆298Updated 3 years ago
- A PyTorch implementation of the paper Multimodal Transformer with Multiview Visual Representation for Image Captioning☆25Updated 4 years ago
- ☆220Updated 3 years ago
- Implementation of our AAAI2022 paper, Show Your Faith: Cross-Modal Confidence-Aware Network for Image-Text Matching.☆36Updated last year
- The PyTorch code of the AAAI2021 paper "Non-Autoregressive Coarse-to-Fine Video Captioning".☆58Updated last year
- Microsoft COCO Caption Evaluation Tool - Python 3☆33Updated 5 years ago
- Local self-attention in Transformer for visual question answering☆12Updated last year
- [CVPR 2021] Counterfactual VQA: A Cause-Effect Look at Language Bias☆121Updated 3 years ago
- Implementation of our CVPR2020 paper, Graph Structured Network for Image-Text Matching☆167Updated 4 years ago
- Implementation of the Object Relation Transformer for Image Captioning☆177Updated 6 months ago
- A Fast and Accurate One-Stage Approach to Visual Grounding, ICCV 2019 (Oral)☆146Updated 4 years ago
- Counterfactual Samples Synthesizing for Robust VQA☆78Updated 2 years ago
- Grid features pre-training code for visual question answering☆269Updated 3 years ago
- Bottom-up features extractor implemented in PyTorch.☆72Updated 5 years ago
- A curated list of Multimodal Captioning related research(including image captioning, video captioning, and text captioning)☆110Updated 2 years ago
- Optimized code based on M2 for faster image captioning training☆20Updated 2 years ago