SatyamGaba / visual_question_answeringLinks
Visual Question Answering in PyTorch with various Attention Models
☆20Updated 5 years ago
Alternatives and similar repositories for visual_question_answering
Users that are interested in visual_question_answering are comparing it to the libraries listed below
Sorting:
- Pytorch implementation of VQA: Visual Question Answering (https://arxiv.org/pdf/1505.00468.pdf) using VQA v2.0 dataset for open-ended ta…☆21Updated 5 years ago
- CNN+LSTM, Attention based, and MUTAN-based models for Visual Question Answering☆77Updated 6 years ago
- [ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos☆126Updated 2 years ago
- GRIT: Faster and Better Image-captioning Transformer (ECCV 2022)☆198Updated 2 years ago
- Pytorch implementation of image captioning using transformer-based model.☆68Updated 2 years ago
- Implementation code of the work "Exploiting Multiple Sequence Lengths in Fast End to End Training for Image Captioning"☆94Updated last year
- Image Captioning Using Transformer☆271Updated 3 years ago
- Code of Dense Relational Captioning☆69Updated 2 years ago
- A length-controllable and non-autoregressive image captioning model.☆68Updated 4 years ago
- Using VideoBERT to tackle video prediction☆134Updated 4 years ago
- A simplified pytorch version of densecap☆42Updated last year
- Pytorch VQA : Visual Question Answering (https://arxiv.org/pdf/1505.00468.pdf)☆98Updated 2 years ago
- PyTorch implementation of Multi-modal Dense Video Captioning (CVPR 2020 Workshops)☆144Updated 2 years ago
- Using LSTM or Transformer to solve Image Captioning in Pytorch☆79Updated 4 years ago
- CapDec: SOTA Zero Shot Image Captioning Using CLIP and GPT2, EMNLP 2022 (findings)☆202Updated 2 years ago
- Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for …☆61Updated 3 years ago
- An implementation that downstreams pre-trained V+L models to VQA tasks. Now support: VisualBERT, LXMERT, and UNITER☆165Updated 3 years ago
- Implementation of the Object Relation Transformer for Image Captioning☆180Updated last year
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Updated 2 years ago
- The repository collects many various multi-modal transformer architectures, including image transformer, video transformer, image-languag…☆233Updated 3 years ago
- A reading list of papers about Visual Question Answering.☆35Updated 3 years ago
- Code for CVPR 2023 paper "Procedure-Aware Pretraining for Instructional Video Understanding"☆50Updated last year
- Implementation of 'End-to-End Transformer Based Model for Image Captioning' [AAAI 2022]☆69Updated last year
- [ICLR 2022] code for "How Much Can CLIP Benefit Vision-and-Language Tasks?" https://arxiv.org/abs/2107.06383☆419Updated 3 years ago
- [TMM 2023] VideoXum: Cross-modal Visual and Textural Summarization of Videos☆53Updated last year
- A curated list of research papers in Video Captioning☆121Updated 5 years ago
- ☆76Updated 3 years ago
- A unified framework to jointly model images, text, and human attention traces.☆79Updated 4 years ago
- (ACL'2023) MultiCapCLIP: Auto-Encoding Prompts for Zero-Shot Multilingual Visual Captioning☆36Updated last year
- A self-evident application of the VQA task is to design systems that aid blind people with sight reliant queries. The VizWiz VQA dataset …☆15Updated 2 years ago