apugoneappu / ask_me_anything
An easy-to-use app to visualise attentions of various VQA models.
☆41Updated 2 years ago
Alternatives and similar repositories for ask_me_anything
Users that are interested in ask_me_anything are comparing it to the libraries listed below
Sorting:
- CNN+LSTM, Attention based, and MUTAN-based models for Visual Question Answering☆75Updated 5 years ago
- Generic framework for ML projects☆19Updated 2 years ago
- PyTorch code for Learning Cooperative Visual Dialog Agents using Deep Reinforcement Learning☆169Updated 6 years ago
- Labs for the course on Meta Learning at BITS-Goa☆34Updated 4 years ago
- Starter code in PyTorch for the Visual Dialog challenge☆191Updated 2 years ago
- Baseline model for nocaps benchmark, ICCV 2019 paper "nocaps: novel object captioning at scale".☆75Updated last year
- PyTorch VQA implementation that achieved top performances in the (ECCV18) VizWiz Grand Challenge: Answering Visual Questions from Blind P…☆60Updated 6 years ago
- Neural Turing Machines in Pytorch.☆46Updated 6 years ago
- [CVPR 2021] VirTex: Learning Visual Representations from Textual Annotations☆563Updated last year
- A neural network architecture(CNN+LSTM) that automatically generates captions from the images. The model uses ResNet architecture to trai…☆25Updated 5 years ago
- Implementation for "Large-scale Pretraining for Visual Dialog" https://arxiv.org/abs/1912.02379☆96Updated 5 years ago
- A tool to prepare for GRE using command line terminal. Build in process.☆8Updated 3 years ago
- Repository to generate CLEVR-Dialog: A diagnostic dataset for Visual Dialog☆46Updated 5 years ago
- Improved Fusion of Visual and Language Representations by Dense Symmetric Co-Attention for Visual Question Answering☆106Updated 5 years ago
- Attention-based Visual Question Answering in Torch☆100Updated 7 years ago
- ☆18Updated 5 years ago
- A collection of multimodal datasets, and visual features for VQA and captionning in pytorch. Just run "pip install multimodal"☆82Updated 3 years ago
- An implementation that downstreams pre-trained V+L models to VQA tasks. Now support: VisualBERT, LXMERT, and UNITER☆163Updated 2 years ago
- Example of a Cover letter for AI Residency☆80Updated 5 years ago
- Code for our paper: *Shamsian, *Kleinfeld, Globerson & Chechik, "Learning Object Permanence from Video"☆68Updated 5 months ago
- List of AI Residency & Research programs, Ph.D Fellowships, Research Internships☆158Updated 4 years ago
- ML/DL meeting group at IIT Kharagpur☆47Updated 4 years ago
- The Easy Visual Question Answering dataset.☆33Updated last year
- Strong baseline for visual question answering☆239Updated 2 years ago
- Memory, Attention and Composition (MAC) Network for CLEVR implemented in PyTorch☆85Updated 6 years ago
- Grid features pre-training code for visual question answering☆269Updated 3 years ago
- Transformer-based image captioning extension for pytorch/fairseq☆316Updated 4 years ago
- Recognition to Cognition Networks (code for the model in "From Recognition to Cognition: Visual Commonsense Reasoning", CVPR 2019)☆465Updated 4 years ago
- Code implementation for our ICPR, 2020 paper titled "Improving Word Recognition using Multiple Hypotheses and Deep Embeddings"☆21Updated 3 years ago
- Code for ICML 2019 paper "Probabilistic Neural-symbolic Models for Interpretable Visual Question Answering" [long-oral]☆66Updated last year