siwooyong / Codalab-Microsoft-COCO-Image-Captioning-ChallengeLinks
π₯ Codalab-Microsoft-COCO-Image-Captioning-Challenge 3rd place solution(06.30.21)
β23Updated 3 years ago
Alternatives and similar repositories for Codalab-Microsoft-COCO-Image-Captioning-Challenge
Users that are interested in Codalab-Microsoft-COCO-Image-Captioning-Challenge are comparing it to the libraries listed below
Sorting:
- A length-controllable and non-autoregressive image captioning model.β68Updated 4 years ago
- Code for "Aligning Visual Regions and Textual Concepts for Semantic-Grounded Image Representations" οΌNeurIPS 2019οΌβ65Updated 5 years ago
- Implementation of paper "Improving Image Captioning with Better Use of Caption"β33Updated 5 years ago
- [CVPR 2020] Transform and Tell: Entity-Aware News Image Captioningβ93Updated last year
- Implementation for MAF: Multimodal Alignment Frameworkβ46Updated 5 years ago
- β24Updated 3 years ago
- β44Updated 7 months ago
- Human-like Controllable Image Captioning with Verb-specific Semantic Roles.β36Updated 3 years ago
- Show, Edit and Tell: A Framework for Editing Image Captions, CVPR 2020β81Updated 5 years ago
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integrationβ56Updated 2 years ago
- A PyTorch implementation of the paper Multimodal Transformer with Multiview Visual Representation for Image Captioningβ25Updated 5 years ago
- β64Updated 4 years ago
- pre-trained vision and language model summaryβ12Updated 4 years ago
- CVPR 2021 Official Pytorch Code for UC2: Universal Cross-lingual Cross-modal Vision-and-Language Pre-trainingβ34Updated 4 years ago
- Research Code for NeurIPS 2020 Spotlight paper "Large-Scale Adversarial Training for Vision-and-Language Representation Learning": UNITERβ¦β119Updated 5 years ago
- An implementation that downstreams pre-trained V+L models to VQA tasks. Now support: VisualBERT, LXMERT, and UNITERβ165Updated 3 years ago
- β67Updated 3 years ago
- Controllable mage captioning model with unsupervised modesβ21Updated 2 years ago
- [ACL 2020] PyTorch code for MART: Memory-Augmented Recurrent Transformer for Coherent Video Paragraph Captioningβ171Updated 5 years ago
- β53Updated 4 years ago
- A reading list of papers about Visual Question Answering.β35Updated 3 years ago
- Multitask Multilingual Multimodal Pre-trainingβ72Updated 3 years ago
- Code for ACL 2020 paper "Dense-Caption Matching and Frame-Selection Gating for Temporal Localization in VideoQA." Hyounghun Kim, Zineng Tβ¦β34Updated 5 years ago
- Code of Dense Relational Captioningβ69Updated 2 years ago
- [ECCV 2020] PyTorch code of MMT (a multimodal transformer captioning model) on TVCaption datasetβ90Updated 2 years ago
- Data Release for VALUE Benchmarkβ30Updated 3 years ago
- Simple is not Easy: A Simple Strong Baseline for TextVQA and TextCaps[AAAI2021]β57Updated 3 years ago
- Code and Resources for the Transformer Encoder Reasoning Network (TERN) - https://arxiv.org/abs/2004.09144β58Updated 2 years ago
- Extended Intramodal and Intermodal Semantic Similarity Judgments for MS-COCOβ54Updated 5 years ago
- Research Code for NeurIPS 2020 Spotlight paper "Large-Scale Adversarial Training for Vision-and-Language Representation Learning": LXMERTβ¦β21Updated 5 years ago