uakarsh / latr
Implementation of LaTr: Layout-aware transformer for scene-text VQA,a novel multimodal architecture for Scene Text Visual Question Answering (STVQA)
☆53Updated 5 months ago
Alternatives and similar repositories for latr:
Users that are interested in latr are comparing it to the libraries listed below
- Simple is not Easy: A Simple Strong Baseline for TextVQA and TextCaps[AAAI2021]☆57Updated 3 years ago
- TAP: Text-Aware Pre-training for Text-VQA and Text-Caption, CVPR 2021 (Oral)☆72Updated last year
- The imdb files with SBD-Trans OCR for TextVQA dataset.☆11Updated 3 years ago
- Official code for paper "Spatially Aware Multimodal Transformers for TextVQA" published at ECCV, 2020.☆64Updated 3 years ago
- STVQA and TextVQA OCR results from Amazon Text in Image pipeline☆11Updated 2 years ago
- RUArt: A Novel Text-Centered Solution for Text-Based Visual Question Answering☆10Updated 2 years ago
- Towards Video Text Visual Question Answering: Benchmark and Baseline☆38Updated last year
- ☆188Updated 11 months ago
- UniTAB: Unifying Text and Box Outputs for Grounded VL Modeling, ECCV 2022 (Oral Presentation)☆86Updated last year
- ☆21Updated 2 years ago
- ☆16Updated 3 years ago
- Scene Text Aware Cross Modal Retrieval (StacMR)☆26Updated 3 years ago
- ☆84Updated 2 years ago
- ☆83Updated 3 years ago
- A modular framework for Visual Question Answering research by the FAIR A-STAR team☆45Updated 3 years ago
- natual language guided image captioning☆81Updated last year
- ROSITA: Enhancing Vision-and-Language Semantic Alignments via Cross- and Intra-modal Knowledge Integration☆56Updated last year
- Source code for EMNLP 2022 paper “PEVL: Position-enhanced Pre-training and Prompt Tuning for Vision-language Models”☆48Updated 2 years ago
- Official PyTorch implementation of our CVPR 2022 paper: Beyond a Pre-Trained Object Detector: Cross-Modal Textual and Visual Context for …☆60Updated 2 years ago
- Visually-Situated Natural Language Understanding with Contrastive Reading Model and Frozen Large Language Models, EMNLP 2023☆45Updated 10 months ago
- ☆37Updated 2 years ago
- [AAAI 2021] Confidence-aware Non-repetitive Multimodal Transformers for TextCaps☆24Updated 2 years ago
- [CVPR 2022] The code for our paper 《Object-aware Video-language Pre-training for Retrieval》☆62Updated 2 years ago
- Cross-View Language Modeling: Towards Unified Cross-Lingual Cross-Modal Pre-training (ACL 2023))☆90Updated last year
- Used in M4C feature extraction script: https://github.com/facebookresearch/mmf/blob/project/m4c/projects/M4C/scripts/extract_ocr_frcn_fea…☆13Updated 5 years ago
- [ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos☆120Updated last year
- [ICCV 2021] Official implementation of the paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering"☆66Updated 3 years ago
- (ACL'2023) MultiCapCLIP: Auto-Encoding Prompts for Zero-Shot Multilingual Visual Captioning☆35Updated 8 months ago
- An unofficial pytorch implementation of "TransVG: End-to-End Visual Grounding with Transformers".☆52Updated 3 years ago
- ☆32Updated 3 years ago