michelecafagna26 / faster-rcnn-bottom-up-pyLinks
Extract features and bounding boxes using the original Bottom-up Attention Faster-RCNN in a few lines of Python code
☆11Updated 2 years ago
Alternatives and similar repositories for faster-rcnn-bottom-up-py
Users that are interested in faster-rcnn-bottom-up-py are comparing it to the libraries listed below
Sorting:
- Original VinVL visual backbone with simplified APIs to easily extract features, boxes, object detections, in a few lines of Python code.☆9Updated 2 years ago
- Original VinVL (and Oscar) repo with API designed for an easy inference☆8Updated 2 years ago
- [EMNLP’24 Main] Encoding and Controlling Global Semantics for Long-form Video Question Answering☆19Updated 9 months ago
- [CVPR 2023] Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation☆62Updated 4 months ago
- [AAAI 2023 Oral] VLTinT: Visual-Linguistic Transformer-in-Transformer for Coherent Video Paragraph Captioning☆67Updated last year
- A Good Prompt Is Worth Millions of Parameters: Low-resource Prompt-based Learning for Vision-Language Models (ACL 2022)☆42Updated 3 years ago
- Code and data for ImageCoDe, a contextual vison-and-language benchmark☆40Updated last year
- Recent Advances in Visual Dialog☆30Updated 2 years ago
- [ECCV'24] Official Implementation of Autoregressive Visual Entity Recognizer.☆14Updated last year
- Pytorch code for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners☆115Updated 2 years ago
- ☆84Updated 2 years ago
- ICCV 2023 (Oral) Open-domain Visual Entity Recognition Towards Recognizing Millions of Wikipedia Entities☆41Updated last month
- NLX-GPT: A Model for Natural Language Explanations in Vision and Vision-Language Tasks, CVPR 2022 (Oral)☆48Updated last year
- Video Graph Transformer for Video Question Answering (ECCV'22)☆48Updated 2 years ago
- CLEVR-X: A Visual Reasoning Dataset for Natural Language Explanations☆28Updated last year
- ☆104Updated 3 years ago
- ☆39Updated 2 years ago
- [ICCV 2021] Official implementation of the paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering"☆66Updated 3 years ago
- The SVO-Probes Dataset for Verb Understanding☆31Updated 3 years ago
- An easy to use, user-friendly and efficient code for extracting OpenAI CLIP (Global/Grid) features from image and text respectively.☆129Updated 6 months ago
- Research code for "KAT: A Knowledge Augmented Transformer for Vision-and-Language"☆65Updated 3 years ago
- Official PyTorch Implementation for CVPR'23 Paper, "The Dialog Must Go On: Improving Visual Dialog via Generative Self-Training"☆20Updated last year
- (ACL'2023) MultiCapCLIP: Auto-Encoding Prompts for Zero-Shot Multilingual Visual Captioning☆35Updated 11 months ago
- MuKEA: Multimodal Knowledge Extraction and Accumulation for Knowledge-based Visual Question Answering☆96Updated 2 years ago
- Multimodal Graph Network (MGN): Code repo, examples from the paper☆25Updated 4 years ago
- Code and data for "Broaden the Vision: Geo-Diverse Visual Commonsense Reasoning" (EMNLP 2021).☆28Updated 3 years ago
- Official implementation of "ConZIC: Controllable Zero-shot Image Captioning by Sampling-Based Polishing"☆73Updated last year
- CLIP4IDC: CLIP for Image Difference Captioning (AACL 2022)☆34Updated 2 years ago
- [CVPR21] Visual Semantic Role Labeling for Video Understanding (https://arxiv.org/abs/2104.00990)☆60Updated 3 years ago
- Code for the ICCV'21 paper "Context-aware Scene Graph Generation with Seq2Seq Transformers"☆43Updated 3 years ago