passerby233 / Collection-of-Visual-Storytelling-StoryNLPLinks
This repository aims to collect the articles and codes for the Visual Storytelling (VIST) task. VIST is a vision-and-language task. It aims to summarize the idea of a photo stream and tells a story about it (in natural language). Be careful about its difference from the "storytelling with data", which is more related to data visualization.
☆22Updated 4 years ago
Alternatives and similar repositories for Collection-of-Visual-Storytelling-StoryNLP
Users that are interested in Collection-of-Visual-Storytelling-StoryNLP are comparing it to the libraries listed below
Sorting:
- ☆76Updated 3 years ago
- [ICCV 2021 Oral + TPAMI] Just Ask: Learning to Answer Questions from Millions of Narrated Videos☆124Updated 2 years ago
- Align and Prompt: Video-and-Language Pre-training with Entity Prompts☆187Updated 6 months ago
- Some papers about *diverse* image (a few videos) captioning☆26Updated 2 years ago
- Research code for CVPR 2022 paper: "EMScore: Evaluating Video Captioning via Coarse-Grained and Fine-Grained Embedding Matching"☆26Updated 3 years ago
- [CVPR21] Visual Semantic Role Labeling for Video Understanding (https://arxiv.org/abs/2104.00990)☆60Updated 4 years ago
- [ACL 2021] mTVR: Multilingual Video Moment Retrieval☆27Updated 3 years ago
- Pytorch code for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners☆115Updated 3 years ago
- Github repository for Plot and Rework: Modeling Storylines for Visual Storytelling (ACL-IJCNLP2021 Findings)☆21Updated 3 years ago
- [EMNLP 2020] What is More Likely to Happen Next? Video-and-Language Future Event Prediction☆51Updated 3 years ago
- Hierarchical Video-Moment Retrieval and Step-Captioning (CVPR 2023)☆106Updated 9 months ago
- MAD: A Scalable Dataset for Language Grounding in Videos from Movie Audio Descriptions☆168Updated 2 years ago
- A Unified Framework for Video-Language Understanding☆60Updated 2 years ago
- Source code of our MM'22 paper Partially Relevant Video Retrieval☆54Updated 11 months ago
- [ECCV 2020] PyTorch code of MMT (a multimodal transformer captioning model) on TVCaption dataset☆90Updated 2 years ago
- A length-controllable and non-autoregressive image captioning model.☆68Updated 4 years ago
- ☆43Updated 4 years ago
- A curated list of research papers in Video Captioning☆121Updated 4 years ago
- MDMMT: Multidomain Multimodal Transformer for Video Retrieval☆26Updated 4 years ago
- Narrative movie understanding benchmark☆76Updated 4 months ago
- Code for paper, "TL;DW? Summarizing Instructional Videos with Task Relevance & Cross-Modal Saliency" ECCV 2022☆39Updated 2 years ago
- [ACL 2023] Official PyTorch code for Singularity model in "Revealing Single Frame Bias for Video-and-Language Learning"☆136Updated 2 years ago
- Use CLIP to represent video for Retrieval Task☆70Updated 4 years ago
- Starter Code for VALUE benchmark☆80Updated 3 years ago
- A reading list of papers about Visual Question Answering.☆34Updated 3 years ago
- [CVPR 2023] VoP: Text-Video Co-operative Prompt Tuning for Cross-Modal Retrieval☆38Updated 2 years ago
- ☆25Updated 3 years ago
- [TMM 2023] VideoXum: Cross-modal Visual and Textural Summarization of Videos☆50Updated last year
- [SIGIR 2022] CenterCLIP: Token Clustering for Efficient Text-Video Retrieval. Also, a text-video retrieval toolbox based on CLIP + fast p…☆133Updated 3 years ago
- Using LLMs and pre-trained caption models for super-human performance on image captioning.☆42Updated 2 years ago