Sejong-VLI / V2T-Action-Graph-JKSUCIS-2023
The implementation of a paper entitled "Action Knowledge for Video Captioning with Graph Neural Networks" (JKSUCIS 2023).
☆13Updated last year
Alternatives and similar repositories for V2T-Action-Graph-JKSUCIS-2023:
Users that are interested in V2T-Action-Graph-JKSUCIS-2023 are comparing it to the libraries listed below
- [CVPR2022] Official code for Hierarchical Modular Network for Video Captioning. Our proposed HMN is implemented with PyTorch.☆52Updated 2 years ago
- A Video-to-Text Framework☆10Updated last year
- The first unofficial implementation of CLIP4Caption: CLIP for Video Caption (ACMMM 2021)☆14Updated 2 years ago
- Official pytorch implementation of "Explore-And-Match: Bridging Proposal-Based and Proposal-Free With Transformer for Sentence Grounding …☆42Updated 2 years ago
- Source code of our TCSVT'22 paper Reading-strategy Inspired Visual Representation Learning for Text-to-Video Retrieval☆19Updated 2 years ago
- source code of our MGPN in SIGIR 2022☆18Updated 2 years ago
- Temporal Sentence Grounding in Videos / Natural Language Video Localization / Video Moment Retrieval的相关工作☆29Updated 2 years ago
- Dual Learning with Dynamic Knowledge Distillation for Partially Relevant Video Retrieval☆14Updated 3 weeks ago
- The code of IJCAI22 paper "GL-RG: Global-Local Representation Granularity for Video Captioning".☆18Updated last year
- [ICCV 2023] The official PyTorch implementation of the paper: "Localizing Moments in Long Video Via Multimodal Guidance"☆17Updated 4 months ago
- ☆34Updated last year
- Official pytorch repository for "Knowing Where to Focus: Event-aware Transformer for Video Grounding" (ICCV 2023)☆49Updated last year
- Repository of proposal-free temporal moment localization work☆33Updated 7 months ago
- [ICCV 2023 CLVL Workshop] Zero-Shot and Few-Shot Video Question Answering with Multi-Modal Prompts☆11Updated 2 weeks ago
- Video Graph Transformer for Video Question Answering (ECCV'22)☆46Updated last year
- CPL: Weakly Supervised Temporal Sentence Grounding with Gaussian-based Contrastive Proposal Learning☆59Updated 9 months ago
- Scanning Only Once: An End-to-end Framework for Fast Temporal Grounding in Long Videos☆22Updated 7 months ago
- [arXiv22] Disentangled Representation Learning for Text-Video Retrieval☆94Updated 2 years ago
- Source code of our CVPR2024 paper TeachCLIP for Text-to-Video Retrieval☆26Updated 2 weeks ago
- ☆31Updated 2 years ago
- Video Corpus Moment Retrieval with Contrastive Learning (SIGIR 2021)☆54Updated 3 years ago
- Entity-Aware and Motion-Aware Transformers for Language-driven Action Localization(IJCAI-22)☆12Updated 2 years ago
- ☆13Updated 3 years ago
- NExT-QA: Next Phase of Question-Answering to Explaining Temporal Actions (CVPR'21)☆27Updated last year
- [ICCV 2023] Simple Baselines for Interactive Video Retrieval with Questions and Answers☆13Updated 9 months ago
- ☆14Updated last year
- (TIP'2023) Concept-Aware Video Captioning: Describing Videos with Effective Prior Information☆25Updated last month
- ☆7Updated last year
- Official pytorch implementation of the AAAI 2021 paper "Semantic Grouping Network for Video Captioning"☆51Updated 3 years ago
- Official implementation for Hierarchical Deep Residual Reasoning for Temporal Moment Localization☆9Updated 3 years ago