TencentARC / UMT
UMT is a unified and flexible framework which can handle different input modality combinations, and output video moment retrieval and/or highlight detection results.
☆213Updated last year
Alternatives and similar repositories for UMT:
Users that are interested in UMT are comparing it to the libraries listed below
- Official pytorch repository for "QD-DETR : Query-Dependent Video Representation for Moment Retrieval and Highlight Detection" (CVPR 2023 …☆226Updated last year
- [NeurIPS 2021] Moment-DETR code and QVHighlights dataset☆301Updated last year
- Official pytorch repository for CG-DETR "Correlation-guided Query-Dependency Calibration in Video Representation Learning for Temporal Gr…☆129Updated 7 months ago
- Video Feature Extraction Code for EMNLP 2020 paper "HERO: Hierarchical Encoder for Video+Language Omni-representation Pre-training"☆107Updated 3 years ago
- ☆244Updated 2 years ago
- Official code for "Bridging Video-text Retrieval with Multiple Choice Questions", CVPR 2022 (Oral).☆138Updated 2 years ago
- Hierarchical Video-Moment Retrieval and Step-Captioning (CVPR 2023)☆100Updated 2 months ago
- ☆193Updated 2 years ago
- [ICCV2023 Oral] Unmasked Teacher: Towards Training-Efficient Video Foundation Models☆327Updated 10 months ago
- [arXiv22] Disentangled Representation Learning for Text-Video Retrieval☆95Updated 3 years ago
- Implementation of Cross-category Video Highlight Detection via Set-based Learning (ICCV 2021).☆74Updated 3 years ago
- An official implementation for "X-CLIP: End-to-End Multi-grained Contrastive Learning for Video-Text Retrieval"☆156Updated last year
- Frozen in Time: A Joint Video and Image Encoder for End-to-End Retrieval [ICCV'21]☆360Updated 2 years ago
- Research code for CVPR 2022 paper "SwinBERT: End-to-End Transformers with Sparse Attention for Video Captioning"☆238Updated 2 years ago
- https://layer6ai-labs.github.io/xpool/☆122Updated last year
- 前沿论文持续更新--视频时刻定位 or 时域语言定位 or 视频片段检索。☆249Updated last year
- ☆128Updated last year
- 🌀 R2-Tuning: Efficient Image-to-Video Transfer Learning for Video Temporal Grounding (ECCV 2024)☆82Updated 9 months ago
- MAD: A Scalable Dataset for Language Grounding in Videos from Movie Audio Descriptions☆161Updated last year
- [CVPR2023] All in One: Exploring Unified Video-Language Pre-training☆281Updated 2 years ago
- [TPAMI2024] Codes and Models for VALOR: Vision-Audio-Language Omni-Perception Pretraining Model and Dataset☆286Updated 3 months ago
- [ECCV2022] A pytorch implementation for TS2-Net: Token Shift and Selection Transformer for Text-Video Retrieval☆74Updated 2 years ago
- PyTorch implementation of BEVT (CVPR 2022) https://arxiv.org/abs/2112.01529☆159Updated 2 years ago
- "Video Moment Retrieval from Text Queries via Single Frame Annotation" in SIGIR 2022☆69Updated 2 years ago
- ☆37Updated 5 months ago
- [CVPR2024] The official implementation of AdaTAD: End-to-End Temporal Action Detection with 1B Parameters Across 1000 Frames☆36Updated 9 months ago
- [NIPS2023] Code and Model for VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and Dataset☆275Updated last year
- [ICCV 2023] UniVTG: Towards Unified Video-Language Temporal Grounding☆347Updated 11 months ago
- Temporal Moment(Action) Localization via Language / Temporal Language Grounding / Video Moment Retrieval☆97Updated 3 years ago
- Official pytorch repository for "Knowing Where to Focus: Event-aware Transformer for Video Grounding" (ICCV 2023)☆50Updated last year