hwjiang1510 / VQLoCView external linksLinks
(NeurIPS 2023) Open-set visual object query search & localization in long-form videos
☆26Feb 1, 2024Updated 2 years ago
Alternatives and similar repositories for VQLoC
Users that are interested in VQLoC are comparing it to the libraries listed below
Sorting:
- [CVPR 2024] Context-Guided Spatio-Temporal Video Grounding☆65Jun 28, 2024Updated last year
- [NeurIPS'24] MemVLT: Vision-Language Tracking with Adaptive Memory-based Prompts☆18Oct 7, 2024Updated last year
- The official code of Towards Balanced Alignment: Modal-Enhanced Semantic Modeling for Video Moment Retrieval (AAAI2024)☆32Mar 29, 2024Updated last year
- ☆19Oct 9, 2024Updated last year
- DisTime: Distribution-based Time Representation for Video Large Language Models.☆18Jul 10, 2025Updated 7 months ago
- Official implementation of "A Backpack Full of Skills: Egocentric Video Understanding with Diverse Task Perspectives", accepted at CVPR 2…☆24Jun 13, 2024Updated last year
- Official repository of paper "Let All be Whitened: Multi-teacher Distillation for Efficient Visual Retrieval"☆10Dec 20, 2023Updated 2 years ago
- ☆25Dec 23, 2024Updated last year
- [ICLR2023] Video Scene Graph Generation from Single-Frame Weak Supervision☆12Sep 17, 2023Updated 2 years ago
- PiVOT uses a foundational model for online automatic visual prompt refinement to aid tracking.☆15May 15, 2025Updated 8 months ago
- This is a repository contains the implementation of our NeurIPS'24 paper "Temporal Sentence Grounding with Relevance Feedback in Videos"☆13Aug 22, 2025Updated 5 months ago
- Source code of the paper: Overlapped Trajectory-Enhanced Visual Tracking☆11Sep 3, 2024Updated last year
- Code and data release for the paper "Learning Object State Changes in Videos: An Open-World Perspective" (CVPR 2024)☆35Sep 9, 2024Updated last year
- [NeurIPS 2022] Embracing Consistency: A One-Stage Approach for Spatio-Temporal Video Grounding☆53Mar 5, 2024Updated last year
- Structuring Hour-Long Videos into Navigable Chapters and Hierarchical Summaries☆34Nov 19, 2025Updated 2 months ago
- TrackGPT: Track What You Need in Videos via Text Prompts☆25May 16, 2023Updated 2 years ago
- All in One: Exploring Unified Vision-Language Tracking with Multi-Modal Alignment☆19Feb 11, 2025Updated last year
- SpaceVLLM: Endowing Multimodal Large Language Model with Spatio-Temporal Video Grounding Capability☆16May 8, 2025Updated 9 months ago
- [ICCV 2025] Factorized Learning for Temporally Grounded Video-Language Models☆26Jan 1, 2026Updated last month
- Repo for paper "MUSEG: Reinforcing Video Temporal Understanding via Timestamp-Aware Multi-Segment Grounding".☆39Jun 9, 2025Updated 8 months ago
- [CVPR 2025] LLaVA-ST: A Multimodal Large Language Model for Fine-Grained Spatial-Temporal Understanding☆81Jul 4, 2025Updated 7 months ago
- A vision-language tracking paper list, articles related to visual language tracking have been documented.☆42Dec 15, 2024Updated last year
- [2023 ACL] CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding☆31Aug 5, 2023Updated 2 years ago
- ☆68Nov 5, 2025Updated 3 months ago
- ☆13Jul 20, 2024Updated last year
- [ICASSP'25] Enhancing Vision-Language Tracking by Effectively Converting Textual Cues into Visual Cues☆17Dec 31, 2024Updated last year
- Awesome Visual Tracking☆24Oct 3, 2025Updated 4 months ago
- Official implementation of paper VideoLLM Knows When to Speak: Enhancing Time-Sensitive Video Comprehension with Video-Text Duet Interact…☆42Feb 5, 2025Updated last year
- ☆14Jul 15, 2024Updated last year
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆42Mar 11, 2025Updated 11 months ago
- Pytorch Code for "Unified Coarse-to-Fine Alignment for Video-Text Retrieval" (ICCV 2023)☆66Jun 7, 2024Updated last year
- TStar is a unified temporal search framework for long-form video question answering☆86Sep 2, 2025Updated 5 months ago
- [ICCV'2023] Compositional Feature Augmentation for Unbiased Scene Graph Generation☆15Dec 5, 2023Updated 2 years ago
- ☆16Oct 4, 2024Updated last year
- TEMPURA enables video-language models to reason about causal event relationships and generate fine-grained, timestamped descriptions of u…☆25Jun 4, 2025Updated 8 months ago
- Github repository for "Why Is Spatial Reasoning Hard for VLMs? An Attention Mechanism Perspective on Focus Areas" (ICML 2025)☆68May 2, 2025Updated 9 months ago
- Code for "Are “Hierarchical” Visual Representations Hierarchical?" in NeurIPS Workshop for Symmetry and Geometry in Neural Representation…☆22Nov 8, 2023Updated 2 years ago
- This repo contains the code for the recipe of the winning entry to the Ego4d VQ2D challenge at CVPR 2022.☆41Mar 7, 2023Updated 2 years ago
- [ICLR'25] Official code for "Can Video LLMs Refuse to Answer? Alignment for Answerability in Video Large Language Models"☆34Dec 26, 2025Updated last month