liveseongho / Awesome-Video-Language-UnderstandingLinks
A Survey on video and language understanding.
☆50Updated 2 years ago
Alternatives and similar repositories for Awesome-Video-Language-Understanding
Users that are interested in Awesome-Video-Language-Understanding are comparing it to the libraries listed below
Sorting:
- FunQA benchmarks funny, creative, and magic videos for challenging tasks including timestamp localization, video description, reasoning, …☆102Updated 8 months ago
- 🦩 Visual Instruction Tuning with Polite Flamingo - training multi-modal LLMs to be both clever and polite! (AAAI-24 Oral)☆64Updated last year
- [NeurIPS 2023] Self-Chained Image-Language Model for Video Localization and Question Answering☆187Updated last year
- LAVIS - A One-stop Library for Language-Vision Intelligence☆48Updated last year
- ☆138Updated 10 months ago
- ☆108Updated 2 years ago
- ☆91Updated last year
- ☆152Updated 9 months ago
- ☆72Updated last year
- [ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, …☆120Updated 4 months ago
- ☆133Updated last year
- Hierarchical Video-Moment Retrieval and Step-Captioning (CVPR 2023)☆104Updated 6 months ago
- Official code for "What Makes for Good Visual Tokenizers for Large Language Models?".☆58Updated 2 years ago
- [ECCV2024] Official code implementation of Merlin: Empowering Multimodal LLMs with Foresight Minds☆94Updated last year
- A Unified Framework for Video-Language Understanding☆57Updated 2 years ago
- ☆66Updated last year
- [ICLR2024] Codes and Models for COSA: Concatenated Sample Pretrained Vision-Language Foundation Model☆43Updated 7 months ago
- A PyTorch implementation of EmpiricalMVM☆41Updated last year
- [ACL 2023] Official PyTorch code for Singularity model in "Revealing Single Frame Bias for Video-and-Language Learning"☆135Updated 2 years ago
- ☆76Updated last year
- Narrative movie understanding benchmark☆76Updated 2 months ago
- Language Repository for Long Video Understanding☆32Updated last year
- ☆76Updated 8 months ago
- PG-Video-LLaVA: Pixel Grounding in Large Multimodal Video Models☆257Updated last week
- ☆50Updated last year
- Touchstone: Evaluating Vision-Language Models by Language Models☆83Updated last year
- [ICLR2024] The official implementation of paper "UniAdapter: Unified Parameter-Efficient Transfer Learning for Cross-modal Modeling", by …☆75Updated last year
- SVIT: Scaling up Visual Instruction Tuning☆164Updated last year
- Code release for "EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone" [ICCV, 2023]☆99Updated last year
- [NeurIPS2024] Official code for (IMA) Implicit Multimodal Alignment: On the Generalization of Frozen LLMs to Multimodal Inputs☆20Updated 10 months ago