JUNJIE99 / MLVUView external linksLinks
🔥🔥MLVU: Multi-task Long Video Understanding Benchmark
☆241Aug 21, 2025Updated 5 months ago
Alternatives and similar repositories for MLVU
Users that are interested in MLVU are comparing it to the libraries listed below
Sorting:
- VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs☆54Mar 9, 2025Updated 11 months ago
- Long Context Transfer from Language to Vision☆398Mar 18, 2025Updated 10 months ago
- [Neurips 24' D&B] Official Dataloader and Evaluation Scripts for LongVideoBench.☆113Jul 27, 2024Updated last year
- 🔥🔥First-ever hour scale video understanding models☆611Jul 14, 2025Updated 7 months ago
- ✨✨[CVPR 2025] Video-MME: The First-Ever Comprehensive Evaluation Benchmark of Multi-modal LLMs in Video Analysis☆730Dec 8, 2025Updated 2 months ago
- [ICCV 2025] LVBench: An Extreme Long Video Understanding Benchmark☆136Jul 9, 2025Updated 7 months ago
- ☆32Jul 29, 2024Updated last year
- Official implementation of paper AdaReTaKe: Adaptive Redundancy Reduction to Perceive Longer for Video-language Understanding☆88Apr 23, 2025Updated 9 months ago
- ☆109Dec 30, 2024Updated last year
- [ICLR2026] VideoChat-Flash: Hierarchical Compression for Long-Context Video Modeling☆503Nov 18, 2025Updated 2 months ago
- This is the official implementation of ICCV 2025 "Flash-VStream: Efficient Real-Time Understanding for Long Video Streams"☆269Oct 15, 2025Updated 3 months ago
- [ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, …☆128Apr 4, 2025Updated 10 months ago
- Official implementation of paper ReTaKe: Reducing Temporal and Knowledge Redundancy for Long Video Understanding☆39Mar 16, 2025Updated 10 months ago
- ☆37Sep 16, 2024Updated last year
- VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs☆1,277Jan 23, 2025Updated last year
- [CVPR 2024] MovieChat: From Dense Token to Sparse Memory for Long Video Understanding☆682Jan 29, 2025Updated last year
- ☆80Nov 24, 2024Updated last year
- Official InfiniBench: A Benchmark for Large Multi-Modal Models in Long-Form Movies and TV Shows☆19Nov 4, 2025Updated 3 months ago
- ☆155Oct 31, 2024Updated last year
- ✨First Open-Source R1-like Video-LLM [2025/02/18]☆381Feb 23, 2025Updated 11 months ago
- Official code of *Towards Event-oriented Long Video Understanding*☆12Jul 26, 2024Updated last year
- Awesome papers & datasets specifically focused on long-term videos.☆352Oct 9, 2025Updated 4 months ago
- VILA is a family of state-of-the-art vision language models (VLMs) for diverse multimodal AI tasks across the edge, data center, and clou…☆3,737Nov 28, 2025Updated 2 months ago
- ☆160Jan 16, 2025Updated last year
- Official repository for the paper PLLaVA☆676Jul 28, 2024Updated last year
- Official Repository of paper VideoGPT+: Integrating Image and Video Encoders for Enhanced Video Understanding☆292Aug 5, 2025Updated 6 months ago
- A Versatile Video-LLM for Long and Short Video Understanding with Superior Temporal Localization Ability☆105Nov 28, 2024Updated last year
- [CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.☆3,336Jan 18, 2025Updated last year
- One-for-All Multimodal Evaluation Toolkit Across Text, Image, Video, and Audio Tasks☆3,635Updated this week
- ☆37Nov 8, 2024Updated last year
- ☆53Mar 19, 2025Updated 10 months ago
- A lightweight flexible Video-MLLM developed by TencentQQ Multimedia Research Team.☆74Oct 14, 2024Updated last year
- ☆138Sep 29, 2024Updated last year
- Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model☆281Jun 25, 2024Updated last year
- ☆4,552Sep 14, 2025Updated 5 months ago
- [EMNLP'23] The official GitHub page for ''Evaluating Object Hallucination in Large Vision-Language Models''☆105Aug 21, 2025Updated 5 months ago
- Repo for paper "T2Vid: Translating Long Text into Multi-Image is the Catalyst for Video-LLMs"☆48Sep 3, 2025Updated 5 months ago
- ☆242Jun 4, 2025Updated 8 months ago
- ☆138Nov 17, 2025Updated 2 months ago