llyx97 / video_reason_benchLinks
A benchmark for evaluating vision-centric, complex video reasoning.
☆29Updated last week
Alternatives and similar repositories for video_reason_bench
Users that are interested in video_reason_bench are comparing it to the libraries listed below
Sorting:
- Official implement of MIA-DPO☆63Updated 6 months ago
- Assessing Context-Aware Creative Intelligence in MLLMs☆21Updated 3 weeks ago
- Official implementation of the paper: RICO: Improving Accuracy and Completeness in Image Recaptioning via Visual Reconstruction☆14Updated last month
- Data and Code for CVPR 2025 paper "MMVU: Measuring Expert-Level Multi-Discipline Video Understanding"☆69Updated 5 months ago
- Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆52Updated last month
- Reinforcement Learning Tuning for VideoLLMs: Reward Design and Data Efficiency☆47Updated 2 months ago
- NoisyRollout: Reinforcing Visual Reasoning with Data Augmentation☆84Updated 2 months ago
- Official repository of MMDU dataset☆93Updated 10 months ago
- A Large-scale Dataset for training and evaluating model's ability on Dense Text Image Generation☆73Updated 5 months ago
- ☆52Updated this week
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆63Updated 3 weeks ago
- ☆99Updated 4 months ago
- ACL'24 (Oral) Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback☆72Updated 10 months ago
- The Next Step Forward in Multimodal LLM Alignment☆170Updated 3 months ago
- Official repository of the video reasoning benchmark MMR-V. Can Your MLLMs "Think with Video"?☆35Updated last month
- Repo for paper "T2Vid: Translating Long Text into Multi-Image is the Catalyst for Video-LLMs"☆49Updated 5 months ago
- TStar is a unified temporal search framework for long-form video question answering☆59Updated 4 months ago
- [ECCV 2024] Paying More Attention to Image: A Training-Free Method for Alleviating Hallucination in LVLMs☆132Updated 9 months ago
- This repository is the official implementation of "Look-Back: Implicit Visual Re-focusing in MLLM Reasoning".☆34Updated last month
- [CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?☆78Updated 2 weeks ago
- ☆87Updated last month
- [Neurips 24' D&B] Official Dataloader and Evaluation Scripts for LongVideoBench.☆104Updated last year
- TokLIP: Marry Visual Tokens to CLIP for Multimodal Comprehension and Generation☆103Updated 2 months ago
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆121Updated 2 months ago
- (ICLR 2025 Spotlight) Official code repository for Interleaved Scene Graph.☆27Updated this week
- [ICLR 2025] ChartMimic: Evaluating LMM’s Cross-Modal Reasoning Capability via Chart-to-Code Generation☆117Updated last month
- (ICCV2025) Official repository of paper "ViSpeak: Visual Instruction Feedback in Streaming Videos"☆37Updated last month
- ☆28Updated 9 months ago
- code for "Strengthening Multimodal Large Language Model with Bootstrapped Preference Optimization"☆57Updated 11 months ago
- [NeurIPS 2024] This repo contains evaluation code for the paper "Are We on the Right Way for Evaluating Large Vision-Language Models"☆189Updated 10 months ago