Official code for MotionBench (CVPR 2025)
☆64Mar 3, 2025Updated 11 months ago
Alternatives and similar repositories for MotionBench
Users that are interested in MotionBench are comparing it to the libraries listed below
Sorting:
- [ICLR 2026] MotionSight's official code implementation.☆46Feb 13, 2026Updated 2 weeks ago
- Soft-QMIX: Integrating Maximum Entropy For Monotonic Value Function Factorization☆15Jul 3, 2024Updated last year
- [EMNLP 2024] Preserving Multi-Modal Capabilities of Pre-trained VLMs for Improving Vision-Linguistic Compositionality☆21Oct 8, 2024Updated last year
- (ICCV2025) Official repository of paper "ViSpeak: Visual Instruction Feedback in Streaming Videos"☆45Jul 1, 2025Updated 8 months ago
- [CVPR 2025] PVC: Progressive Visual Token Compression for Unified Image and Video Processing in Large Vision-Language Models☆51Jun 12, 2025Updated 8 months ago
- ☆14Sep 11, 2025Updated 5 months ago
- ☆11Aug 7, 2025Updated 6 months ago
- Official InfiniBench: A Benchmark for Large Multi-Modal Models in Long-Form Movies and TV Shows☆19Nov 4, 2025Updated 3 months ago
- ☆20Oct 15, 2025Updated 4 months ago
- This is the official code repository for the paper: Towards General Continuous Memory for Vision-Language Models.☆20Jul 3, 2025Updated 7 months ago
- ☆14Jun 2, 2025Updated 9 months ago
- Scripting Multi-Scene Videos with Time-Aware and Structural Audio-Visual Captions☆21Feb 11, 2026Updated 2 weeks ago
- ☆11Aug 4, 2024Updated last year
- This is an implementation of the paper "Are We Done with Object-Centric Learning?"☆12Sep 11, 2025Updated 5 months ago
- From Accuracy to Robustness: A Study of Rule- and Model-based Verifiers in Mathematical Reasoning.☆25Oct 7, 2025Updated 4 months ago
- TemporalBench: Benchmarking Fine-grained Temporal Understanding for Multimodal Video Models☆37Nov 10, 2024Updated last year
- Official implementation of Deep Factorized Metric Learning.☆20Jun 6, 2023Updated 2 years ago
- ☆16Mar 26, 2025Updated 11 months ago
- 🕵 Code for our EMNLP 2025 Main paper: "FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in Diverse Adventure Games"☆24Dec 14, 2025Updated 2 months ago
- Extending context length of visual language models☆12Dec 18, 2024Updated last year
- ☆37Nov 8, 2024Updated last year
- From Flatland to Space (SPAR). Accepted to NeurIPS 2025 Datasets & Benchmarks. A large-scale dataset & benchmark for 3D spatial perceptio…☆78Jan 5, 2026Updated last month
- Code for paper: Unified Text-to-Image Generation and Retrieval☆16Jul 6, 2024Updated last year
- [CVPR 2025] OmniMMI: A Comprehensive Multi-modal Interaction Benchmark in Streaming Video Contexts☆21Dec 22, 2025Updated 2 months ago
- Doe-1: Closed-Loop Autonomous Driving with Large World Model☆114Jan 21, 2025Updated last year
- KMM: Key Frame Mask Mamba for Extended Motion Generation☆19Sep 22, 2025Updated 5 months ago
- Transactions on Multimedia (TMM25)☆19Apr 8, 2025Updated 10 months ago
- [CVPR 2025 Oral] VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection☆137Jul 28, 2025Updated 7 months ago
- ☆48Nov 1, 2024Updated last year
- [NeurIPS 2024] Official PyTorch implementation of "Improving Compositional Reasoning of CLIP via Synthetic Vision-Language Negatives"☆46Dec 1, 2024Updated last year
- VideoNSA: Native Sparse Attention Scales Video Understanding☆81Nov 16, 2025Updated 3 months ago
- Inferix: A Block-Diffusion based Next-Generation Inference Engine for World Simulation☆110Updated this week
- TStar is a unified temporal search framework for long-form video question answering☆87Sep 2, 2025Updated 6 months ago
- [ICLR 2025] Video Action Differencing☆52Jul 3, 2025Updated 7 months ago
- Code, Data and Model for Paper "Learning from Peers in Reasoning Models"☆27May 13, 2025Updated 9 months ago
- This project is the official implementation of 'DreamOmni3: Scribble-based Editing and Generation''☆38Dec 30, 2025Updated 2 months ago
- iLLaVA: An Image is Worth Fewer Than 1/3 Input Tokens in Large Multimodal Models☆21Jan 29, 2025Updated last year
- DiP: Taming Diffusion Models in Pixel Space☆55Nov 27, 2025Updated 3 months ago
- Code for "CLIP Behaves like a Bag-of-Words Model Cross-modally but not Uni-modally"☆20Feb 14, 2025Updated last year