AntResearchNLP / ViLaSRLinks
Reinforcing Spatial Reasoning in Vision-Language Models with Interwoven Thinking and Visual Drawing
☆53Updated last week
Alternatives and similar repositories for ViLaSR
Users that are interested in ViLaSR are comparing it to the libraries listed below
Sorting:
- SpaceR: The first MLLM empowered by SG-RLVR for video spatial reasoning☆71Updated 3 weeks ago
- Official repo for "Streaming Video Understanding and Multi-round Interaction with Memory-enhanced Knowledge" ICLR2025☆64Updated 4 months ago
- A paper list for spatial reasoning☆127Updated last month
- Ego-R1: Chain-of-Tool-Thought for Ultra-Long Egocentric Video Reasoning☆97Updated last month
- Pixel-Level Reasoning Model trained with RL☆180Updated last month
- [ICLR'25] Reconstructive Visual Instruction Tuning☆101Updated 3 months ago
- ☆93Updated 4 months ago
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, …☆162Updated 3 months ago
- [ICML 2025] Streamline Without Sacrifice - Squeeze out Computation Redundancy in LMM☆19Updated 2 months ago
- Uni-CoT: Towards Unified Chain-of-Thought Reasoning Across Text and Vision☆49Updated this week
- ☆41Updated last month
- VCR-Bench: A Comprehensive Evaluation Framework for Video Chain-of-Thought Reasoning☆32Updated 3 weeks ago
- Official implementation of "Ross3D: Reconstructive Visual Instruction Tuning with 3D-Awareness".☆47Updated 2 weeks ago
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆62Updated 3 weeks ago
- Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆52Updated last month
- ☆62Updated last month
- Autoregressive Semantic Visual Reconstruction Helps VLMs Understand Better☆36Updated last month
- SpatialScore: Towards Unified Evaluation for Multimodal Spatial Understanding☆54Updated 3 weeks ago
- ☆82Updated last week
- [ICCV2025]Code Release of Harmonizing Visual Representations for Unified Multimodal Understanding and Generation☆145Updated 2 months ago
- [CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?☆77Updated last week
- Code and dataset link for "DenseWorld-1M: Towards Detailed Dense Grounded Caption in the Real World"☆89Updated last month
- Official repository of DoraemonGPT: Toward Understanding Dynamic Scenes with Large Language Models☆86Updated 11 months ago
- Machine Mental Imagery: Empower Multimodal Reasoning with Latent Visual Tokens (arXiv 2025)☆115Updated this week
- From Flatland to Space: Teaching Vision-Language Models to Perceive and Reason in 3D☆52Updated 2 months ago
- ☆59Updated 4 months ago
- [ECCV 2024] M3DBench introduces a comprehensive 3D instruction-following dataset with support for interleaved multi-modal prompts.☆60Updated 10 months ago
- [LLaVA-Video-R1]✨First Adaptation of R1 to LLaVA-Video (2025-03-18)☆30Updated 2 months ago
- Structured Video Comprehension of Real-World Shorts☆132Updated this week
- A collection of vision foundation models unifying understanding and generation.☆57Updated 7 months ago