weijiawu / Awesome-Visual-Reinforcement-LearningLinks
π This is a repository for organizing papers, codes and other resources related to Visual Reinforcement Learning.
β322Updated last week
Alternatives and similar repositories for Awesome-Visual-Reinforcement-Learning
Users that are interested in Awesome-Visual-Reinforcement-Learning are comparing it to the libraries listed below
Sorting:
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, β¦β192Updated 6 months ago
- Pixel-Level Reasoning Model trained with RL [NeuIPS25]β244Updated last month
- Official repo and evaluation implementation of VSI-Benchβ618Updated 3 months ago
- Visual Planning: Let's Think Only with Imagesβ279Updated 5 months ago
- Machine Mental Imagery: Empower Multimodal Reasoning with Latent Visual Tokens (arXiv 2025)β186Updated 3 months ago
- SpaceR: The first MLLM empowered by SG-RLVR for video spatial reasoningβ93Updated 4 months ago
- A paper list for spatial reasoningβ157Updated this week
- Official Code for "Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual Search"β357Updated last month
- β102Updated 3 months ago
- [NeurIPS 2025]βοΈ Reason-RFT: Reinforcement Fine-Tuning for Visual Reasoning.β230Updated last month
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).β121Updated last year
- [ICLR 2025] VILA-U: a Unified Foundation Model Integrating Visual Understanding and Generationβ402Updated 6 months ago
- [NeurIPS 2025] Official Repo of Omni-R1: Reinforcement Learning for Omnimodal Reasoning via Two-System Collaborationβ87Updated 5 months ago
- ACTIVE-O3: Empowering Multimodal Large Language Models with Active Perception via GRPOβ74Updated 5 months ago
- This repository collects papers on VLLM applications. We will update new papers irregularly.β173Updated 2 months ago
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasksβ178Updated last month
- [CVPR2025 Highlight] Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Modelsβ229Updated 4 months ago
- [CVPR 2025] EgoLife: Towards Egocentric Life Assistantβ341Updated 7 months ago
- Official repository of "GoT: Unleashing Reasoning Capability of Multimodal Large Language Model for Visual Generation and Editing"β293Updated last month
- TStar is a unified temporal search framework for long-form video question answeringβ71Updated 2 months ago
- Official implementation of Spatial-MLLM: Boosting MLLM Capabilities in Visual-based Spatial Intelligenceβ375Updated 4 months ago
- [NeurIPS'24] This repository is the implementation of "SpatialRGPT: Grounded Spatial Reasoning in Vision Language Models"β273Updated 10 months ago
- [NIPS2025] VideoChat-R1 & R1.5: Enhancing Spatio-Temporal Perception and Reasoning via Reinforcement Fine-Tuningβ220Updated 3 weeks ago
- [NeurIPS 2025] Official code implementation of Perception R1: Pioneering Perception Policy with Reinforcement Learningβ266Updated 3 months ago
- Collections of Papers and Projects for Multimodal Reasoning.β104Updated 6 months ago
- π₯π₯π₯ Latest Papers, Codes and Datasets on Video-LMM Post-Trainingβ154Updated last week
- https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoTβ96Updated last week
- Vision Manus: Your versatile Visual AI assistantβ290Updated 3 weeks ago
- Official Repo of From Masks to Worlds: A Hitchhikerβs Guide to World Models.β45Updated 2 weeks ago
- π This is a repository for organizing papers, codes, and other resources related to unified multimodal models.β324Updated 3 weeks ago