sjh0354 / World-Aware-PlanningLinks
β19Updated 4 months ago
Alternatives and similar repositories for World-Aware-Planning
Users that are interested in World-Aware-Planning are comparing it to the libraries listed below
Sorting:
- VLA-RFT: Vision-Language-Action Models with Reinforcement Fine-Tuningβ96Updated 2 months ago
- π¦Ύ A Dual-System VLA with System2 Thinkingβ122Updated 3 months ago
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"β205Updated last month
- InstructVLA: Vision-Language-Action Instruction Tuning from Understanding to Manipulationβ70Updated 2 months ago
- Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoningβ78Updated 6 months ago
- Official code for "Embodied-R1: Reinforced Embodied Reasoning for General Robotic Manipulation"β108Updated 3 months ago
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasksβ183Updated 2 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policyβ296Updated 3 weeks ago
- β56Updated this week
- Official implementation of "OneTwoVLA: A Unified Vision-Language-Action Model with Adaptive Reasoning"β200Updated 6 months ago
- Nav-R1: Reasoning and Navigation in Embodied Scenesβ75Updated last month
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintainedπ₯]β173Updated last month
- β86Updated 6 months ago
- Evaluate Multimodal LLMs as Embodied Agentsβ54Updated 9 months ago
- [CVPR2024] This is the official implement of MP5β106Updated last year
- The repo of paper `RoboMamba: Multimodal State Space Model for Efficient Robot Reasoning and Manipulation`β143Updated 11 months ago
- The Official Implementation of RoboMatrixβ104Updated 6 months ago
- β59Updated 8 months ago
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimizationβ152Updated 8 months ago
- [NeurIPS'25] SSR: Enhancing Depth Perception in Vision-Language Models via Rationale-Guided Spatial Reasoningβ33Updated last month
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"β119Updated 9 months ago
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledgeβ245Updated 2 months ago
- β61Updated 9 months ago
- Visual Embodied Brain: Let Multimodal Large Language Models See, Think, and Control in Spacesβ87Updated 6 months ago
- Official implementation of: Bootstrapping Language-Guided Navigation Learning with Self-Refining Data Flywheelβ31Updated 6 months ago
- Embodied Question Answering (EQA) benchmark and method (ICCV 2025)β43Updated 3 months ago
- RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulationβ268Updated this week
- β54Updated last year
- ICCV2025β143Updated 3 weeks ago
- [NeurIPS 2025] VIKIβR: Coordinating Embodied Multi-Agent Cooperation via Reinforcement Learningβ62Updated last month