aopolin-lv / RoboMP2Links
[ICML 2024] RoboMP2: A Robotic Multimodal Perception-Planning Framework with Multimodal Large Language Models
☆11Updated 4 months ago
Alternatives and similar repositories for RoboMP2
Users that are interested in RoboMP2 are comparing it to the libraries listed below
Sorting:
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization☆145Updated 7 months ago
- [IROS24 Oral]ManipVQA: Injecting Robotic Affordance and Physically Grounded Information into Multi-Modal Large Language Models☆97Updated last year
- Official implementation of GR-MG☆90Updated 9 months ago
- ☆60Updated 10 months ago
- ☆47Updated last year
- An example RLDS dataset builder for X-embodiment dataset conversion.☆45Updated 8 months ago
- [ICRA 2025] RACER: Rich Language-Guided Failure Recovery Policies for Imitation Learning☆37Updated last year
- Official implementation of "OneTwoVLA: A Unified Vision-Language-Action Model with Adaptive Reasoning"☆194Updated 5 months ago
- ☆33Updated last year
- The official codebase for ManipLLM: Embodied Multimodal Large Language Model for Object-Centric Robotic Manipulation(cvpr 2024)☆142Updated last year
- [ICML 2025] OTTER: A Vision-Language-Action Model with Text-Aware Visual Feature Extraction☆109Updated 6 months ago
- Code for ICRA24 paper "Think, Act, and Ask: Open-World Interactive Personalized Robot Navigation" Paper//arxiv.org/abs/2310.07968 …☆31Updated last year
- ☆45Updated last year
- A list of robotics related papers accepted by ICLR'25☆22Updated 2 months ago
- [MMM 2025 Best Paper] RoLD: Robot Latent Diffusion for Multi-Task Policy Modeling☆22Updated last year
- Reimplementation of GR-1, a generalized policy for robotics manipulation.☆143Updated last year
- Data pre-processing and training code on Open-X-Embodiment with pytorch☆11Updated 9 months ago
- [CoRL 2023] REFLECT: Summarizing Robot Experiences for Failure Explanation and Correction☆101Updated last year
- Official Implementation of CAPEAM (ICCV'23)☆13Updated 11 months ago
- official repo for AGNOSTOS, a cross-task manipulation benchmark, and X-ICM method, a cross-task in-context manipulation (VLA) method☆49Updated last week
- ☆39Updated 4 months ago
- Code for Reinforcement Learning from Vision Language Foundation Model Feedback☆126Updated last year
- MOKA: Open-World Robotic Manipulation through Mark-based Visual Prompting (RSS 2024)☆89Updated last year
- ICCV2025☆140Updated 2 months ago
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆145Updated last month
- ☆43Updated 3 months ago
- 🦾 A Dual-System VLA with System2 Thinking☆114Updated 2 months ago
- [ICLR 2025 Oral] Seer: Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulation☆252Updated 4 months ago
- A simple testbed for robotics manipulation policies☆102Updated 6 months ago
- ☆38Updated 4 months ago