OpenGVLab / Instruct2ActLinks
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
☆372Updated last year
Alternatives and similar repositories for Instruct2Act
Users that are interested in Instruct2Act are comparing it to the libraries listed below
Sorting:
- [arXiv 2023] Embodied Task Planning with Large Language Models☆193Updated 2 years ago
- Implementation of "PaLM-E: An Embodied Multimodal Language Model"☆335Updated 2 years ago
- Code for RoboFlamingo☆421Updated last year
- Official Task Suite Implementation of ICML'23 Paper "VIMA: General Robot Manipulation with Multimodal Prompts"☆324Updated 2 years ago
- Pytorch implementation of the models RT-1-X and RT-2-X from the paper: "Open X-Embodiment: Robotic Learning Datasets and RT-X Models"☆234Updated 2 weeks ago
- VoxPoser: Composable 3D Value Maps for Robotic Manipulation with Language Models☆780Updated 11 months ago
- Democratization of RT-2 "RT-2: New model translates vision and language into action"☆548Updated last year
- Generating Robotic Simulation Tasks via Large Language Models☆346Updated last year
- [ICML 2024] 3D-VLA: A 3D Vision-Language-Action Generative World Model☆617Updated last year
- ☆430Updated 2 months ago
- Embodied Chain of Thought: A robotic policy that reason to solve the task.☆364Updated 10 months ago
- The official codebase for ManipLLM: Embodied Multimodal Large Language Model for Object-Centric Robotic Manipulation(cvpr 2024)☆146Updated last year
- [ICML 2024] LEO: An Embodied Generalist Agent in 3D World☆475Updated 9 months ago
- [CVPR 2025] RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete. Official Repository.☆364Updated 3 months ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆301Updated last year
- The Official Implementation of RoboMatrix☆104Updated 8 months ago
- Official Code for RVT-2 and RVT☆395Updated 11 months ago
- ☆86Updated 2 years ago
- Official Algorithm Implementation of ICML'23 Paper "VIMA: General Robot Manipulation with Multimodal Prompts"☆844Updated last year
- This repository compiles a list of papers related to the application of video technology in the field of robotics! Star⭐ the repo and fol…☆183Updated last year
- 🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.☆640Updated 7 months ago
- ☆262Updated last year
- [CoRL 2023] This repository contains data generation and training code for Scaling Up & Distilling Down☆405Updated last year
- Embodied Agent Interface (EAI): Benchmarking LLMs for Embodied Decision Making (NeurIPS D&B 2024 Oral)☆278Updated 11 months ago
- ☆36Updated 10 months ago
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.☆379Updated 2 months ago
- [ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy☆227Updated 10 months ago
- [IROS24 Oral]ManipVQA: Injecting Robotic Affordance and Physically Grounded Information into Multi-Modal Large Language Models☆99Updated last year
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.☆334Updated 4 months ago
- [AAAI'26 Oral] DexGraspVLA: A Vision-Language-Action Framework Towards General Dexterous Grasping☆465Updated 5 months ago