Gary3410 / TaPA
[arXiv 2023] Embodied Task Planning with Large Language Models
☆185Updated last year
Alternatives and similar repositories for TaPA:
Users that are interested in TaPA are comparing it to the libraries listed below
- Embodied Chain of Thought: A robotic policy that reason to solve the task.☆236Updated last month
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.☆215Updated last week
- Embodied Agent Interface (EAI): Benchmarking LLMs for Embodied Decision Making (NeurIPS D&B 2024 Oral)☆194Updated 2 months ago
- Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model☆359Updated 10 months ago
- The official codebase for ManipLLM: Embodied Multimodal Large Language Model for Object-Centric Robotic Manipulation(cvpr 2024)☆131Updated 10 months ago
- ProgPrompt for Virtualhome☆133Updated last year
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"☆91Updated 2 months ago
- Code for RoboFlamingo☆374Updated last year
- ☆175Updated last year
- ☆136Updated last month
- [IROS24 Oral]ManipVQA: Injecting Robotic Affordance and Physically Grounded Information into Multi-Modal Large Language Models☆90Updated 8 months ago
- ☆340Updated 3 months ago
- Official Task Suite Implementation of ICML'23 Paper "VIMA: General Robot Manipulation with Multimodal Prompts"☆299Updated last year
- The Official Implementation of RoboMatrix☆90Updated 4 months ago
- Pytorch implementation of the models RT-1-X and RT-2-X from the paper: "Open X-Embodiment: Robotic Learning Datasets and RT-X Models"☆206Updated 2 weeks ago
- [ICCV'23] LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Models☆184Updated last month
- Code for MultiPLY: A Multisensory Object-Centric Embodied Large Language Model in 3D World☆128Updated 6 months ago
- Official code for the paper: Embodied Multi-Modal Agent trained by an LLM from a Parallel TextWorld☆56Updated 7 months ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆247Updated last year
- ☆102Updated 3 weeks ago
- Official repository of Learning to Act from Actionless Videos through Dense Correspondences.☆214Updated last year
- [ICLR 2025] LAPA: Latent Action Pretraining from Videos☆244Updated 3 months ago
- ☆83Updated last year
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization☆117Updated last month
- [CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"☆145Updated last month
- Official Implementation of ReALFRED (ECCV'24)☆39Updated 6 months ago
- Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success☆363Updated last week
- ☆29Updated 7 months ago
- The repo of paper `RoboMamba: Multimodal State Space Model for Efficient Robot Reasoning and Manipulation`☆116Updated 4 months ago
- ☆106Updated last year