microsoft / GPT4Vision-Robot-Manipulation-PromptsLinks
This repository provides the sample code designed to interpret human demonstration videos and convert them into high-level tasks for robots.
☆44Updated last year
Alternatives and similar repositories for GPT4Vision-Robot-Manipulation-Prompts
Users that are interested in GPT4Vision-Robot-Manipulation-Prompts are comparing it to the libraries listed below
Sorting:
- PyTorch implementation of YAY Robot☆164Updated last year
- [IROS24 Oral]ManipVQA: Injecting Robotic Affordance and Physically Grounded Information into Multi-Modal Large Language Models☆97Updated last year
- ☆62Updated 7 months ago
- [ICRA 2025] PyTorch Code for Local Policies Enable Zero-shot Long-Horizon Manipulation☆126Updated 6 months ago
- An official implementation of Vision-Language Interpreter (ViLaIn)☆42Updated last year
- code implementation of GraspGPT and FoundationGrasp☆130Updated 2 weeks ago
- [CoRL2024] ThinkGrasp: A Vision-Language System for Strategic Part Grasping in Clutter. https://arxiv.org/abs/2407.11298☆107Updated 3 months ago
- ☆74Updated last year
- A library of long-horizon Task-and-Motion-Planning (TAMP) problems in kitchen and household scenes, as well as planners to solve them☆150Updated 5 months ago
- Manipulate-Anything: Automating Real-World Robots using Vision-Language Models [CoRL 2024]☆46Updated 7 months ago
- ☆38Updated last year
- Public release for "Distillation and Retrieving Generalizable Knowledge for Robot Manipulation via Language Corrections"☆46Updated last year
- Code for the paper: "Active Vision Might Be All You Need: Exploring Active Vision in Bimanual Robotic Manipulation"☆47Updated last month
- RLAfford: End-to-End Affordance Learning for Robotic Manipulation, ICRA 2023☆120Updated last year
- Official implementation for paper "EquiBot: SIM(3)-Equivariant Diffusion Policy for Generalizable and Data Efficient Learning".☆160Updated last year
- Code for the RA-L paper "Language Models as Zero-Shot Trajectory Generators" available at https://arxiv.org/abs/2310.11604.☆102Updated 7 months ago
- Official implementation of CoPa: General Robotic Manipulation through Spatial Constraints of Parts with Foundation Models☆97Updated 9 months ago
- This repository contains the implementation of Ground4Act, a two-stage approach for collaborative pushing and grasping in clutter using a…☆32Updated 7 months ago
- ACE: A Cross-platform Visual-Exoskeletons for Low-Cost Dexterous Teleoperation☆122Updated last year
- ☆48Updated last week
- [NeurIPS 2025 Spotlight 🎊] DexGarmentLab: Dexterous Garment Manipulation Environment with Generalizable Policy☆99Updated last month
- LLM3: Large Language Model-based Task and Motion Planning with Motion Failure Reasoning☆93Updated last year
- A collection of papers, codes and talks of visual imitation learning/imitation learning from video for robotics.☆79Updated 2 years ago
- [ICRA 2024] AirExo: Low-Cost Exoskeletons for Learning Whole-Arm Manipulation in the Wild☆45Updated last year
- [CoRL 2023] REFLECT: Summarizing Robot Experiences for Failure Explanation and Correction☆101Updated last year
- This is the official implementation of RoboBERT, which is a novel end-to-end mutiple-modality robotic operations training framework.☆56Updated 5 months ago
- ☆17Updated 8 months ago
- Code for Point Policy: Unifying Observations and Actions with Key Points for Robot Manipulation☆81Updated 3 months ago
- ☆39Updated 4 months ago
- This is the repo of CoRL 2024 paper "Learning to Manipulate Anywhere: A Visual Generalizable Framework For Reinforcement Learning"☆80Updated 10 months ago