DelinQu / awesome-vision-language-action-model
Latest Advances on Vison-Language-Action Models.
☆17Updated 2 months ago
Alternatives and similar repositories for awesome-vision-language-action-model:
Users that are interested in awesome-vision-language-action-model are comparing it to the libraries listed below
- ManiCM: Real-time 3D Diffusion Policy via Consistency Model for Robotic Manipulation☆92Updated 7 months ago
- G3Flow: Generative 3D Semantic Flow for Pose-aware and Generalizable Object Manipulation☆33Updated 2 weeks ago
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization☆76Updated 2 weeks ago
- [CoRL 2023 Oral] GNFactor: Multi-Task Real Robot Learning with Generalizable Neural Feature Fields☆127Updated last year
- Code for FLIP: Flow-Centric Generative Planning for General-Purpose Manipulation Tasks☆41Updated 2 months ago
- Official implementation of "Towards Generalizable Vision-Language Robotic Manipulation: A Benchmark and LLM-guided 3D Policy."☆49Updated last week
- OVExp: Open Vocabulary Exploration for Object-Oriented Navigation☆33Updated 7 months ago
- Efficiently apply modification functions to RLDS/TFDS datasets.☆24Updated 8 months ago
- [ECCV 2024] ManiGaussian: Dynamic Gaussian Splatting for Multi-task Robotic Manipulation☆207Updated 3 months ago
- Dreamitate: Real-World Visuomotor Policy Learning via Video Generation (CoRL 2024)☆43Updated 7 months ago
- ☆61Updated 5 months ago
- ☆91Updated 6 months ago
- Hand-object interaction Pretraining From Videos☆71Updated 3 months ago
- ☆43Updated 2 months ago
- [CoRL 2024] Im2Flow2Act: Flow as the Cross-domain Manipulation Interface☆96Updated 4 months ago
- Human Demo Videos to Robot Action Plans☆43Updated 3 months ago
- code for the paper Predicting Point Tracks from Internet Videos enables Diverse Zero-Shot Manipulation☆76Updated 6 months ago
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"☆66Updated this week
- ☆44Updated last month
- ☆61Updated 3 months ago
- [CoRL2024] Official repo of `A3VLM: Actionable Articulation-Aware Vision Language Model`☆104Updated 4 months ago
- A Vision-Language Model for Spatial Affordance Prediction in Robotics☆98Updated last week
- [NeurIPS 2024 D&B] Point Cloud Matters: Rethinking the Impact of Different Observation Spaces on Robot Learning☆63Updated 4 months ago
- [ECCV 2024] 🎉 Official repository of "Robo-ABC: Affordance Generalization Beyond Categories via Semantic Correspondence for Robot Manipu…☆71Updated 2 months ago
- [CoRL 2024] RoboEXP: Action-Conditioned Scene Graph via Interactive Exploration for Robotic Manipulation☆92Updated 4 months ago
- [ICLR 2025🎉] This is the official implementation of paper "Robots Pre-Train Robots: Manipulation-Centric Robotic Representation from Lar…☆60Updated 3 weeks ago