Video Prediction Policy: A Generalist Robot Policy with Predictive Visual Representations https://video-prediction-policy.github.io
☆352May 17, 2025Updated 9 months ago
Alternatives and similar repositories for video-prediction-policy
Users that are interested in video-prediction-policy are comparing it to the libraries listed below
Sorting:
- [RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions☆1,017Nov 19, 2025Updated 3 months ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)☆342Jul 23, 2025Updated 7 months ago
- [RSS 2024] 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations☆1,274Oct 17, 2025Updated 4 months ago
- [IROS 2025] Generalizable Humanoid Manipulation with 3D Diffusion Policies. Part 1: Train & Deploy of iDP3☆508Jun 16, 2025Updated 8 months ago
- Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success☆1,057Sep 9, 2025Updated 6 months ago
- [CoRL 2024] Im2Flow2Act: Flow as the Cross-domain Manipulation Interface☆152Oct 17, 2024Updated last year
- Code for the paper "3D Diffuser Actor: Policy Diffusion with 3D Scene Representations"☆384Aug 17, 2024Updated last year
- ICCV 2025 | TesserAct: Learning 4D Embodied World Models☆382Aug 4, 2025Updated 7 months ago
- Evaluating and reproducing real-world robot manipulation policies (e.g., RT-1, RT-1-X, Octo) in simulation under common setups (e.g., Goo…☆991Dec 20, 2025Updated 2 months ago
- RDT-1B: a Diffusion Foundation Model for Bimanual Manipulation☆1,632Jan 21, 2026Updated last month
- ☆89Sep 23, 2025Updated 5 months ago
- [ICLR 2025] LAPA: Latent Action Pretraining from Videos☆478Jan 22, 2025Updated last year
- [RSS25] Official implementation of DemoGen: Synthetic Demonstration Generation for Data-Efficient Visuomotor Policy Learning☆239Jul 18, 2025Updated 7 months ago
- OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulation☆347Aug 27, 2025Updated 6 months ago
- Unfied World Models: Coupling Video and Action Diffusion for Pretraining on Large Robotic Datasets☆190Oct 8, 2025Updated 5 months ago
- This is the official implementation of the paper "ConRFT: A Reinforced Fine-tuning Method for VLA Models via Consistency Policy".☆329Nov 11, 2025Updated 3 months ago
- [ICLR 2025 Oral] Seer: Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulation☆280Jul 8, 2025Updated 8 months ago
- CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks☆848Sep 8, 2025Updated 6 months ago
- Re-implementation of pi0 vision-language-action (VLA) model from Physical Intelligence☆1,404Jan 31, 2025Updated last year
- [IROS 2025 Best Paper Award Finalist & IEEE TRO 2026] The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems☆2,804Dec 16, 2025Updated 2 months ago
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.☆395Nov 11, 2025Updated 3 months ago
- Official codebase for "Any-point Trajectory Modeling for Policy Learning"☆273Jun 19, 2025Updated 8 months ago
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆339Oct 3, 2025Updated 5 months ago
- [CVPR 2025] RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete. Official Repository.☆371Oct 13, 2025Updated 4 months ago
- Code for FLIP: Flow-Centric Generative Planning for General-Purpose Manipulation Tasks☆79Dec 12, 2024Updated last year
- ☆247May 12, 2025Updated 9 months ago
- ☆10,475Dec 27, 2025Updated 2 months ago
- [CVPR 2024] Hierarchical Diffusion Policy for Multi-Task Robotic Manipulation☆229Apr 9, 2024Updated last year
- ☆443Nov 29, 2025Updated 3 months ago
- ICCV2025☆161Dec 10, 2025Updated 3 months ago
- Code for Point Policy: Unifying Observations and Actions with Key Points for Robot Manipulation☆90Jul 21, 2025Updated 7 months ago
- [ICML 2024] 3D-VLA: A 3D Vision-Language-Action Generative World Model☆623Oct 29, 2024Updated last year
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆164Oct 1, 2025Updated 5 months ago
- OpenVLA: An open-source vision-language-action model for robotic manipulation.☆5,461Mar 23, 2025Updated 11 months ago
- code for the paper Predicting Point Tracks from Internet Videos enables Diverse Zero-Shot Manipulation☆100Jul 31, 2024Updated last year
- [ICLR 2026] SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning☆1,449Jan 6, 2026Updated 2 months ago
- [CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"☆234Nov 6, 2025Updated 4 months ago
- ☆75Jan 8, 2025Updated last year
- RynnVLA-002: A Unified Vision-Language-Action and World Model☆912Dec 2, 2025Updated 3 months ago