[ICLR 2025] LAPA: Latent Action Pretraining from Videos
☆506Jan 22, 2025Updated last year
Alternatives and similar repositories for LAPA
Users that are interested in LAPA are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- [ICCV2025 Oral] Latent Motion Token as the Bridging Language for Learning Robot Manipulation from Videos☆171Oct 1, 2025Updated 6 months ago
- [RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions☆1,049Nov 19, 2025Updated 4 months ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)☆364Jul 23, 2025Updated 8 months ago
- Evaluating and reproducing real-world robot manipulation policies (e.g., RT-1, RT-1-X, Octo) in simulation under common setups (e.g., Goo…☆1,032Dec 20, 2025Updated 3 months ago
- ☆461Nov 29, 2025Updated 4 months ago
- Virtual machines for every use case on DigitalOcean • AdGet dependable uptime with 99.99% SLA, simple security tools, and predictable monthly pricing with DigitalOcean's virtual machines, called Droplets.
- [ICLR 2025 Oral] Seer: Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulation☆290Jul 8, 2025Updated 9 months ago
- Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success☆1,132Sep 9, 2025Updated 7 months ago
- A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation☆417Oct 30, 2025Updated 5 months ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆307Apr 22, 2024Updated last year
- [ICML 2024] 3D-VLA: A 3D Vision-Language-Action Generative World Model☆619Oct 29, 2024Updated last year
- Re-implementation of pi0 vision-language-action (VLA) model from Physical Intelligence☆1,443Jan 31, 2025Updated last year
- RDT-1B: a Diffusion Foundation Model for Bimanual Manipulation☆1,668Jan 21, 2026Updated 2 months ago
- Code for the ICLR 2024 spotlight paper: "Learning to Act without Actions" (introducing Latent Action Policies)☆138Jul 31, 2024Updated last year
- OpenVLA: An open-source vision-language-action model for robotic manipulation.☆5,874Mar 23, 2025Updated last year
- AI Agents on DigitalOcean Gradient AI Platform • AdBuild production-ready AI agents using customizable tools or access multiple LLMs through a single endpoint. Create custom knowledge bases or connect external data.
- Embodied Chain of Thought: A robotic policy that reason to solve the task.☆381Apr 5, 2025Updated last year
- [CoRL25] GraspVLA: a Grasping Foundation Model Pre-trained on Billion-scale Synthetic Action Data☆362Dec 29, 2025Updated 3 months ago
- [CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"☆234Nov 6, 2025Updated 5 months ago
- CALVIN - A benchmark for Language-Conditioned Policy Learning for Long-Horizon Robot Manipulation Tasks☆880Sep 8, 2025Updated 7 months ago
- [IROS 2025 Best Paper Award Finalist & IEEE TRO 2026] The Large-scale Manipulation Platform for Scalable and Intelligent Embodied Systems☆2,905Dec 16, 2025Updated 4 months ago
- ☆51Apr 15, 2025Updated last year
- Heterogeneous Pre-trained Transformer (HPT) as Scalable Policy Learner.☆535Dec 6, 2024Updated last year
- Octo is a transformer-based robot policy trained on a diverse mix of 800k robot trajectories.☆1,611Jul 31, 2024Updated last year
- ☆11,178Mar 29, 2026Updated 2 weeks ago
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- Repo for Bring Your Own Vision-Language-Action (VLA) model, arxiv 2024☆37Jan 22, 2025Updated last year
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆346Oct 3, 2025Updated 6 months ago
- Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning☆79May 17, 2025Updated 10 months ago
- Official implementation of GR-MG☆92Jan 12, 2025Updated last year
- 🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.☆677Jun 23, 2025Updated 9 months ago
- Benchmarking Knowledge Transfer in Lifelong Robot Learning☆1,692Mar 15, 2025Updated last year
- Code for FLIP: Flow-Centric Generative Planning for General-Purpose Manipulation Tasks☆83Dec 12, 2024Updated last year
- Code for the paper "3D Diffuser Actor: Policy Diffusion with 3D Scene Representations"☆388Aug 17, 2024Updated last year
- RoboVerse: Towards a Unified Platform, Dataset and Benchmark for Scalable and Generalizable Robot Learning☆1,708Apr 8, 2026Updated last week
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- ReKep: Spatio-Temporal Reasoning of Relational Keypoint Constraints for Robotic Manipulation☆931Feb 20, 2025Updated last year
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimization☆159Apr 6, 2025Updated last year
- Official repository of Learning to Act from Actionless Videos through Dense Correspondences.☆250Apr 25, 2024Updated last year
- DynaMo: In-Domain Dynamics Pretraining for Visuo-Motor Control☆117Oct 27, 2024Updated last year
- [IROS 2025] Generalizable Humanoid Manipulation with 3D Diffusion Policies. Part 1: Train & Deploy of iDP3☆521Jun 16, 2025Updated 10 months ago
- ICCV2025☆164Dec 10, 2025Updated 4 months ago
- [RSS 2024] 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations☆1,313Oct 17, 2025Updated 5 months ago