OpenHelix-Team / LLaVA-VLALinks
LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintainedπ₯]
β88Updated this week
Alternatives and similar repositories for LLaVA-VLA
Users that are interested in LLaVA-VLA are comparing it to the libraries listed below
Sorting:
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Modelβ254Updated last month
- ICCV2025β103Updated 2 weeks ago
- Official implementation of "OneTwoVLA: A Unified Vision-Language-Action Model with Adaptive Reasoning"β147Updated last month
- Latest Advances on Vison-Language-Action Models.β83Updated 4 months ago
- β242Updated 3 months ago
- SoFar: Language-Grounded Orientation Bridges Spatial Reasoning and Object Manipulationβ175Updated 2 weeks ago
- β73Updated 2 months ago
- Single-file implementation to advance vision-language-action (VLA) models with reinforcement learning.β163Updated last week
- OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulationβ216Updated last month
- WorldVLA: Towards Autoregressive Action World Modelβ248Updated last week
- A comprehensive list of papers about dual-system VLA models, including papers, codes, and related websites.β53Updated last month
- β55Updated 4 months ago
- β70Updated last month
- The Official Implementation of RoboMatrixβ93Updated last month
- [CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"β184Updated 3 months ago
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"β99Updated 5 months ago
- GRAPE: Guided-Reinforced Vision-Language-Action Preference Optimizationβ132Updated 3 months ago
- [RSS 2024] Code for "Multimodal Diffusion Transformer: Learning Versatile Behavior from Multimodal Goals" for CALVIN experiments with preβ¦β145Updated 9 months ago
- This repository summarizes recent advances in the VLA + RL paradigm and provides a taxonomic classification of relevant works.β166Updated 3 weeks ago
- π₯ SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.β392Updated 3 weeks ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"β263Updated last year
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.β253Updated 2 weeks ago
- [ICCV 2025] Latent Motion Token as the Bridging Language for Robot Manipulationβ110Updated 2 months ago
- Online RL with Simple Reward Enables Training VLA Models with Only One Trajectoryβ269Updated 3 weeks ago
- [ICLR 2025 Oral] Seer: Predictive Inverse Dynamics Models are Scalable Learners for Robotic Manipulationβ205Updated last week
- β80Updated last month
- [NeurIPS 2024] CLOVER: Closed-Loop Visuomotor Control with Generative Expectation for Robotic Manipulationβ119Updated 2 weeks ago
- [ICLR 2025] LAPA: Latent Action Pretraining from Videosβ326Updated 5 months ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)β232Updated 2 weeks ago
- [Arxiv 2025: MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Manipulation]β36Updated 3 months ago