YuZhaoshu / Efficient-VLAs-SurveyLinks
π₯This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the repository, so follow us to keep up with the latest developments!!!
β75Updated last week
Alternatives and similar repositories for Efficient-VLAs-Survey
Users that are interested in Efficient-VLAs-Survey are comparing it to the libraries listed below
Sorting:
- StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developingβ433Updated last week
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintainedπ₯]β172Updated 3 weeks ago
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Modelβ318Updated last month
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"β117Updated 9 months ago
- β358Updated 3 weeks ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policyβ269Updated last week
- OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulationβ323Updated 2 months ago
- Paper list in the survey: A Survey on Vision-Language-Action Models: An Action Tokenization Perspectiveβ323Updated 4 months ago
- [NeurIPS 2025] VLA-Cache: Towards Efficient Vision-Language-Action Model via Adaptive Token Caching in Robotic Manipulationβ40Updated 2 months ago
- WorldVLA: Towards Autoregressive Action World Modelβ539Updated last month
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledgeβ224Updated 2 months ago
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).β121Updated last year
- Unified Vision-Language-Action Modelβ226Updated last month
- [AAAI 2026] Official code for MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Manβ¦β51Updated 3 months ago
- Latest Advances on Vison-Language-Action Models.β119Updated 8 months ago
- A curated list of large VLM-based VLA models for robotic manipulation.β250Updated last week
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"β195Updated 3 weeks ago
- Galaxea's first VLA releaseβ312Updated 3 weeks ago
- The repo of paper `RoboMamba: Multimodal State Space Model for Efficient Robot Reasoning and Manipulation`β139Updated 10 months ago
- Single-file implementation to advance vision-language-action (VLA) models with reinforcement learning.β338Updated last week
- [NeurIPS 2025 Spotlight] SoFar: Language-Grounded Orientation Bridges Spatial Reasoning and Object Manipulationβ203Updated 4 months ago
- Building General-Purpose Robots Based on Embodied Foundation Modelβ592Updated this week
- Running VLA at 30Hz frame rate and 480Hz trajectory frequencyβ242Updated last week
- Dexbotic: Open-Source Vision-Language-Action Toolboxβ467Updated 2 weeks ago
- π₯ SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.β570Updated 4 months ago
- SimpleVLA-RL: Scaling VLA Training via Reinforcement Learningβ987Updated last month
- F1: A Vision Language Action Model Bridging Understanding and Generation to Actionsβ135Updated last month
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.β317Updated 2 months ago
- ICCV2025β142Updated this week
- A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulationβ374Updated 3 weeks ago