YuZhaoshu / Efficient-VLAs-SurveyLinks
π₯This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the repository, so follow us to keep up with the latest developments!!!
β91Updated 3 weeks ago
Alternatives and similar repositories for Efficient-VLAs-Survey
Users that are interested in Efficient-VLAs-Survey are comparing it to the libraries listed below
Sorting:
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Modelβ324Updated 2 months ago
- Paper list in the survey: A Survey on Vision-Language-Action Models: An Action Tokenization Perspectiveβ354Updated 5 months ago
- [NeurIPS 2025] VLA-Cache: Towards Efficient Vision-Language-Action Model via Adaptive Token Caching in Robotic Manipulationβ47Updated 2 months ago
- Evo-1: Lightweight Vision-Language-Action Model with Preserved Semantic Alignmentβ170Updated last week
- StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developingβ513Updated 3 weeks ago
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"β120Updated 9 months ago
- β380Updated last month
- OpenHelix: An Open-source Dual-System VLA Model for Robotic Manipulationβ330Updated 3 months ago
- [AAAI 2026] Official code for MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Manβ¦β56Updated 4 months ago
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintainedπ₯]β173Updated last month
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policyβ305Updated last month
- Galaxea's first VLA releaseβ317Updated last month
- The repo of paper `RoboMamba: Multimodal State Space Model for Efficient Robot Reasoning and Manipulation`β143Updated 11 months ago
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledgeβ245Updated 2 months ago
- Running VLA at 30Hz frame rate and 480Hz trajectory frequencyβ297Updated last week
- Official Code For VLA-OS.β128Updated 5 months ago
- Latest Advances on Vison-Language-Action Models.β120Updated 9 months ago
- Dexbotic: Open-Source Vision-Language-Action Toolboxβ554Updated last week
- Building General-Purpose Robots Based on Embodied Foundation Modelβ626Updated 2 weeks ago
- A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulationβ383Updated last month
- A curated list of large VLM-based VLA models for robotic manipulation.β280Updated 3 weeks ago
- Heterogeneous Pre-trained Transformer (HPT) as Scalable Policy Learner.β519Updated last year
- Unified Vision-Language-Action Modelβ245Updated last month
- π₯ SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.β584Updated 5 months ago
- β12Updated 9 months ago
- Real-Time VLAs via Future-state-aware Asynchronous Inference.β207Updated this week
- Single-file implementation to advance vision-language-action (VLA) models with reinforcement learning.β362Updated last month
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.β319Updated 2 months ago
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).β122Updated last year
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"β207Updated last month