guanweifan / awesome-efficient-vlaLinks
🔥 A curated roadmap to the Efficient VLA landscape. We’re keeping this list live—contribute your latest work!
☆72Updated this week
Alternatives and similar repositories for awesome-efficient-vla
Users that are interested in awesome-efficient-vla are comparing it to the libraries listed below
Sorting:
- 🔥This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the r…☆118Updated 3 weeks ago
- Paper list in the survey: A Survey on Vision-Language-Action Models: An Action Tokenization Perspective☆419Updated 6 months ago
- [NeurIPS'24]Efficient and accurate memory saving method towards W4A4 large multi-modal models.☆93Updated last year
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).☆121Updated last year
- ☆489Updated 3 months ago
- siiRL: Shanghai Innovation Institute RL Framework for Advanced LLMs and Multi-Agent Systems☆330Updated this week
- RynnVLA-002: A Unified Vision-Language-Action and World Model☆859Updated last month
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintained🔥]☆174Updated 3 months ago
- [ACM CSUR 2025] Understanding World or Predicting Future? A Comprehensive Survey of World Models☆413Updated 2 months ago
- The first decoder-only multimodal state space model☆100Updated 8 months ago
- 🏆 Official implementation of LangCoop: Collaborative Driving with Natural Language☆75Updated 4 months ago
- StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing☆949Updated this week
- [NeurIPS 2025] Official code implementation of Perception R1: Pioneering Perception Policy with Reinforcement Learning☆285Updated 6 months ago
- Running VLA at 30Hz frame rate and 480Hz trajectory frequency☆393Updated last week
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.☆332Updated 4 months ago
- Unleashing the Power of VLMs in Autonomous Driving via Reinforcement Learning and Reasoning☆314Updated 10 months ago
- Building General-Purpose Robots Based on Embodied Foundation Model☆731Updated last month
- [ICML'25] Official implementation of paper "SparseVLM: Visual Token Sparsification for Efficient Vision-Language Model Inference" and "Sp…☆231Updated last month
- ☆37Updated last week
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆336Updated 3 months ago
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledge☆282Updated 3 weeks ago
- DrivePI: Spatial-aware 4D MLLM for Unified Autonomous Driving Understanding, Perception, Prediction and Planning☆76Updated last month
- LightVLA☆73Updated last month
- [NeurIPS 24] MoE Jetpack: From Dense Checkpoints to Adaptive Mixture of Experts for Vision Tasks☆134Updated last year
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"☆124Updated 11 months ago
- 📚 Collection of token-level model compression resources.☆189Updated 4 months ago
- Doe-1: Closed-Loop Autonomous Driving with Large World Model☆116Updated last year
- [ICLR 2026] Unified Vision-Language-Action Model☆268Updated 3 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆344Updated 3 weeks ago
- ☆87Updated 8 months ago