LiAutoAD / LightVLALinks
Official Code for LightVLA (ICRA 2026)
☆74Updated last week
Alternatives and similar repositories for LightVLA
Users that are interested in LightVLA are comparing it to the libraries listed below
Sorting:
- ☆87Updated 8 months ago
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).☆121Updated last year
- [NeurIPS 2025] VLA-Cache: Towards Efficient Vision-Language-Action Model via Adaptive Token Caching in Robotic Manipulation☆65Updated 4 months ago
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"☆226Updated last month
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintained🔥]☆175Updated 3 months ago
- [ACM CSUR 2025] Understanding World or Predicting Future? A Comprehensive Survey of World Models☆457Updated 2 months ago
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆336Updated 4 months ago
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledge☆286Updated last month
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasks☆186Updated 4 months ago
- 🔥This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the r…☆126Updated last month
- Latest Advances on Vison-Language-Action Models.☆128Updated 11 months ago
- [NeurIPS 2025]⭐️ Reason-RFT: Reinforcement Fine-Tuning for Visual Reasoning.☆267Updated 4 months ago
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.☆334Updated 4 months ago
- A Multi-Modal Large Language Model with Retrieval-augmented In-context Learning capacity designed for generalisable and explainable end-t…☆120Updated last year
- Official code for "Embodied-R1: Reinforced Embodied Reasoning for General Robotic Manipulation"☆121Updated 5 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆355Updated last month
- [ICLR 2026] Unified Vision-Language-Action Model☆273Updated 3 months ago
- Official code of paper "DeeR-VLA: Dynamic Inference of Multimodal Large Language Models for Efficient Robot Execution"☆124Updated 11 months ago
- MiMo-Embodied☆349Updated 2 months ago
- Nav-R1: Reasoning and Navigation in Embodied Scenes☆110Updated 3 months ago
- Paper list in the survey: A Survey on Vision-Language-Action Models: An Action Tokenization Perspective☆431Updated 7 months ago
- 🔥 A curated roadmap to the Efficient VLA landscape. We’re keeping this list live—contribute your latest work!☆74Updated last week
- 🏆 Official implementation of LangCoop: Collaborative Driving with Natural Language☆76Updated 4 months ago
- Official Implementation of Paper: WMPO: World Model-based Policy Optimization for Vision-Language-Action Models☆146Updated last month
- This repository collects papers on VLLM applications. We will update new papers irregularly.☆202Updated last month
- Embodied Question Answering (EQA) benchmark and method (ICCV 2025)☆46Updated 5 months ago
- ☆452Updated last week
- VLA-RFT: Vision-Language-Action Models with Reinforcement Fine-Tuning☆124Updated 4 months ago
- Code of the paper "EvolveNav: Empowering LLM-Based Vision-Language Navigation via Self-Improving Embodied Reasoning"☆28Updated 3 months ago
- [AAAI-25 Oral] Official Implementation of "FLAME: Learning to Navigate with Multimodal LLM in Urban Environments"☆69Updated 3 months ago