JiuTian-VL / Large-VLM-based-VLA-for-Robotic-ManipulationView external linksLinks
A curated list of large VLM-based VLA models for robotic manipulation.
☆340Dec 21, 2025Updated last month
Alternatives and similar repositories for Large-VLM-based-VLA-for-Robotic-Manipulation
Users that are interested in Large-VLM-based-VLA-for-Robotic-Manipulation are comparing it to the libraries listed below
Sorting:
- Spatial Aptitude Training for Multimodal Langauge Models☆24Feb 8, 2026Updated last week
- [ICRA 2025] Official implementation of Open-Nav: Exploring Zero-Shot Vision-and-Language Navigation in Continuous Environment with Open-S…☆122May 31, 2025Updated 8 months ago
- Official repository for LeLaN training and inference code☆131Sep 27, 2024Updated last year
- Implementation of paper: VLA-Touch: Enhancing Vision-Language-Action Models with Dual-Level Tactile Feedback☆54Jan 4, 2026Updated last month
- ☆28Dec 16, 2025Updated 2 months ago
- This is the source code to paper “DAgger Diffusion Navigation: DAgger Boosted Diffusion Policy for Vision-Language Navigation”.☆30Aug 13, 2025Updated 6 months ago
- [RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions☆984Nov 19, 2025Updated 2 months ago
- A curated list of state-of-the-art research in embodied AI, focusing on vision-language-action (VLA) models, vision-language navigation (…☆2,550Updated this week
- RynnVLA-002: A Unified Vision-Language-Action and World Model☆889Dec 2, 2025Updated 2 months ago
- [CVPR 2025] RoomTour3D - Geometry-aware, cheap and automatic data from web videos for embodied navigation☆69Mar 17, 2025Updated 11 months ago
- Can 3D Vision-Language Models Truly Understand Natural Language?☆20Mar 28, 2024Updated last year
- Official implementation of SPGrasp: A framework for dynamic grasp synthesis from sparse spatiotemporal prompts.☆19Jan 6, 2026Updated last month
- Official implementation of Get a Grip: Multi-Finger Grasp Evaluation at Scale Enables Robust Sim-to-Real Transfer☆32Feb 12, 2025Updated last year
- [NeurIPS 2025] CogVLA: Cognition-Aligned Vision-Language-Action Models via Instruction-Driven Routing & Sparsification☆130Dec 10, 2025Updated 2 months ago
- [ICLR 2026] SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning☆1,380Jan 6, 2026Updated last month
- ☆82Aug 20, 2025Updated 5 months ago
- Paper Survey for Visual Language Navigation☆70Updated this week
- ☆248Oct 30, 2025Updated 3 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆363Jan 4, 2026Updated last month
- Torque estimation framework for dynamic robotic actuators.☆15Sep 26, 2024Updated last year
- Learning Robot Locomotion from Diverse Datasets☆14Apr 24, 2025Updated 9 months ago
- source code for project instinct website☆15Feb 4, 2025Updated last year
- LLM-Powered Open-Vocabulary Scene Segmentation with Language Embedded 3D Gaussians☆22Jan 10, 2025Updated last year
- [AAAI 2025] The official implementation for the "Motion Decoupled 3D Gaussian Splatting for Dynamic Object Representation"☆18Jul 18, 2025Updated 6 months ago
- ☆22Oct 4, 2021Updated 4 years ago
- Official implementation of ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver.☆204Jan 25, 2026Updated 3 weeks ago
- Official implementation of the paper: "NavDP: Learning Sim-to-Real Navigation Diffusion Policy with Privileged Information Guidance"☆529Jan 12, 2026Updated last month
- ☆175Jan 19, 2026Updated 3 weeks ago
- [CVPR 2025] RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete. Official Repository.☆364Oct 13, 2025Updated 4 months ago
- GBC: Generalized Behavior-Cloning Framework for Whole-Body Humanoid Imitation☆43Jan 27, 2026Updated 3 weeks ago
- StereoVLA is powered by stereo vision and supports flexible deployment with high tolerance to camera pose variations.☆50Jan 12, 2026Updated last month
- ☆78Jan 11, 2026Updated last month
- RoboBrain 2.5: Advanced version of RoboBrain. Depth in Sight, Time in Mind. 🎉🎉🎉☆818Jan 27, 2026Updated 3 weeks ago
- Benchmarking Knowledge Transfer in Lifelong Robot Learning☆1,485Mar 15, 2025Updated 11 months ago
- Code of 3DMIT: 3D MULTI-MODAL INSTRUCTION TUNING FOR SCENE UNDERSTANDING☆31Jul 26, 2024Updated last year
- OpenVLA: An open-source vision-language-action model for robotic manipulation.☆5,251Mar 23, 2025Updated 10 months ago
- Official Repository for MolmoAct☆299Jan 13, 2026Updated last month
- LeIsaac provides teleoperation functionality in IsaacLab using the SO101Leader (LeRobot), including data collection, data conversion, and…☆535Jan 30, 2026Updated 2 weeks ago
- Galaxea's open-source VLA repository☆513Jan 17, 2026Updated last month