om-ai-lab / OpenTrackVLALinks
Open & Reproducible Research for Tracking VLAs
☆125Updated 3 weeks ago
Alternatives and similar repositories for OpenTrackVLA
Users that are interested in OpenTrackVLA are comparing it to the libraries listed below
Sorting:
- [CoRL 2025] Repository relating to "TrackVLA: Embodied Visual Tracking in the Wild"☆323Updated last month
- The offical Implementation of "Soft-Prompted Transformer as Scalable Cross-Embodiment Vision-Language-Action Model"☆436Updated last week
- ☆166Updated this week
- Official implementation of the paper: "StreamVLN: Streaming Vision-and-Language Navigation via SlowFast Context Modeling"☆376Updated 2 months ago
- The Official Implementation of RoboMatrix☆104Updated 8 months ago
- 🤖 RoboOS: A Universal Embodied Operating System for Cross-Embodied and Multi-Robot Collaboration☆280Updated last month
- ✨✨【NeurIPS 2025】Official implementation of BridgeVLA☆164Updated 4 months ago
- Vision-Language Navigation Benchmark in Isaac Lab☆288Updated 4 months ago
- Official implementation of "OneTwoVLA: A Unified Vision-Language-Action Model with Adaptive Reasoning"☆206Updated 7 months ago
- RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulation☆277Updated last month
- A curated list of large VLM-based VLA models for robotic manipulation.☆313Updated last month
- ☆87Updated 8 months ago
- ☆190Updated 9 months ago
- PoliFormer: Scaling On-Policy RL with Transformers Results in Masterful Navigators☆105Updated last year
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆342Updated 2 weeks ago
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆333Updated 3 months ago
- Official GitHub Repository for Paper "Bridging Zero-shot Object Navigation and Foundation Models through Pixel-Guided Navigation Skill", …☆128Updated last year
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintained🔥]☆174Updated 2 months ago
- [RSS'25] This repository is the implementation of "NaVILA: Legged Robot Vision-Language-Action Model for Navigation"☆479Updated 5 months ago
- [RSS 2024 & RSS 2025] VLN-CE evaluation code of NaVid and Uni-NaVid☆361Updated 3 months ago
- [ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models☆234Updated last year
- Fast-in-Slow: A Dual-System Foundation Model Unifying Fast Manipulation within Slow Reasoning☆138Updated 5 months ago
- 🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.☆632Updated 7 months ago
- [TMLR 2024] repository for VLN with foundation models☆242Updated 2 months ago
- [CVPR 2025] The offical Implementation of "Universal Actions for Enhanced Embodied Foundation Models"☆224Updated 2 months ago
- [ICRA 2025] Official implementation of Open-Nav: Exploring Zero-Shot Vision-and-Language Navigation in Continuous Environment with Open-S…☆112Updated 7 months ago
- Official Code for "From Cognition to Precognition: A Future-Aware Framework for Social Navigation" (ICRA 2025)☆99Updated last month
- ☆28Updated 3 weeks ago
- Code of the paper "NavCoT: Boosting LLM-Based Vision-and-Language Navigation via Learning Disentangled Reasoning" (TPAMI 2025)☆126Updated 7 months ago
- [RSS 2025] Uni-NaVid: A Video-based Vision-Language-Action Model for Unifying Embodied Navigation Tasks.☆214Updated last month