XiaomiMiMo / MiMo-EmbodiedLinks
MiMo-Embodied
☆345Updated 2 months ago
Alternatives and similar repositories for MiMo-Embodied
Users that are interested in MiMo-Embodied are comparing it to the libraries listed below
Sorting:
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).☆121Updated last year
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasks☆186Updated 4 months ago
- A Pragmatic VLA Foundation Model☆247Updated this week
- ☆61Updated 10 months ago
- LLaVA-VLA: A Simple Yet Powerful Vision-Language-Action Model [Actively Maintained🔥]☆174Updated 3 months ago
- Official repository for "RLVR-World: Training World Models with Reinforcement Learning" (NeurIPS 2025), https://arxiv.org/abs/2505.13934☆204Updated 3 months ago
- [ICLR 2026] Unified Vision-Language-Action Model☆268Updated 3 months ago
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"☆222Updated last month
- ☆104Updated 3 months ago
- RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulation☆281Updated last week
- Visual Embodied Brain: Let Multimodal Large Language Models See, Think, and Control in Spaces☆87Updated 7 months ago
- ☆64Updated last month
- RynnVLA-002: A Unified Vision-Language-Action and World Model☆866Updated 2 months ago
- NORA: A Small Open-Sourced Generalist Vision Language Action Model for Embodied Tasks☆206Updated 3 weeks ago
- Multi-SpatialMLLM Multi-Frame Spatial Understanding with Multi-Modal Large Language Models☆167Updated 3 months ago
- ☆379Updated last week
- Official implementation for BitVLA: 1-bit Vision-Language-Action Models for Robotics Manipulation☆102Updated 6 months ago
- Cosmos-RL is a flexible and scalable Reinforcement Learning framework specialized for Physical AI applications.☆304Updated this week
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.☆332Updated 4 months ago
- VLA-RFT: Vision-Language-Action Models with Reinforcement Fine-Tuning☆121Updated 3 months ago
- InternVLA-A1: Unifying Understanding, Generation, and Action for Robotic Manipulation☆305Updated last week
- Official Code for EnerVerse-AC: Envisioning EmbodiedEnvironments with Action Condition☆144Updated 6 months ago
- The official implementation of "DynamicVLA: A Vision-Language-Action Model for Dynamic Object Manipulation". (arXiv 2601.22153)☆69Updated this week
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, …☆203Updated 8 months ago
- HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model☆336Updated 3 months ago
- Official code for "Embodied-R1: Reinforced Embodied Reasoning for General Robotic Manipulation"☆121Updated 5 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆344Updated 3 weeks ago
- Official code for EWMBench: Evaluating Scene, Motion, and Semantic Quality in Embodied World Models☆97Updated 7 months ago
- [CoRL 2024] VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding☆129Updated 8 months ago
- EO: Open-source Unified Embodied Foundation Model Series☆44Updated 2 weeks ago