nvidia-cosmos / cosmos-reason2Links
Cosmos-Reason2 models understand the physical common sense and generate appropriate embodied decisions in natural language through long chain-of-thought reasoning processes.
☆149Updated last week
Alternatives and similar repositories for cosmos-reason2
Users that are interested in cosmos-reason2 are comparing it to the libraries listed below
Sorting:
- MiMo-Embodied☆345Updated 2 months ago
- Cosmos-RL is a flexible and scalable Reinforcement Learning framework specialized for Physical AI applications.☆304Updated this week
- ☆133Updated 3 months ago
- Cosmos-Predict2.5, the latest version of the Cosmos World Foundation Models (WFMs) family, specialized for simulating and predicting the …☆735Updated last week
- NORA: A Small Open-Sourced Generalist Vision Language Action Model for Embodied Tasks☆206Updated 3 weeks ago
- Cosmos-Transfer2.5, built on top of Cosmos-Predict2.5, produces high-quality world simulations conditioned on multiple spatial control in…☆422Updated this week
- Open source repo for Locate 3D Model, 3D-JEPA and Locate 3D Dataset☆410Updated 7 months ago
- Cosmos-Predict1 is a collection of general-purpose world foundation models for Physical AI that can be fine-tuned into customized world m…☆397Updated 3 weeks ago
- [ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy☆226Updated 10 months ago
- [CVPR 2025] Source codes for the paper "3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning"☆210Updated 3 months ago
- ☆58Updated 8 months ago
- A Large-scale Video Action Dataset☆341Updated 2 weeks ago
- [CoRL 2024] VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding☆128Updated 8 months ago
- Official Reporsitory of "RoboEngine: Plug-and-Play Robot Data Augmentation with Semantic Robot Segmentation and Background Generation"☆146Updated 8 months ago
- ☆169Updated 11 months ago
- VLA-0: Building State-of-the-Art VLAs with Zero Modification☆436Updated 3 weeks ago
- Scaling Vision Pre-Training to 4K Resolution☆221Updated 3 weeks ago
- [ICCV 2025] Detect Anything 3D in the Wild☆245Updated last month
- Robot Utility Models are trained on a diverse set of environments and objects, and then can be deployed in novel environments with novel …☆240Updated last week
- Nvidia GEAR Lab's initiative to solve the robotics data problem using world models☆453Updated 3 months ago
- [ICCV 2025] A Simple yet Effective Pathway to Empowering LLaVA to Understand and Interact with 3D World☆369Updated 3 months ago
- Multi-SpatialMLLM Multi-Frame Spatial Understanding with Multi-Modal Large Language Models☆167Updated 3 months ago
- OpenVLA: An open-source vision-language-action model for robotic manipulation.☆329Updated 10 months ago
- Detect corn stalks for micro-sensor insertion☆13Updated last year
- ☆112Updated last week
- ☆49Updated 7 months ago
- VaViM and VaVAM: Autonomous Driving through Video Generative Modeling (official repository).☆139Updated 6 months ago
- [ICLR 2026] Unified Vision-Language-Action Model☆268Updated 3 months ago
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"☆222Updated last month
- Spot Sim2Real Infrastructure☆100Updated 8 months ago