facebookresearch / nwmLinks
Official code for the CVPR 2025 paper "Navigation World Models".
☆493Updated last month
Alternatives and similar repositories for nwm
Users that are interested in nwm are comparing it to the libraries listed below
Sorting:
- Official implementation of the paper: "StreamVLN: Streaming Vision-and-Language Navigation via SlowFast Context Modeling"☆365Updated 2 months ago
- ☆222Updated 5 months ago
- Official implementation of Spatial-MLLM: Boosting MLLM Capabilities in Visual-based Spatial Intelligence☆419Updated last month
- [NeurIPS 2025] Official implementation of "RoboRefer: Towards Spatial Referring with Reasoning in Vision-Language Models for Robotics"☆218Updated 3 weeks ago
- Official PyTorch Implementation of Unified Video Action Model (RSS 2025)☆308Updated 5 months ago
- Unified Vision-Language-Action Model☆257Updated 2 months ago
- [ICLR 2025 Spotlight] MetaUrban: An Embodied AI Simulation Platform for Urban Micromobility☆218Updated 2 months ago
- InternRobotics' open platform for building generalized navigation foundation models.☆565Updated last week
- [CVPR2025] CityWalker: Learning Embodied Urban Navigation from Web-Scale Videos☆171Updated 3 months ago
- [CVPR 2025] Source codes for the paper "3D-Mem: 3D Scene Memory for Embodied Exploration and Reasoning"☆205Updated 3 months ago
- [NeurIPS 2025] DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledge☆265Updated 3 months ago
- [ICLR 2025] SPA: 3D Spatial-Awareness Enables Effective Embodied Representation☆171Updated 6 months ago
- [ICCV 2025] A Simple yet Effective Pathway to Empowering LLaVA to Understand and Interact with 3D World☆361Updated 2 months ago
- [NeurIPS'24] This repository is the implementation of "SpatialRGPT: Grounded Spatial Reasoning in Vision Language Models"☆303Updated last year
- [CVPR'25] SeeGround: See and Ground for Zero-Shot Open-Vocabulary 3D Visual Grounding☆201Updated 8 months ago
- [ECCV 2024] ManiGaussian: Dynamic Gaussian Splatting for Multi-task Robotic Manipulation☆260Updated 9 months ago
- VLM-3R: Vision-Language Models Augmented with Instruction-Aligned 3D Reconstruction☆319Updated 4 months ago
- [ICLR 2025] LAPA: Latent Action Pretraining from Videos☆434Updated 11 months ago
- [NeurIPS 2025 Spotlight] SoFar: Language-Grounded Orientation Bridges Spatial Reasoning and Object Manipulation☆218Updated 6 months ago
- InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for Generalist Robot Policy☆330Updated this week
- [TMLR 2024] repository for VLN with foundation models☆231Updated 2 months ago
- A curated list of large VLM-based VLA models for robotic manipulation.☆293Updated 2 weeks ago
- VITRA: Scalable Vision-Language-Action Model Pretraining for Robotic Manipulation with Real-Life Human Activity Videos☆239Updated 2 weeks ago
- [ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models☆233Updated last year
- 🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.☆617Updated 6 months ago
- ☆130Updated 3 months ago
- [ICCV 2025] GLEAM: Learning Generalizable Exploration Policy for Active Mapping in Complex 3D Indoor Scene☆156Updated this week
- Cosmos-Transfer2.5, built on top of Cosmos-Predict2.5, produces high-quality world simulations conditioned on multiple spatial control in…☆304Updated last week
- [CVPR 2025] The code for paper ''Video-3D LLM: Learning Position-Aware Video Representation for 3D Scene Understanding''.☆189Updated 7 months ago
- [CVPR 2025] UniGoal: Towards Universal Zero-shot Goal-oriented Navigation☆289Updated 3 months ago