silence143 / EMMOE
EMMOE: A Comprehensive Benchmark for Embodied Mobile Manipulation in Open Environments
☆14Updated 2 weeks ago
Alternatives and similar repositories for EMMOE:
Users that are interested in EMMOE are comparing it to the libraries listed below
- [NeurIPS 2024 D&B] Point Cloud Matters: Rethinking the Impact of Different Observation Spaces on Robot Learning☆76Updated 6 months ago
- ☆49Updated 6 months ago
- List of papers on video-centric robot learning☆19Updated 5 months ago
- [RSS 2025] Novel Demonstration Generation with Gaussian Splatting Enables Robust One-Shot Manipulation☆61Updated last week
- [CVPR 25] G3Flow: Generative 3D Semantic Flow for Pose-aware and Generalizable Object Manipulation☆58Updated 3 weeks ago
- [NeurIPS 2024] Official code repository for MSR3D paper☆50Updated last week
- Code for paper "Grounding Video Models to Actions through Goal Conditioned Exploration" (ICLR 2025 Spotlight).☆44Updated 4 months ago
- ☆69Updated 7 months ago
- Official implementation of SAME: Learning Generic Language-Guided Visual Navigation with State-Adaptive Mixture of Experts☆18Updated 4 months ago
- HandsOnVLM: Vision-Language Models for Hand-Object Interaction Prediction☆28Updated 4 months ago
- FLaRe: Achieving Masterful and Adaptive Robot Policies with Large-Scale Reinforcement Learning Fine-Tuning☆18Updated 3 months ago
- [RAL 2024] OpenObj: Open-Vocabulary Object-Level Neural Radiance Fields with Fine-Grained Understanding☆25Updated 2 months ago
- Code for FLIP: Flow-Centric Generative Planning for General-Purpose Manipulation Tasks☆60Updated 4 months ago
- Implementation of our ICCV 2023 paper DREAMWALKER: Mental Planning for Continuous Vision-Language Navigation☆19Updated last year
- ☆27Updated 3 weeks ago
- [CVPR 2024] Situational Awareness Matters in 3D Vision Language Reasoning☆37Updated 4 months ago
- ☆58Updated 3 months ago
- Official Implementation of Learning Navigational Visual Representations with Semantic Map Supervision (ICCV2023)☆25Updated last year
- Single-file implementation to advance vision-language-action (VLA) models with reinforcement learning.☆56Updated this week
- Splat-MOVER: Multi-Stage, Open-Vocabulary Robotic Manipulation via Editable Gaussian Splatting☆32Updated 6 months ago
- Planning as In-Painting: A Diffusion-Based Embodied Task Planning Framework for Environments under Uncertainty☆20Updated last year
- Official implementation of Sim-to-Real Transfer via 3D Feature Fields for Vision-and-Language Navigation (CoRL'24).☆62Updated last month
- Dreamitate: Real-World Visuomotor Policy Learning via Video Generation (CoRL 2024)☆44Updated 10 months ago
- Official code for the CVPR 2025 paper "Navigation World Models".☆60Updated 2 weeks ago
- Unifying 2D and 3D Vision-Language Understanding☆74Updated 2 weeks ago
- EgoVid-5M: A Large-Scale Video-Action Dataset for Egocentric Video Generation☆103Updated 5 months ago
- Official implementation of "SUGAR: Pre-training 3D Visual Representations for Robotics" (CVPR'24).☆38Updated 9 months ago
- ☆26Updated 3 months ago
- This is the official implementation of Video Generation part of This&That: Language-Gesture Controlled Video Generation for Robot Plannin…☆39Updated 2 months ago
- Official implementation of "Re3Sim: Generating High-Fidelity Simulation Data via 3D-Photorealistic Real-to-Sim for Robotic Manipulation"☆87Updated last month