sled-group / DOROTHIE
Official Code for DOROTHIE: Spoken Dialogue for Handling Unexpected Situations in Interactive Autonomous Driving Agents (Findings of EMNLP 2022)
☆20Updated last year
Alternatives and similar repositories for DOROTHIE:
Users that are interested in DOROTHIE are comparing it to the libraries listed below
- A Multi-Modal Large Language Model with Retrieval-augmented In-context Learning capacity designed for generalisable and explainable end-t…☆88Updated 5 months ago
- A comprehensive list of papers for the definition of World Models and using World Models for General Video Generation, Embodied AI, and A…☆84Updated 3 weeks ago
- Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning☆54Updated 2 months ago
- [Official] [IROS 2024] A goal-oriented planning to lift VLN performance for Closed-Loop Navigation: Simple, Yet Effective☆28Updated 11 months ago
- ☆51Updated last month
- Code for MultiPLY: A Multisensory Object-Centric Embodied Large Language Model in 3D World☆127Updated 5 months ago
- Benchmark and model for step-by-step reasoning in autonomous driving.☆38Updated 2 weeks ago
- ☆16Updated last month
- [RSS 2024] Learning Manipulation by Predicting Interaction☆101Updated 7 months ago
- Simulator designed to generate diverse driving scenarios.☆40Updated last month
- Code of the paper "NavCoT: Boosting LLM-Based Vision-and-Language Navigation via Learning Disentangled Reasoning" (TPAMI 2025)☆50Updated 2 weeks ago
- [NeurIPS 2024] CLOVER: Closed-Loop Visuomotor Control with Generative Expectation for Robotic Manipulation☆106Updated 3 months ago
- ☆33Updated last year
- Official implementation of: Bootstrapping Language-Guided Navigation Learning with Self-Refining Data Flywheel☆22Updated 3 months ago
- [ICLR 2023] SQA3D for embodied scene understanding and reasoning☆128Updated last year
- Reason2Drive: Towards Interpretable and Chain-based Reasoning for Autonomous Driving☆79Updated last year
- [ECCV 2024] Official implementation of C-Instructor: Controllable Navigation Instruction Generation with Chain of Thought Prompting☆23Updated 3 months ago
- ☆49Updated 5 months ago
- ☆29Updated 6 months ago
- [ICCV'23] Learning Vision-and-Language Navigation from YouTube Videos☆52Updated 3 months ago
- ☆69Updated 3 months ago
- ☆46Updated 3 months ago
- [ICML 2024] The offical Implementation of "DecisionNCE: Embodied Multimodal Representations via Implicit Preference Learning"☆78Updated 6 months ago
- [CVPR'24 Highlight] The official code and data for paper "EgoThink: Evaluating First-Person Perspective Thinking Capability of Vision-Lan…☆58Updated last week
- [ECCV 2024] The official code for "Dolphins: Multimodal Language Model for Driving“☆67Updated last month
- Planning as In-Painting: A Diffusion-Based Embodied Task Planning Framework for Environments under Uncertainty☆20Updated last year
- ☆45Updated 11 months ago
- ☆54Updated last year
- Evaluate Multimodal LLMs as Embodied Agents☆39Updated last month
- A paper list of world model☆25Updated 10 months ago