raphael-sch / VELMALinks
VELMA agent for VLN in Street View
☆28Updated 2 years ago
Alternatives and similar repositories for VELMA
Users that are interested in VELMA are comparing it to the libraries listed below
Sorting:
- Code of the paper "NavCoT: Boosting LLM-Based Vision-and-Language Navigation via Learning Disentangled Reasoning" (TPAMI 2025)☆107Updated 5 months ago
- Official GitHub Repository for Paper "Bridging Zero-shot Object Navigation and Foundation Models through Pixel-Guided Navigation Skill", …☆122Updated last year
- [AAAI 2024] Official implementation of NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models☆285Updated 2 years ago
- Official implementation of Think Global, Act Local: Dual-scale GraphTransformer for Vision-and-Language Navigation (CVPR'22 Oral).☆223Updated 2 years ago
- [CVPR 2024] The code for paper 'Towards Learning a Generalist Model for Embodied Navigation'☆209Updated last year
- ☆111Updated last year
- [AAAI-25 Oral] Official Implementation of "FLAME: Learning to Navigate with Multimodal LLM in Urban Environments"☆64Updated last week
- [ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models☆216Updated last year
- ☆118Updated last year
- [ACL 24] The official implementation of MapGPT: Map-Guided Prompting with Adaptive Path Planning for Vision-and-Language Navigation.☆114Updated 6 months ago
- [TMLR 2024] repository for VLN with foundation models☆206Updated 2 weeks ago
- ☆37Updated last year
- [CVPR 2024] The code for paper 'Towards Learning a Generalist Model for Embodied Navigation'☆54Updated last year
- [ICCV'23] Learning Vision-and-Language Navigation from YouTube Videos☆62Updated 10 months ago
- Code for LGX (Language Guided Exploration). We use LLMs to perform embodied robot navigation in a zero-shot manner.☆66Updated 2 years ago
- Open Vocabulary Object Navigation☆96Updated 5 months ago
- ☆28Updated 5 months ago
- Embodied Question Answering (EQA) benchmark and method (ICCV 2025)☆40Updated 2 months ago
- Repository for Vision-and-Language Navigation via Causal Learning (Accepted by CVPR 2024)☆89Updated 5 months ago
- [ICCV 2023] PEANUT: Predicting and Navigating to Unseen Targets☆52Updated last year
- ☆176Updated 7 months ago
- Code of the paper "Correctable Landmark Discovery via Large Models for Vision-Language Navigation" (TPAMI 2024)☆16Updated last year
- ZSON: Zero-Shot Object-Goal Navigation using Multimodal Goal Embeddings. NeurIPS 2022☆93Updated 2 years ago
- Leveraging Large Language Models for Visual Target Navigation☆138Updated 2 years ago
- official implementation of NeurIPS 2023 paper "FGPrompt: Fine-grained Goal Prompting for Image-goal Navigation"☆37Updated last year
- [CVPR 2023] CoWs on Pasture: Baselines and Benchmarks for Language-Driven Zero-Shot Object Navigation☆144Updated 2 years ago
- Official implementation of GridMM: Grid Memory Map for Vision-and-Language Navigation (ICCV'23).☆98Updated last year
- Official implementation of Why Only Text: Empowering Vision-and-Language Navigation with Multi-modal Prompts(IJCAI 2024)☆15Updated last year
- [NeurIPS 2024] SG-Nav: Online 3D Scene Graph Prompting for LLM-based Zero-shot Object Navigation☆277Updated last month
- Fast-Slow Test-time Adaptation for Online Vision-and-Language Navigation☆28Updated last year