PyTorch code for ICLR 2019 paper: Self-Monitoring Navigation Agent via Auxiliary Progress Estimation
☆122Oct 3, 2023Updated 2 years ago
Alternatives and similar repositories for selfmonitoring-agent
Users that are interested in selfmonitoring-agent are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- PyTorch code for CVPR 2019 paper: The Regretful Agent: Heuristic-Aided Navigation through Progress Estimation☆125Oct 3, 2023Updated 2 years ago
- Code for "Tactical Rewind: Self-Correction via Backtracking in Vision-and-Language Navigation"☆62Sep 24, 2019Updated 6 years ago
- Code release for Fried et al., Speaker-Follower Models for Vision-and-Language Navigation. in NeurIPS, 2018.☆137Nov 22, 2022Updated 3 years ago
- PyTorch Code of NAACL 2019 paper "Learning to Navigate Unseen Environments: Back Translation with Environmental Dropout"☆144Oct 23, 2021Updated 4 years ago
- The repository of ECCV 2020 paper `Active Visual Information Gathering for Vision-Language Navigation`☆44Apr 9, 2022Updated 3 years ago
- Bare Metal GPUs on DigitalOcean Gradient AI • AdPurpose-built for serious AI teams training foundational models, running large-scale inference, and pushing the boundaries of what's possible.
- Code for the paper "Improving Vision-and-Language Navigation with Image-Text Pairs from the Web" (ECCV 2020)☆59Oct 7, 2022Updated 3 years ago
- large scale pretrain for navigation task☆95Mar 2, 2023Updated 3 years ago
- ☆55Apr 1, 2022Updated 3 years ago
- Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation☆57Oct 26, 2021Updated 4 years ago
- Code of the NeurIPS 2021 paper: Language and Visual Entity Relationship Graph for Agent Navigation☆46Oct 31, 2021Updated 4 years ago
- PyTorch code for the ACL 2020 paper: "BabyWalk: Going Farther in Vision-and-Language Navigationby Taking Baby Steps"☆42Apr 13, 2022Updated 3 years ago
- AI Research Platform for Reinforcement Learning from Real Panoramic Images.☆684Jul 12, 2024Updated last year
- Dataset for Bilingual VLN☆11Dec 5, 2020Updated 5 years ago
- Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation☆202Aug 13, 2022Updated 3 years ago
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Code of the CVPR 2022 paper "HOP: History-and-Order Aware Pre-training for Vision-and-Language Navigation"☆30Aug 21, 2023Updated 2 years ago
- Feature resources of "Diagnosing the Environment Bias in Vision-and-Language Navigation"☆16May 6, 2020Updated 5 years ago
- Code accompanying the CVPR 2019 paper: https://arxiv.org/abs/1812.04155☆61Mar 30, 2022Updated 3 years ago
- ☆33Aug 19, 2023Updated 2 years ago
- Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language N…☆149Oct 31, 2023Updated 2 years ago
- Vision and Language Agent Navigation☆85Jan 29, 2021Updated 5 years ago
- Code and Data for our CVPR 2021 paper "Structured Scene Memory for Vision-Language Navigation"☆43Jul 31, 2021Updated 4 years ago
- code of the paper "Vision-Language Navigation with Multi-granularity Observation and Auxiliary Reasoning Tasks"☆23Mar 23, 2021Updated 5 years ago
- Know What and Know Where: An Object-and-Room Informed Sequential BERT for Indoor Vision-Language Navigation☆16Feb 7, 2022Updated 4 years ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click and start building anything your business needs.
- REVERIE: Remote Embodied Visual Referring Expression in Real Indoor Environments☆152Updated this week
- Visual Navigation with Natural Multimodal Assistance (EMNLP 2019)☆29Jun 30, 2020Updated 5 years ago
- Official implementation of Why Only Text: Empowering Vision-and-Language Navigation with Multi-modal Prompts(IJCAI 2024)☆15Oct 16, 2024Updated last year
- Vision-and-Language Navigation in Continuous Environments using Habitat☆748Jan 7, 2025Updated last year
- Codebase for the Airbert paper☆46Mar 20, 2023Updated 3 years ago
- Ideas and thoughts about the fascinating Vision-and-Language Navigation☆296Jun 28, 2023Updated 2 years ago
- Official implementation of Think Global, Act Local: Dual-scale GraphTransformer for Vision-and-Language Navigation (CVPR'22 Oral).☆260Jun 27, 2023Updated 2 years ago
- Official implementation of History Aware Multimodal Transformer for Vision-and-Language Navigation (NeurIPS'21).☆144Jun 14, 2023Updated 2 years ago
- Official implementation of KERM: Knowledge Enhanced Reasoning for Vision-and-Language Navigation (CVPR'23)☆45Aug 6, 2024Updated last year
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- Official implementation of the ECCV 2022 Oral paper: Sim-2-Sim Transfer for Vision-and-Language Navigation in Continuous Environments☆35Dec 16, 2023Updated 2 years ago
- ☆19Mar 11, 2022Updated 4 years ago
- Implementation of our ICCV 2023 paper DREAMWALKER: Mental Planning for Continuous Vision-Language Navigation☆20Jul 24, 2023Updated 2 years ago
- [ICLR 2022] code for "How Much Can CLIP Benefit Vision-and-Language Tasks?" https://arxiv.org/abs/2107.06383☆420Oct 28, 2022Updated 3 years ago
- Cooperative Vision-and-Dialog Navigation☆72Nov 22, 2022Updated 3 years ago
- Episodic Transformer (E.T.) is a novel attention-based architecture for vision-and-language navigation. E.T. is based on a multimodal tra…☆93Jul 11, 2023Updated 2 years ago
- Dynamic Robot Instruction Following☆39Dec 28, 2021Updated 4 years ago