facebookresearch / open-eqa
OpenEQA Embodied Question Answering in the Era of Foundation Models
β276Updated 7 months ago
Alternatives and similar repositories for open-eqa:
Users that are interested in open-eqa are comparing it to the libraries listed below
- Compose multimodal datasets πΉβ366Updated 2 weeks ago
- Official repo and evaluation implementation of VSI-Benchβ475Updated 2 months ago
- π₯[ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policyβ209Updated last month
- [CVPR 2024] The code for paper 'Towards Learning a Generalist Model for Embodied Navigation'β183Updated 10 months ago
- [ICCV'23] LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Modelsβ184Updated last month
- [ICLR 2025] LAPA: Latent Action Pretraining from Videosβ244Updated 3 months ago
- [ECCV2024] πOctopus, an embodied vision-language model trained with RLEF, emerging superior in embodied visual planning and programming.β286Updated 11 months ago
- Embodied Chain of Thought: A robotic policy that reason to solve the task.β236Updated last month
- [ICML 2024] Official code repository for 3D embodied generalist agent LEOβ436Updated 2 weeks ago
- Cosmos-Reason1 models understand the physical common sense and generate appropriate embodied decisions in natural language through long cβ¦β305Updated last month
- Implementation of "PaLM-E: An Embodied Multimodal Language Model"β301Updated last year
- Embodied Agent Interface (EAI): Benchmarking LLMs for Embodied Decision Making (NeurIPS D&B 2024 Oral)β194Updated 2 months ago
- [arXiv 2023] Embodied Task Planning with Large Language Modelsβ185Updated last year
- Embodied Reasoning Question Answer (ERQA) Benchmarkβ146Updated last month
- Code for MultiPLY: A Multisensory Object-Centric Embodied Large Language Model in 3D Worldβ128Updated 6 months ago
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasksβ91Updated 3 weeks ago
- MetaSpatial leverages reinforcement learning to enhance 3D spatial reasoning in vision-language models (VLMs), enabling more structured, β¦β112Updated last week
- Official Implementation of ReALFRED (ECCV'24)β39Updated 6 months ago
- [NeurIPS'24] This repository is the implementation of "SpatialRGPT: Grounded Spatial Reasoning in Vision Language Models"β190Updated 4 months ago
- β128Updated 9 months ago
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.β250Updated 3 months ago
- β69Updated 5 months ago
- [ICLR 2023] SQA3D for embodied scene understanding and reasoningβ131Updated last year
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.β215Updated last week
- The ProcTHOR-10K Houses Datasetβ102Updated 2 years ago
- LoTa-Bench: Benchmarking Language-oriented Task Planners for Embodied Agents (ICLR 2024)β70Updated last month
- Evaluate Multimodal LLMs as Embodied Agentsβ45Updated 2 months ago
- A flexible and efficient codebase for training visually-conditioned language models (VLMs)β672Updated 10 months ago
- [COLM-2024] List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMsβ142Updated 8 months ago
- Official repository of Learning to Act from Actionless Videos through Dense Correspondences.β214Updated last year