fkenghagho / RobotVQALinks
RobotVQA is a project that develops a Deep Learning-based Cognitive Vision System to support household robots' perception while they perfom human-scale daily manipulation tasks like cooking in a normal kitchen. The system relies on dense description of objects in the scene and their relationships
☆18Updated last year
Alternatives and similar repositories for RobotVQA
Users that are interested in RobotVQA are comparing it to the libraries listed below
Sorting:
- NeurIPS 2022 Paper "VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation"☆96Updated 5 months ago
- Codebase for HiP☆89Updated last year
- ☆45Updated last year
- Implementation of Language-Conditioned Path Planning (Amber Xie, Youngwoon Lee, Pieter Abbeel, Stephen James)☆23Updated 2 years ago
- Voltron Evaluation: Diverse Evaluation Tasks for Robotic Representation Learning☆36Updated 2 years ago
- Chain-of-Thought Predictive Control☆58Updated 2 years ago
- Instruction Following Agents with Multimodal Transforemrs☆53Updated 2 years ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆44Updated last year
- code for TIDEE: Novel Room Reorganization using Visuo-Semantic Common Sense Priors☆41Updated last year
- Code to evaluate a solution in the BEHAVIOR benchmark: starter code, baselines, submodules to iGibson and BDDL repos☆67Updated last year
- Code for the paper Watch-And-Help: A Challenge for Social Perception and Human-AI Collaboration☆99Updated 3 years ago
- MiniGrid Implementation of BEHAVIOR Tasks☆54Updated last month
- [ICRA2023] Grounding Language with Visual Affordances over Unstructured Data☆46Updated last year
- Codebase for ICLR 2023 paper, "SMART: Self-supervised Multi-task pretrAining with contRol Transformers"☆54Updated last year
- ☆34Updated 2 years ago
- This code corresponds to transformer training and evaluation code used as part of the OPTIMUS project.☆81Updated last year
- PyTorch implementation of the Hiveformer research paper☆49Updated 2 years ago
- Official implementation of "Cross-Domain Transfer via Semantic Skill Imitation", Pertsch et al., CoRL 2022☆14Updated 2 years ago
- Pytorch code for ICRA'21 paper: "Hierarchical Cross-Modal Agent for Robotics Vision-and-Language Navigation"☆84Updated last year
- [EMNLP 2023 (Findings)] This repository contains data processing, evaluation, and fine-tuning code for NEWTON: Are Large Language Models …☆39Updated 11 months ago
- 🔀 Visual Room Rearrangement☆122Updated 2 years ago
- Official repository for "LIV: Language-Image Representations and Rewards for Robotic Control" (ICML 2023)☆124Updated 2 years ago
- Pytorch code for ICRA 2022 Paper StructFormer☆47Updated 3 years ago
- Hierarchical Universal Language Conditioned Policies☆76Updated last year
- Official implementation of Matcha-agent, https://arxiv.org/abs/2303.08268☆27Updated last year
- Reshaping Robot Trajectories Using Natural Language Commands: A Study of Multi-Modal Data Alignment Using Transformers☆60Updated 2 years ago
- Official codebase for EmbCLIP☆132Updated 2 years ago
- Official code for the paper "Housekeep: Tidying Virtual Households using Commonsense Reasoning" published at ECCV, 2022☆51Updated 2 years ago
- ☆38Updated 3 years ago
- General-purpose Visual Understanding Evaluation☆20Updated last year