fkenghagho / RobotVQALinks
RobotVQA is a project that develops a Deep Learning-based Cognitive Vision System to support household robots' perception while they perfom human-scale daily manipulation tasks like cooking in a normal kitchen. The system relies on dense description of objects in the scene and their relationships
☆18Updated last year
Alternatives and similar repositories for RobotVQA
Users that are interested in RobotVQA are comparing it to the libraries listed below
Sorting:
- NeurIPS 2022 Paper "VLMbench: A Compositional Benchmark for Vision-and-Language Manipulation"☆98Updated 8 months ago
- Code for "Unleashing Large-Scale Video Generative Pre-training for Visual Robot Manipulation"☆44Updated last year
- [ICRA2023] Grounding Language with Visual Affordances over Unstructured Data☆45Updated 2 years ago
- code for TIDEE: Novel Room Reorganization using Visuo-Semantic Common Sense Priors☆40Updated 2 years ago
- ☆46Updated last year
- Codebase for HiP☆90Updated 2 years ago
- Voltron Evaluation: Diverse Evaluation Tasks for Robotic Representation Learning☆37Updated 2 years ago
- Instruction Following Agents with Multimodal Transforemrs☆53Updated 3 years ago
- Implementation of Language-Conditioned Path Planning (Amber Xie, Youngwoon Lee, Pieter Abbeel, Stephen James)☆25Updated 2 years ago
- Chain-of-Thought Predictive Control☆57Updated 2 years ago
- PyTorch implementation of the Hiveformer research paper☆49Updated 2 years ago
- ☆33Updated last year
- ☆37Updated 2 years ago
- MiniGrid Implementation of BEHAVIOR Tasks☆56Updated 3 months ago
- ☆44Updated 2 years ago
- 🔀 Visual Room Rearrangement☆123Updated 2 years ago
- ☆33Updated last year
- Official codebase for EmbCLIP☆131Updated 2 years ago
- Official repository for "LIV: Language-Image Representations and Rewards for Robotic Control" (ICML 2023)☆130Updated 2 years ago
- ☆78Updated 7 months ago
- ☆89Updated last year
- Code to evaluate a solution in the BEHAVIOR benchmark: starter code, baselines, submodules to iGibson and BDDL repos☆69Updated last year
- General-purpose Visual Understanding Evaluation☆20Updated 2 years ago
- Codebase for ICLR 2023 paper, "SMART: Self-supervised Multi-task pretrAining with contRol Transformers"☆54Updated last year
- Code for the RSS 2023 paper "Energy-based Models are Zero-Shot Planners for Compositional Scene Rearrangement"☆21Updated 2 years ago
- ☆44Updated 3 years ago
- ☆75Updated last year
- ☆46Updated 2 years ago
- ☆79Updated last year
- ☆56Updated last year