dongyh20 / Octopus
[ECCV2024] πOctopus, an embodied vision-language model trained with RLEF, emerging superior in embodied visual planning and programming.
β284Updated 10 months ago
Alternatives and similar repositories for Octopus:
Users that are interested in Octopus are comparing it to the libraries listed below
- [ICLR 2024] Source codes for the paper "Building Cooperative Embodied Agents Modularly with Large Language Models"β247Updated last week
- Open Platform for Embodied Agentsβ298Updated 2 months ago
- Code for "Learning to Model the World with Language." ICML 2024 Oral.β379Updated last year
- OpenEQA Embodied Question Answering in the Era of Foundation Modelsβ260Updated 5 months ago
- (ECCV 2024) Code for V-IRL: Grounding Virtual Intelligence in Real Lifeβ339Updated 3 months ago
- Embodied Agent Interface (EAI): Benchmarking LLMs for Embodied Decision Making (NeurIPS D&B 2024 Oral)β177Updated 2 weeks ago
- [NeurIPSw'24] This repo is the official implementation of "MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simuβ¦β83Updated last month
- Towards Large Multimodal Models as Visual Foundation Agentsβ192Updated last month
- [arXiv 2023] Embodied Task Planning with Large Language Modelsβ170Updated last year
- [CVPR2024] This is the official implement of MP5β96Updated 8 months ago
- β125Updated 8 months ago
- Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Modelβ355Updated 8 months ago
- Code for MultiPLY: A Multisensory Object-Centric Embodied Large Language Model in 3D Worldβ127Updated 4 months ago
- Compose multimodal datasets πΉβ309Updated this week
- [ICCV'23] LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Modelsβ167Updated last week
- [ICML 2024] Official code repository for 3D embodied generalist agent LEOβ417Updated 2 months ago
- β44Updated last year
- Official Repo for Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learningβ322Updated 3 months ago
- Implementation of "Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agenβ¦β270Updated last year
- [ICLR 2024 Spotlight] Code for the paper "Text2Reward: Reward Shaping with Language Models for Reinforcement Learning"β153Updated 3 months ago
- [COLM-2024] List Items One by One: A New Data Source and Learning Paradigm for Multimodal LLMsβ139Updated 6 months ago
- SmartPlay is a benchmark for Large Language Models (LLMs). Uses a variety of games to test various important LLM capabilities as agents. β¦β134Updated 11 months ago
- β71Updated last year
- [ACL 2024] PCA-Bench: Evaluating Multimodal Large Language Models in Perception-Cognition-Action Chainβ102Updated last year
- [ICLR 2024 Spotlight] DreamLLM: Synergistic Multimodal Comprehension and Creationβ426Updated 3 months ago
- Generating Robotic Simulation Tasks via Large Language Modelsβ316Updated 11 months ago
- Implementation of "PaLM-E: An Embodied Multimodal Language Model"β286Updated last year
- Official implementation of WebVLN: Vision-and-Language Navigation on Websitesβ28Updated last year
- [NeurIPS 2023 Datasets and Benchmarks Track] LAMM: Multi-Modal Large Language Models and Applications as AI Agentsβ309Updated 11 months ago