thunlp / LEGENTLinks
Open Platform for Embodied Agents
β334Updated 10 months ago
Alternatives and similar repositories for LEGENT
Users that are interested in LEGENT are comparing it to the libraries listed below
Sorting:
- [ECCV2024] πOctopus, an embodied vision-language model trained with RLEF, emerging superior in embodied visual planning and programming.β293Updated last year
- [CVPR 2025] RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete. Official Repository.β349Updated last month
- Official Repo for Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learningβ401Updated 11 months ago
- [ICLR 2024] Source codes for the paper "Building Cooperative Embodied Agents Modularly with Large Language Models"β286Updated 8 months ago
- [CVPR2024] This is the official implement of MP5β106Updated last year
- Embodied-Reasoner: Synergizing Visual Search, Reasoning, and Action for Embodied Interactive Tasksβ183Updated 2 months ago
- β59Updated 8 months ago
- Embodied Agent Interface (EAI): Benchmarking LLMs for Embodied Decision Making (NeurIPS D&B 2024 Oral)β272Updated 9 months ago
- [arXiv 2023] Embodied Task Planning with Large Language Modelsβ193Updated 2 years ago
- Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Modelβ370Updated last year
- Training VLM agents with multi-turn reinforcement learningβ338Updated last week
- OpenEQA Embodied Question Answering in the Era of Foundation Modelsβ332Updated last year
- [ICML 2025 Oral] Official repo of EmbodiedBench, a comprehensive benchmark designed to evaluate MLLMs as embodied agents.β227Updated last month
- [ICML 2024] Official code repository for 3D embodied generalist agent LEOβ468Updated 7 months ago
- Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).β122Updated last year
- Towards Large Multimodal Models as Visual Foundation Agentsβ245Updated 7 months ago
- Code for RoboFlamingoβ412Updated last year
- RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulationβ268Updated last week
- ALFWorld: Aligning Text and Embodied Environments for Interactive Learningβ596Updated 4 months ago
- β118Updated 8 months ago
- β94Updated last month
- The official repo for "SpatialBot: Precise Spatial Understanding with Vision Language Models.β319Updated 2 months ago
- Official repo of VLABench, a large scale benchmark designed for fairly evaluating VLA, Embodied Agent, and VLMs.β339Updated last month
- Official code for the paper: Embodied Multi-Modal Agent trained by an LLM from a Parallel TextWorldβ60Updated last year
- [ICLR 2024 Spotlight] Text2Reward: Reward Shaping with Language Models for Reinforcement Learningβ191Updated 11 months ago
- Implementation of "PaLM-E: An Embodied Multimodal Language Model"β332Updated last year
- β46Updated 2 years ago
- (ECCV 2024) Code for V-IRL: Grounding Virtual Intelligence in Real Lifeβ366Updated last year
- [ICCV'23] LLM-Planner: Few-Shot Grounded Planning for Embodied Agents with Large Language Modelsβ209Updated 8 months ago
- Embodied Chain of Thought: A robotic policy that reason to solve the task.β334Updated 8 months ago