Zhitong Wang
2024
LEGENT: Open Platform for Embodied Agents
Zhili Cheng
|
Zhitong Wang
|
Jinyi Hu
|
Shengding Hu
|
An Liu
|
Yuge Tu
|
Pengkai Li
|
Lei Shi
|
Zhiyuan Liu
|
Maosong Sun
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)
Despite advancements in Large Language Models (LLMs) and Large Multimodal Models (LMMs), their integration into language-grounded, human-like embodied agents remains incomplete, hindering complex real-life task performance in 3D environments. Existing integrations often feature limited open-sourcing, challenging collective progress in this field. We introduce LEGENT, an open, scalable platform for developing embodied agents using LLMs and LMMs. LEGENT offers a dual approach: a rich 3D environment with interactive, communicable, and actionable agents, paired with a user-friendly interface, and a sophisticated data generation pipeline utilizing advanced algorithms to exploit supervision from simulated worlds at scale. In our experiments, an embryonic vision-language-action model trained on LEGENT-generated data surpasses GPT-4V in embodied tasks, showcasing promising generalization capabilities. The demo video is available at the following link https://video.legent.ai.
Search
Co-authors
- Zhili Cheng 1
- Jinyi Hu 1
- Shengding Hu 1
- An Liu 1
- Yuge Tu 1
- show all...
Venues
- acl1