LIAM: Multimodal Transformer for Language Instructions, Images, Actions and Semantic Maps
Wang, Yihao, Memmesheimer, Raphael, Behnke, Sven
–arXiv.org Artificial Intelligence
The availability of large language models and open-vocabulary object perception methods enables more flexibility for domestic service robots. The large variability of domestic tasks can be addressed without implementing each task individually by providing the robot with a task description along with appropriate environment information. In this work, we propose LIAM -- an end-to-end model that predicts action transcripts based on language, image, action, and map inputs. Language and image inputs are encoded with a CLIP backbone, for which we designed two pre-training tasks to fine-tune its weights and pre-align the latent spaces. We evaluate our method on the ALFRED dataset, a simulator-generated benchmark for domestic tasks. Our results demonstrate the importance of pre-aligning embedding spaces from different modalities and the efficacy of incorporating semantic maps.
arXiv.org Artificial Intelligence
Mar-15-2025
- Genre:
- Research Report > New Finding (0.68)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.68)
- Natural Language > Large Language Model (0.50)
- Robots (1.00)
- Vision (1.00)
- Information Technology > Artificial Intelligence