Goto

Collaborating Authors

 alter3


Minimal Self in Humanoid Robot "Alter3" Driven by Large Language Model

Yoshida, Takahide, Baba, Suzune, Masumori, Atsushi, Ikegami, Takashi

arXiv.org Artificial Intelligence

This paper introduces Alter3, a humanoid robot that demonstrates spontaneous motion generation through the integration of GPT-4, Large Language Model (LLM). This overcomes challenges in applying language models to direct robot control. By translating linguistic descriptions into actions, Alter3 can autonomously perform various tasks. The key aspect of humanoid robots is their ability to mimic human movement and emotions, allowing them to leverage human knowledge from language models. This raises the question of whether Alter3+GPT-4 can develop a "minimal self" with a sense of agency and ownership. This paper introduces mirror self-recognition and rubber hand illusion tests to assess Alter3's potential for a sense of self. The research suggests that even disembodied language models can develop agency when coupled with a physical robotic platform.


Meet the humanoid robot that learns from natural language, mimics human emotions

FOX News

Alter3 is a humanoid robot first introduced in 2016. Imagine what it would be like to have a robot friend that can do things like take selfies, toss a ball, eat popcorn and play air guitar? Well, you might not have to wait too long. Researchers at the University of Tokyo have created a robot that can do all that and more, thanks to the power of GPT-4, the latest and most advanced large language model (LLM) in the world. CLICK TO GET KURT'S FREE CYBERGUY NEWSLETTER WITH SECURITY ALERTS, QUICK VIDEO TIPS, TECH REVIEWS, AND EASY HOW-TO'S TO MAKE YOU SMARTER A researcher gives Alter3, a humanoid robot, verbal instructions.


From Text to Motion: Grounding GPT-4 in a Humanoid Robot "Alter3"

Yoshida, Takahide, Masumori, Atsushi, Ikegami, Takashi

arXiv.org Artificial Intelligence

We report the development of Alter3, a humanoid robot capable of generating spontaneous motion using a Large Language Model (LLM), specifically GPT-4. This achievement was realized by integrating GPT-4 into our proprietary android, Alter3, thereby effectively grounding the LLM with Alter's bodily movement. Typically, low-level robot control is hardware-dependent and falls outside the scope of LLM corpora, presenting challenges for direct LLM-based robot control. However, in the case of humanoid robots like Alter3, direct control is feasible by mapping the linguistic expressions of human actions onto the robot's body through program code. Remarkably, this approach enables Alter3 to adopt various poses, such as a 'selfie' stance or 'pretending to be a ghost,' and generate sequences of actions over time without explicit programming for each body part. This demonstrates the robot's zero-shot learning capabilities. Additionally, verbal feedback can adjust poses, obviating the need for fine-tuning. A video of Alter3's generated motions is available at https://tnoinkwms.github.io/ALTER-LLM/