LILA: Language-Informed Latent Actions
Karamcheti, Siddharth, Srivastava, Megha, Liang, Percy, Sadigh, Dorsa
–arXiv.org Artificial Intelligence
We introduce Language-Informed Latent Actions (LILA), a framework for learning natural language interfaces in the context of human-robot collaboration. LILA falls under the shared autonomy paradigm: in addition to providing discrete language inputs, humans are given a low-dimensional controller $-$ e.g., a 2 degree-of-freedom (DoF) joystick that can move left/right and up/down $-$ for operating the robot. LILA learns to use language to modulate this controller, providing users with a language-informed control space: given an instruction like "place the cereal bowl on the tray," LILA may learn a 2-DoF space where one dimension controls the distance from the robot's end-effector to the bowl, and the other dimension controls the robot's end-effector pose relative to the grasp point on the bowl. We evaluate LILA with real-world user studies, where users can provide a language instruction while operating a 7-DoF Franka Emika Panda Arm to complete a series of complex manipulation tasks. We show that LILA models are not only more sample efficient and performant than imitation learning and end-effector control baselines, but that they are also qualitatively preferred by users.
arXiv.org Artificial Intelligence
Nov-4-2021
- Country:
- Europe > United Kingdom
- England > Greater London > London (0.04)
- North America > United States
- California > Santa Clara County > Palo Alto (0.04)
- Europe > United Kingdom
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Education (0.46)
- Leisure & Entertainment (0.46)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks (0.93)
- Natural Language (1.00)
- Robots (1.00)
- Information Technology > Artificial Intelligence