Goto

Collaborating Authors

 Loftin, Robert


Interactive Learning of Environment Dynamics for Sequential Tasks

arXiv.org Artificial Intelligence

In order for robots and other artificial agents to efficiently learn to perform useful tasks defined by an end user, they must understand not only the goals of those tasks, but also the structure and dynamics of that user's environment. While existing work has looked at how the goals of a task can be inferred from a human teacher, the agent is often left to learn about the environment on its own. To address this limitation, we develop an algorithm, Behavior Aware Modeling (BAM), which incorporates a teacher's knowledge into a model of the transition dynamics of an agent's environment. We evaluate BAM both in simulation and with real human teachers, learning from a combination of task demonstrations and evaluative feedback, and show that it can outperform approaches which do not explicitly consider this source of dynamics knowledge.


Training an Agent to Ground Commands with Reward and Punishment

AAAI Conferences

As robots and autonomous assistants becomemore capable, there will be agreater need for humans to easilyconvey to agents the complex tasks they wantthem to carry out. Conveying tasks throughnatural language provides an intuitive interfacethat does not require any technical expertise,but implementing such an interface requires methods forthe agent to learn a grounding of natural language commands.In this work, we demonstrate how high-level task groundings can belearned from a human trainer providing online reward and punishment.Grounding language to high-level tasks for the agent to solveremoves the need for the human to specify low-level solution details intheir command.Using reward and punishment for trainingmakes the training procedure simple enough to be used by people withouttechnical expertise and also allows a human trainer to immediatelycorrect errors in interpretation that the agent has made. We present preliminary results from a single usertraining an agent in a simple simulated home environment and showthat the agent can quickly learn a grounding oflanguage such that the agent can successfully interpretnew commands and executethem in a variety of different environments.