Reducing Human-Robot Goal State Divergence with Environment Design
Sikes, Kelsey, Keren, Sarah, Sreedharan, Sarath
–arXiv.org Artificial Intelligence
One of the most difficult challenges in creating successful human-AI collaborations is aligning a robot's behavior with a human user's expectations. When this fails to occur, a robot may misinterpret their specified goals, prompting it to perform actions with unanticipated, potentially dangerous side effects. To avoid this, we propose a new metric we call Goal State Divergence $\mathcal{(GSD)}$, which represents the difference between a robot's final goal state and the one a human user expected. In cases where $\mathcal{GSD}$ cannot be directly calculated, we show how it can be approximated using maximal and minimal bounds. We then input the $\mathcal{GSD}$ value into our novel human-robot goal alignment (HRGA) design problem, which identifies a minimal set of environment modifications that can prevent mismatches like this. To show the effectiveness of $\mathcal{GSD}$ for reducing differences between human-robot goal states, we empirically evaluate our approach on several standard benchmarks.
arXiv.org Artificial Intelligence
Apr-10-2024
- Country:
- Europe > Spain (0.14)
- North America > United States (0.14)
- Genre:
- Research Report (0.40)
- Technology:
- Information Technology > Artificial Intelligence
- Cognitive Science > Problem Solving (1.00)
- Games > Go (1.00)
- Representation & Reasoning > Planning & Scheduling (0.93)
- Robots > Humanoid Robots (0.81)
- Information Technology > Artificial Intelligence