Value Alignment or Misalignment -- What Will Keep Systems Accountable?

Arnold, Thomas (Tufts University) | Kasenberg, Daniel (Tufts University) | Scheutz, Matthias (Tufts University)

AAAI Conferences 

Machine learning's advances have led to new ideas about the feasibility and importance of machine ethics keeping pace, with increasing emphasis on safety, containment, and alignment. This paper addresses a recent suggestion that inverse reinforcement learning (IRL) could be a means to so-called "value alignment.'' We critically consider how such an approach can engage the social, norm-infused nature of ethical action and outline several features of ethical appraisal that go beyond simple models of behavior, including unavoidably temporal dimensions of norms and counterfactuals.  We propose that a hybrid approach for computational architectures still offers the most promising avenue for machines acting in an ethical fashion.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found