Value Alignment or Misalignment -- What Will Keep Systems Accountable?
Arnold, Thomas (Tufts University) | Kasenberg, Daniel (Tufts University) | Scheutz, Matthias (Tufts University)
Machine learning's advances have led to new ideas about the feasibility and importance of machine ethics keeping pace, with increasing emphasis on safety, containment, and alignment. This paper addresses a recent suggestion that inverse reinforcement learning (IRL) could be a means to so-called "value alignment.'' We critically consider how such an approach can engage the social, norm-infused nature of ethical action and outline several features of ethical appraisal that go beyond simple models of behavior, including unavoidably temporal dimensions of norms and counterfactuals. We propose that a hybrid approach for computational architectures still offers the most promising avenue for machines acting in an ethical fashion.
Feb-4-2017
- Country:
- Europe > United Kingdom
- England (0.14)
- North America > United States
- Arizona (0.14)
- Europe > United Kingdom
- Industry:
- Leisure & Entertainment > Games (0.46)
- Technology: