Conference Papers  |    |  February 12, 2016

Using Stories to Teach Human Values to Artificial Agents

Conference paper by Mark O. Riedl and Brent Harrison.
Presented at the Thirtieth AAAI Conference on Artificial Intelligence: Workshop on AI, Ethics and Society.


Value alignment is a property of an intelligent agent indicating that it can only pursue goals that are beneficial to humans. Successful value alignment should ensure that an artificial general intelligence cannot intentionally or unintentionally perform behaviors that adversely affect humans. This is problematic in practice since it is difficult to exhaustively enumerated by human programmers. In order for successful value alignment, we argue that values should be learned. In this paper, we hypothesize that an artificial intelligence that can read and understand stories can learn the values tacitly held by the culture from which the stories originate. We describe preliminary work on using stories to generate a value-aligned reward signal for reinforcement learning agents that prevents psychotic-appearing behavior.