Pragmatic-Pedagogic Value Alignment

07/20/2017
by   Jaime F. Fisac, et al.
0

For an autonomous system to provide value (e.g., to customers, designers, or society at large) it must have a reliable method to determine the intended goal. This is the essence of the value-alignment problem: ensuring that the objectives of an autonomous system match those of its human users. In robotics, value alignment is crucial to the design of collaborative robots that can integrate into human workflows, successfully learning and adapting to the objectives of their users as they go. We argue that a meaningful solution to the value-alignment problem will combine multi-agent decision theory with rich mathematical models of human cognition, enabling robots to tap into people's natural collaborative capabilities. We present a solution to the cooperative inverse reinforcement learning (CIRL) dynamic game using well-established models of decision making and theory of mind from cognitive science. The solution accounts for two crucial aspects of collaborative value alignment: that the human will not plan her actions in isolation, but will reason pedagogically about how the robot might learn from them; and that the robot should anticipate this and interpret the human's actions pragmatically. To our knowledge, this constitutes the first equilibrium analysis of value alignment grounded in an empirically validated cognitive model of the human.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset