Affiliation:
1. Research Unit on Theory of Mind, Department of Psychology Università Cattolica del Sacro Cuore Milan Italy
2. School of Engineering University of Manchester Manchester UK
3. Graduate School of Education Harvard University Cambridge Massachusetts USA
Abstract
AbstractIncluding robots in children's lives calls for reflection on the psychological and moral aspects of such relationships, especially with respect to children's ability to differentiate intentional from unintentional false statements, that is, lies from mistakes. This ability calls for an understanding of an interlocutor's intentions. This study examined the ability of 5‐6‐year‐olds to recognize, and morally evaluate, lies and mistakes produced by a human as compared to a NAO robot, and to attribute relevant emotions to the deceived party. Irrespective of the agent, children had more difficulty in understanding mistakes than lies. In addition, they were disinclined to attribute a lie to the robot. Children's age and their understanding of intentionality were the strongest predictors of their performance on the lie‐mistake task. Children's Theory of Mind, but not their executive function skills, also correlated with their performance. Our findings suggest that, regardless of age, a robot is perceived as an intentional agent. Robot behaviour was more acceptable for children because his actions could be attributed to someone who programmed it to act in a specific way.Highlights
The ability to recognize an intention to lie or not in different agents represents a significant developmental step.
Children saw a human/robot making intentionally or unintentionally false statements, and understanding the mistake was more difficult than the lie.
Robots may be associated with the human concept by younger children with important implications for use of cHRI in education.
Subject
Developmental and Educational Psychology
Cited by
6 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献