Including robots in children's lives calls for reflection on the psychological and moral aspects of such relationships, especially with respect to children's ability to differentiate intentional from unintentional false statements, that is, lies from mistakes. This ability calls for an understanding of an interlocutor's intentions. This study examined the ability of 5‐6‐year‐olds to recognize, and morally evaluate, lies and mistakes produced by a human as compared to a NAO robot, and to attribute relevant emotions to the deceived party. Irrespective of the agent, children had more difficulty in understanding mistakes than lies. In addition, they were disinclined to attribute a lie to the robot. Children's age and their understanding of intentionality were the strongest predictors of their performance on the lie‐mistake task. Children's Theory of Mind, but not their executive function skills, also correlated with their performance. Our findings suggest that, regardless of age, a robot is perceived as an intentional agent. Robot behaviour was more acceptable for children because his actions could be attributed to someone who programmed it to act in a specific way.
Highlights
The ability to recognize an intention to lie or not in different agents represents a significant developmental step.
Children saw a human/robot making intentionally or unintentionally false statements, and understanding the mistake was more difficult than the lie.
Robots may be associated with the human concept by younger children with important implications for use of cHRI in education.