Can a robot lie? Young children's understanding of intentionality beneath false statements

Giulia Peretti*, Federico Manzi*, Cinzia Di Dio, Angelo Cangelosi, Paul L. Harris, Davide Massaro, Antonella Marchetti

*Autore corrispondente per questo lavoro

Risultato della ricerca: Contributo in rivistaArticolo in rivista

Abstract

Including robots in children's lives calls for reflection on the psychological and moral aspects of such relationships, especially with respect to children's ability to differentiate intentional from unintentional false statements, that is, lies from mistakes. This ability calls for an understanding of an interlocutor's intentions. This study examined the ability of 5-6-year-olds to recognize, and morally evaluate, lies and mistakes produced by a human as compared to a NAO robot, and to attribute relevant emotions to the deceived party. Irrespective of the agent, children had more difficulty in understanding mistakes than lies. In addition, they were disinclined to attribute a lie to the robot. Children's age and their understanding of intentionality were the strongest predictors of their performance on the lie-mistake task. Children's Theory of Mind, but not their executive function skills, also correlated with their performance. Our findings suggest that, regardless of age, a robot is perceived as an intentional agent. Robot behaviour was more acceptable for children because his actions could be attributed to someone who programmed it to act in a specific way.
Lingua originaleEnglish
pagine (da-a)1-25
Numero di pagine25
RivistaInfant and Child Development
DOI
Stato di pubblicazionePubblicato - 2023

Keywords

  • children
  • human–robot interaction
  • intentionality understanding
  • lie-mistake
  • theory of mind

Fingerprint

Entra nei temi di ricerca di 'Can a robot lie? Young children's understanding of intentionality beneath false statements'. Insieme formano una fingerprint unica.

Cita questo