Determining the identity of children is critical to aid in the fight against child exploitation, as well as for passport control and visa issuance purposes. Facial image comparison is one method that may be used to determine identity. Due to the substantial amount of facial growth that occurs in childhood, it is critical to understand facial image comparison performance across both chronological age (the age of the child), and age variation (the age difference between images). In this study we examined the performance of 120 facial comparison practitioners from a government agency on a dataset of 23,760 image pairs selected from the agency’s own database of controlled, operational images. Each chronological age in childhood (0–17 years) and age variations ranging from 0–10 years were examined. Practitioner performance was found to vary considerably across childhood, and depended on whether the pairs were mated (same child) or non-mated (different child). Overall, practitioners were more accurate and confident with image pairs containing older children, and also more accurate and confident with smaller age variations. Chronological age impacted on accuracy with mated pairs, but age variation did not. In contrast, both age and age variation impacted on accuracy with non-mated pairs. These differences in performance show that changes in the face throughout childhood have a significant impact on practitioner performance. We propose that improvements in accuracy may be achievable with a better understanding of which facial features are most appropriate to compare across childhood, and adjusting training and development programs accordingly.
These experiments investigated how the candidate list displayed to facial reviewers affects their performance in a one-to-many unfamiliar face matching task. Automated facial recognition systems present the results of a database search and require selection of an image that matches the target. Few studies investigate how humans in combination with facial recognition algorithms perform within different operational contexts. These experiments investigated how the candidate list displayed to facial reviewers affects their performance in a one-to-many unfamiliar face matching task. We tested candidate list length with inexperienced (Experiment 1) and experienced (Experiment 2) facial reviewers. Candidate list length had a large impact on performance, varying with the operational context. However, response-time analyses show that the accurate responses were resolved quickly, with an error-prone guess process implemented after failed search. Long candidate lists (100 images) produced more false alarms, fewer hits, lower decision confidence, and increased response latencies among both inexperienced and experienced facial reviewers.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.