Social engagement is a key indicator of an individual's socio-emotional and cognitive states. For a child with Autism Spectrum Disorder (ASD), this serves as an important factor in assessing the quality of the interactions and interventions. So far, qualitative measures of social engagement have been used extensively in research and in practice, but a reliable, objective, and quantitative measure is yet to be widely accepted and utilized. In this paper, we present our work on the development of a framework for the automated measurement of social engagement in children with ASD that can be utilized in real-world settings for the long-term clinical monitoring of a child's social behaviors as well as for the evaluation of the intervention methods being used. We present a computational modeling approach to derive the social engagement metric based on a user study with children between the ages of 4 and 12 years. The study was conducted within a child-robot interaction setting that targets sensory processing skills in children. We collected video, audio and motion-tracking data from the subjects and used them to generate personalized models of social engagement by training a multi-channel and multi-layer convolutional neural network. We then evaluated the performance of this network by comparing it with traditional classifiers and assessed its limitations, followed by discussions on the next steps toward finding a comprehensive and accurate metric for social engagement in ASD.
Robotic eye-gaze-based cueing has been studied and proved to be effective, in controlled environments, in achieving social functions as humans gaze. However, its dynamic adaptability in various real interactions has not been explored in-depth. This paper addresses a case where a simplistic robotic gaze fails to achieve effective social cueing in human–robot communication, primarily due to in-attentional blindness (IB), and presents a method that enables the robot to deliver gaze-based social signals adequately. To understand the implications of IB and figure out ways to overcome any limitations from IB, which frequently arise in task-oriented situations, we designed a set of 1-on-1 communication experiments consisting of a robotic tutor and human learners participating in multiple-choice quiz sessions (task-oriented situation). Here, multimedia contents were utilized alongside the robot as visual stimuli competing for the human’s attention. We hypothesized that quiz scores would increase when humans successfully recognize the robot’s gaze-based cue signals hinting answers. We compared the impacts of two different cueing methods: oblique cueing (OC), where cues were straightforwardly provided regardless of whether participants were potentially experiencing IB or not, and leading cueing (LC), where cueing procedures were led through achieving eye contact and securing the participants’ attention before signaling the cue. By comparing participants’ test scores achieved by the control group with no cueing (NC) and two experimental groups of OC and LC, respectively, we found that there was a significant increase in test scores only when the LC method was utilized. This experiment illustrates the importance of proactively guiding a user’s attention through sophisticated interaction design in effectively attaining a user’s attention and successfully delivering cues. In future studies, we aim to evaluate different methods by which a social robot can intentionally shift a human’s attention, such as incorporating stimuli from various multi-modal human communication channels.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.