Self-recognition, being indispensable for successful social communication, has become a major focus in current social neuroscience. The physical aspects of the self are most typically manifested in the face and voice. Compared with the wealth of studies on self-face recognition, self-voice recognition (SVR) has not gained much attention. Converging evidence has suggested that the fundamental frequency (F0) and formant structures serve as the key acoustic cues for other-voice recognition (OVR). However, little is known about which, and how, acoustic cues are utilized for SVR as opposed to OVR. To address this question, we independently manipulated the F0 and formant information of recorded voices and investigated their contributions to SVR and OVR. Japanese participants were presented with recorded vocal stimuli and were asked to identify the speaker—either themselves or one of their peers. Six groups of 5 peers of the same sex participated in the study. Under conditions where the formant information was fully preserved and where only the frequencies lower than the third formant (F3) were retained, accuracies of SVR deteriorated significantly with the modulation of the F0, and the results were comparable for OVR. By contrast, under a condition where only the frequencies higher than F3 were retained, the accuracy of SVR was significantly higher than that of OVR throughout the range of F0 modulations, and the F0 scarcely affected the accuracies of SVR and OVR. Our results indicate that while both F0 and formant information are involved in SVR, as well as in OVR, the advantage of SVR is manifested only when major formant information for speech intelligibility is absent. These findings imply the robustness of self-voice representation, possibly by virtue of auditory familiarity and other factors such as its association with motor/articulatory representation.
Human social activity is a continuous dynamic behavior consisting of live social signal exchanges; thus, studying interactions among multiple humans is critical to understanding social cognition. Indeed, social neuroscience focusing on such aspects-interactive social neuroscience-is an emerging field of interest. Functional near-infrared spectroscopy (fNIRS) has played a significant role in accelerating this field by enabling real-world neuroimaging for various populations. The present paper will first review previous hyperscanning studies using functional magnetic resonance imaging (fMRI), magnetoencephalography (MEG), and electroencephalography (EEG). We will then summarize attempts and findings of fNIRS hyperscanning studies on social interaction in adult populations. Finally, we will review recent investigations of interactive social neuroscience in young populations and show preliminary results from a mother-infant hyperscanning study. These studies have predominantly revealed synchronized brain activities between humans and have identified conditions in which such inter-personal connectivity was found to be increased. Furthermore, these studies suggest possible mechanisms of inter-brain coupling: a process that recruits both mirror system and mentalization networks. Although fNIRS hyperscanning of infants remains limited, the reviewed literature demonstrates significant potential for fNIRS to disclose the interactive social brain and its development.We thank E. Hoshino, M. Asano, and M. Hata for their kind help with the preliminary study. We also thank K. Higuchi for automatic behavioral estimation used in Figure 1.
The present study used functional near-infrared spectroscopy (fNIRS) to measure 5- to 6-month-old infants’ hemodynamic response in the prefrontal cortex (PFC) to visual stimuli differing in saliency and social value. Nineteen Japanese 5- to 6-month-old infants watched video clips of Peek-a-Boo (social signal) performed by an anime character (AC) or a human, and hand movements without social signal performed by an AC. The PFC activity of infants was measured by 22-channel fNIRS, while behaviors including looking time were recorded simultaneously. NIRS data showed that infants’ hemodynamic responses in the PFC generally decreased due to these stimuli, and the decrease was most prominent in the frontopolar (FP), covering medial PFC (MPFC), when infants were viewing Peek-a-Boo performed by an AC. Moreover, the decrease was more pronounced in the dorsolateral PFC (DLPFC) when infants were viewing Peek-a-Boo performed by an AC than by a human. Accordingly, behavioral data revealed significantly longer looking times when Peek-a-Boo was performed by an AC than by a human. No significant difference between Peek-a-Boo and non-Peek-a-Boo conditions was observed in either measure. These findings indicate that infants at this age may prefer stimuli with more salient features, which may be more effective in attracting their attentions. In conjunction with our previous findings on responses to self-name calling in infants of similar age, we hypothesize that the dynamic function of the MPFC and its vicinity (as part of default mode network (DMN): enhanced by self-focused stimuli, attenuated by externally focused stimuli), which is consistently observed in adults, may have already emerged in 5- to 6-month-old infants.
Individuals with autism spectrum disorders (ASD) are impaired not only in social competencies but also in sensory perception, particularly olfaction. The olfactory ability of individuals with ASD has been examined in several psychophysical studies, but the results have been highly variable, which might be primarily due to methodological difficulties in the control of odor stimuli (e.g., the problem of lingering scents). In addition, the neural correlates of olfactory specificities in individuals with ASD remain largely unknown. To date, only one study has investigated this issue using functional magnetic resonance imaging (fMRI). The present study utilized a sophisticated method−a pulse ejection system−to present well-controlled odor stimuli to participants with ASD using an ASD-friendly application. With this advantageous system, we examined their odor detection, identification, and evaluation abilities and measured their brain activity evoked by odors using functional near-infrared spectroscopy (fNIRS). As the odor detection threshold (DT) of participants with ASD was highly variable, these participants were divided into two groups according to their DT: an ASD-Low DT group and an ASD-High DT group. Behavioral results showed that the ASD-High DT group had a significantly higher DT than the typically developing (control) group and the ASD-Low DT group, indicating their insensitivity to the tested odors. In addition, while there was no significant difference in the odor identification ability between groups, there was some discrepancy between the groups’ evaluations of odor pleasantness. The brain data identified, for the first time, that neural activity in the right dorsolateral prefrontal cortex (DLPFC) was significantly weaker in the ASD-High DT group than in the control group. Moreover, the strength of activity in the right DLPFC was negatively correlated with the DT. These findings suggest that participants with ASD have impairments in the higher-order function of olfactory processing, such as olfactory working memory and/or attention.
Vocal control plays a critical role in smooth social communication. Speakers constantly monitor auditory feedback (AF) and make adjustments when their voices deviate from their intentions. Previous studies have shown that when certain acoustic features of the AF are artificially altered, speakers compensate for this alteration in the opposite direction. However, little is known about how the vocal control system implements compensations for alterations of different acoustic features, and associates them with subjective consciousness. The present study investigated whether compensations for the fundamental frequency (F0), which corresponds to perceived pitch, and formants, which contribute to perceived timbre, can be performed unconsciously and independently. Forty native Japanese speakers received two types of altered AF during vowel production that involved shifts of either only the formant frequencies (formant modification; Fm) or both the pitch and formant frequencies (pitch + formant modification; PFm). For each type, three levels of shift (slight, medium, and severe) in both directions (increase or decrease) were used. After the experiment, participants were tested for whether they had perceived a change in the F0 and/or formants. The results showed that (i) only formants were compensated for in the Fm condition, while both the F0 and formants were compensated for in the PFm condition; (ii) the F0 compensation exhibited greater precision than the formant compensation in PFm; and (iii) compensation occurred even when participants misperceived or could not explicitly perceive the alteration in AF. These findings indicate that non-experts can compensate for both formant and F0 modifications in the AF during vocal production, even when the modifications are not explicitly or correctly perceived, which provides further evidence for a dissociation between conscious perception and action in vocal control. We propose that such unconscious control of voice production may enhance rapid adaptation to changing speech environments and facilitate mutual communication.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.