While rhythmic expectancies are thought to be at the base of beat perception in music, the extent to which stress patterns in speech are similarly represented and predicted during on-line language comprehension is debated. The temporal prediction of stress may be advantageous to speech processing, as stress patterns aid segmentation and mark new information in utterances. However, while linguistic stress patterns may be organized into hierarchical metrical structures similarly to musical meter, they do not typically present the same degree of periodicity. We review the theoretical background for the idea that stress patterns are predicted and address the following questions: First, what is the evidence that listeners can predict the temporal location of stress based on preceding rhythm? If they can, is it thanks to neural entrainment mechanisms similar to those utilized for musical beat perception? And lastly, what linguistic factors other than rhythm may account for the prediction of stress in natural speech? We conclude that while expectancies based on the periodic presentation of stresses are at play in some of the current literature, other processes are likely to affect the prediction of stress in more naturalistic, less isochronous speech. Specifically, aspects of prosody other than amplitude changes (e.g., intonation) as well as lexical, syntactic and information structural constraints on the realization of stress may all contribute to the probabilistic expectation of stress in speech.
While listeners can infer the mood expressed by the music of a different culture, the question of whether strong felt emotional responses can also be induced cross-culturally remains unanswered. We address this question by measuring chill responses, sudden increases in emotional arousal, through self-report and skin conductance measures. Excerpts of Western classical, traditional Chinese, and Hindustani classical music were presented to 3 groups of participants, each familiar with one of these styles. Participants felt a similar number of chills to both familiar and unfamiliar musical styles, but significantly fewer chills to scrambled music, which acted as a control. Acoustic analyses revealed that sudden peaks in loudness, brightness, and roughness were correlated with chills across styles, suggesting that similar acoustic elements induced emotional responses cross-culturally. The number of chills was also related to the degree to which participants paid attention to the music, rated the music as emotional and as having affected their own mood, but not to how much participants liked the music. Overall, this research counters the idea of musical emotional meaning as being entirely generated within cultural conventions and supports the claim that some aspects of the way music conveys emotion may be shared across cultures.
In recent years, a growing number of studies have used cortical tracking methods to investigate auditory language processing. Although most studies that employ cortical tracking stem from the field of auditory signal processing, this approach should also be of interest to psycholinguistics—particularly the subfield of sentence processing—given its potential to provide insight into dynamic language comprehension processes. However, there has been limited collaboration between these fields, which we suggest is partly because of differences in theoretical background and methodological constraints, some mutually exclusive. In this article, we first review the theories and methodological constraints that have historically been prioritized in each field and provide concrete examples of how some of these constraints may be reconciled. We then elaborate on how further collaboration between the two fields could be mutually beneficial. Specifically, we argue that the use of cortical tracking methods may help resolve long-standing debates in the field of sentence processing that commonly used behavioral and neural measures (e.g., ERPs) have failed to adjudicate. Similarly, signal processing researchers who use cortical tracking may be able to reduce noise in the neural data and broaden the impact of their results by controlling for linguistic features of their stimuli and by using simple comprehension tasks. Overall, we argue that a balance between the methodological constraints of the two fields will lead to an overall improved understanding of language processing as well as greater clarity on what mechanisms cortical tracking of speech reflects. Increased collaboration will help resolve debates in both fields and will lead to new and exciting avenues for research.
Speech disfluencies such as repeated words and pauses provide information about the cognitive systems underlying speech production. Understanding whether older age leads to changes in speech fluency can therefore help characterize the robustness of these systems over the life span. Older adults have been assumed to be more disfluent, but current evidence is minimal and contradictory. Particularly noteworthy is the lack of longitudinal data that would help establish whether a given individual's disfluency rates change over time. This study examines changes in disfluency rates through a sequential design with a longitudinal component, involving the analysis of 325 recorded interviews conducted with 91 individuals at several points in their lives, spanning the ages of 20-94 years. We analyzed the speech of these individuals to assess the extent to which they became more disfluent in later interviews. We found that, with older age, individuals spoke more slowly and repeated more words. However, older age was not associated with other types of disfluencies such as filled pauses (uh's and um's) and repairs. Overall, this study provides evidence that, although age itself is not a strong predictor of disfluencies, age leads to changes in other speech characteristics among some individuals (i.e., speech rate and indicators of lexical and syntactic complexity), and those changes in turn predict the production of disfluencies over the life span. These findings help resolve previous inconsistencies in this literature and set the stage for future experimental work on the cognitive mechanisms underlying changes in speech production in healthy aging. Public Significance StatementThis study measures changes in speech fluency over the life span sequentially using a novel corpus of 91 individuals interviewed at multiple times in their lives. We found that, while older adults spoke more slowly and repeated more words, they were not more likely to produce filled pauses (uh's and um's) and repairs. Overall, we suggest that age is not as strong a predictor of disfluencies as are other individual differences in speech characteristics.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.