2022
DOI: 10.3389/fncom.2022.1057439
|View full text |Cite
|
Sign up to set email alerts
|

On the similarities of representations in artificial and brain neural networks for speech recognition

Abstract: IntroductionIn recent years, machines powered by deep learning have achieved near-human levels of performance in speech recognition. The fields of artificial intelligence and cognitive neuroscience have finally reached a similar level of performance, despite their huge differences in implementation, and so deep learning models can—in principle—serve as candidates for mechanistic models of the human auditory system.MethodsUtilizing high-performance automatic speech recognition systems, and advanced non-invasive… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 80 publications
0
2
0
Order By: Relevance
“…There thus appears to be a potential for elegant and constructive theoretical reductionism over multiple psychological tasks and phenomena. In line with this, artificial neural networks even appear to capture some aspects of how the brain processes information at the neural level (Hao et al, 2020;Jocham et al, 2011;Suri & Schultz, 1999;Wingfield et al, 2022).…”
Section: Discussionmentioning
confidence: 85%
“…There thus appears to be a potential for elegant and constructive theoretical reductionism over multiple psychological tasks and phenomena. In line with this, artificial neural networks even appear to capture some aspects of how the brain processes information at the neural level (Hao et al, 2020;Jocham et al, 2011;Suri & Schultz, 1999;Wingfield et al, 2022).…”
Section: Discussionmentioning
confidence: 85%
“…We see this, for example, in the Kocagoncu et al (2017) RSA analyses for the same dataset, which showed well circumscribed and differentiated patches of model fit for the different lexical and semantic models tested, while Lyu et al (2019) use MEG/ssRSA to reveal in unprecedented neurocognitive detail the neural dynamics of lexical semantic interpretation in sentential contexts. In earlier papers focusing on acoustic and phonological processes in medial and lateral STG, we use MEG/ssRSA techniques to pick out spatiotemporally well-defined patches of model fit corresponding to different subsets of auditory input analyses (Su et al, 2014;Wingfield et al, 2017Wingfield et al, , 2022. For the analyses reported here, there are several examples of spatially differentiated model fit -for example left and right angular gyrus and right middle temporal pole all show strong fit to the phonology model but no fit to the entropy or semantics models (Figure 3 and Table 1).…”
Section: Discussionmentioning
confidence: 99%