2022
DOI: 10.1101/2022.09.28.509945
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Decoding of the speech envelope from EEG using the VLAAI deep neural network

Abstract: To investigate the processing of speech in the brain, commonly simple linear models are used to establish a relationship between brain signals and speech features. However, these linear models are ill-equipped to model a highly-dynamic, complex non-linear system like the brain, and they often require a substantial amount of subject-specific training data. This work introduces a novel speech decoder architecture: the Very Large Augmented Auditory Inference (VLAAI) network. The VLAAI network outperformed state-… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 45 publications
1
4
0
Order By: Relevance
“…Therefore, future work should use richer, more complex speech representations. In addition, finetuning the models for the within-subject test set provides extra accuracy, which is consistent with prior studies [7], [14], [46].…”
Section: A Task 1: Match-mismatchsupporting
confidence: 87%
See 1 more Smart Citation
“…Therefore, future work should use richer, more complex speech representations. In addition, finetuning the models for the within-subject test set provides extra accuracy, which is consistent with prior studies [7], [14], [46].…”
Section: A Task 1: Match-mismatchsupporting
confidence: 87%
“…We include a simple linear model as well as the Very Large Augmented Auditory Inference (VLAAI) network [7] as a baseline for task 2. The linear model is implemented using a one-dimensional convolutional layer in TensorFlow with kernel size of 32 (corresponds to a 500 ms integration window).…”
Section: ) Baseline Methodsmentioning
confidence: 99%
“…The data set [57], [58] used in the Auditory EEG Decoding Challenge comprises EEG data that was obtained from 85 young adults who are Dutch native speakers and have normal hearing capabilities. Throughout the experiment, each participant engaged in approximately 8 to 10 trials, with each trial lasting around 15 minutes.…”
Section: Data Setmentioning
confidence: 99%
“…In order to study the neural processing of speech, recent studies have presented natural running speech to participants while the electroencephalogram (EEG) was recorded. Currently, regression is used to either decode features from the speech stimulus from the EEG (also known as a backward model) [1][2][3][4][5] , to predict the EEG from the speech stimulus 1,6 (forward model), or to transform both EEG and speech stimulus to a shared space 7,8 (hybrid model). Deep neural networks have recently been proposed for auditory decoding and have obtained promising results 4,5,[9][10][11][12] .…”
Section: Background and Summarymentioning
confidence: 99%