Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-1190
|View full text |Cite
|
Sign up to set email alerts
|

Self-Attention Channel Combinator Frontend for End-to-End Multichannel Far-Field Speech Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(11 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…We showed that self attention-based algorithms reach state-ofthe-art performance without requiring handcrafted feature extraction. Furthermore, the original self-attention channel combinator (SACC) [22] slightly outperforms MVDR beamformer without requiring to estimate speech signal statistics. The SACC model was also extended in the complex space to preserve all of the information in the STFT.…”
Section: Discussionmentioning
confidence: 99%
See 4 more Smart Citations
“…We showed that self attention-based algorithms reach state-ofthe-art performance without requiring handcrafted feature extraction. Furthermore, the original self-attention channel combinator (SACC) [22] slightly outperforms MVDR beamformer without requiring to estimate speech signal statistics. The SACC model was also extended in the complex space to preserve all of the information in the STFT.…”
Section: Discussionmentioning
confidence: 99%
“…beamforming [15]). In order to automatically combine multiple channels, we apply the SACC method [22] as a feature extractor for the OSD task. This method is also extended in the complex space (cSACC) to preserve all of the STFT information during the weight-estimation procedure.…”
Section: Multichannel Feature Extractionmentioning
confidence: 99%
See 3 more Smart Citations