IberSPEECH 2022 2022
DOI: 10.21437/iberspeech.2022-32
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Transformers for End-to-End Neural Speaker Diarization

Abstract: The recently proposed End-to-End Neural speaker Diarization framework (EEND) handles speech overlap and speech activity detection natively.While extensions of this work have reported remarkable results in both two-speaker and multispeaker diarization scenarios, these come at the cost of a long training process that requires considerable memory and computational power. In this work, we explore the integration of efficient transformer variants into the Self-Attentive EEND with Encoder-Decoder based Attractors (S… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…Both works have also replaced the feature subsampling used on the input by convolutional subsampling and [Leung and Samarakoon, 2021b] introduced a convolutional upsampling to return more fine-grained outputs. Given the quadratic complexity of attention on the sequence length, a linear approximation was evaluated in the context of EEND-EDA in [Izquierdo del Alamo et al, 2022].…”
Section: Eend Extensionsmentioning
confidence: 99%
“…Both works have also replaced the feature subsampling used on the input by convolutional subsampling and [Leung and Samarakoon, 2021b] introduced a convolutional upsampling to return more fine-grained outputs. Given the quadratic complexity of attention on the sequence length, a linear approximation was evaluated in the context of EEND-EDA in [Izquierdo del Alamo et al, 2022].…”
Section: Eend Extensionsmentioning
confidence: 99%