2021 55th Asilomar Conference on Signals, Systems, and Computers 2021
DOI: 10.1109/ieeeconf53345.2021.9723318
|View full text |Cite
|
Sign up to set email alerts
|

The effectiveness of self-supervised representation learning in zero-resource subword modeling

Abstract: For a language with no transcribed speech available (the zero-resource scenario), conventional acoustic modeling algorithms are not applicable. Recently, zero-resource acoustic modeling has gained much interest. One research problem is unsupervised subword modeling (USM), i.e., learning a feature representation that can distinguish subword units and is robust to speaker variation. Previous studies showed that self-supervised learning (SSL) has the potential to separate speaker and phonetic information in speec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 19 publications
(57 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?