Interspeech 2020 2020
DOI: 10.21437/interspeech.2020-1708
|View full text |Cite
|
Sign up to set email alerts
|

Metric Learning Loss Functions to Reduce Domain Mismatch in the x-Vector Space for Language Recognition

Abstract: State-of-the-art language recognition systems are based on discriminative embeddings called x-vectors. Channel and gender distortions produce mismatch in such x-vector space where embeddings corresponding to the same language are not grouped in an unique cluster. To control this mismatch, we propose to train the x-vector DNN with metric learning objective functions. Combining a classification loss with the metric learning n-pair loss allows to improve the language recognition performance. Such a system achieve… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 24 publications
(40 reference statements)
0
5
0
Order By: Relevance
“…A total of 7 and 6 systems were submitted to the closed and open tracks of MERLIon CCS challenge respectively. As the LID task (English and Mandarin) is binary and the challenge dataset is highly imbalanced, equal error rate was selected as a primary evaluation metric [23].…”
Section: Methodsmentioning
confidence: 99%
“…A total of 7 and 6 systems were submitted to the closed and open tracks of MERLIon CCS challenge respectively. As the LID task (English and Mandarin) is binary and the challenge dataset is highly imbalanced, equal error rate was selected as a primary evaluation metric [23].…”
Section: Methodsmentioning
confidence: 99%
“…First, this efficiently minimizes the distribution shift between source and target data [32]. Second, the language recognition literature successfully used MMD for reducing domain mismatch [22].…”
Section: Mmd-based Domain Generalization Approachesmentioning
confidence: 99%
“…• additive angular margin softmax (AAM) [27] as an alternative classification loss function , we used a margin parameter m = 0.1 and a radius s = 40 • regularization of the cross-entropy with maximum mean discrepancy (MMD) between mobile channel and unknown channels [10], we compared different weights λ for the regularization loss functions • regularization of the cross-entropy with n-pair loss [28]…”
Section: Exploring Other Loss Functionsmentioning
confidence: 99%