2018 IEEE 23rd International Conference on Digital Signal Processing (DSP) 2018
DOI: 10.1109/icdsp.2018.8631697
|View full text |Cite
|
Sign up to set email alerts
|

End-to-End Residual CNN with L-GM Loss Speaker Verification System

Abstract: We propose an end-to-end speaker verification system based on the neural network and trained by a loss function with less computational complexity. The end-to-end speaker verification system in this paper consists of a ResNet architecture to extract features from utterance, then produces utterancelevel speaker embeddings, and train using the large-margin Gaussian Mixture loss function. Influenced by the large-margin and likelihood regularization, large-margin Gaussian Mixture loss function benefits the speaker… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(2 citation statements)
references
References 17 publications
0
2
0
Order By: Relevance
“…Deep residual learning has been successfully applied to different tasks like noise-robust speech recognition [26] and speaker verification [27].…”
Section: A Deep Residual Learningmentioning
confidence: 99%
“…Deep residual learning has been successfully applied to different tasks like noise-robust speech recognition [26] and speaker verification [27].…”
Section: A Deep Residual Learningmentioning
confidence: 99%
“…In the first, the network is trained as a multi-class classification network that classifies a large number of classes (speakers in our case). These networks use objectives that include, in addition to traditional classification losses, augmentations that are intended to encourage enhanced within-class clustering of embeddings [3,24,25,26,27,28,29] along with increased separation of embeddings of instances from different classes. The expectation is that this behavior will generalize to data outside the training set.…”
Section: Introductionmentioning
confidence: 99%