IEEE International Conference on Acoustics Speech and Signal Processing 2002
DOI: 10.1109/icassp.2002.5743811
|View full text |Cite
|
Sign up to set email alerts
|

A Monte-Carlo method for score normalization in Automatic Speaker Verification using Kullback-Leibler distances

Abstract: In this paper, we propose a new score normalization technique in Automatic Speaker Verification (ASV): the DNorm. The main advantage of this score normalization is that it does not need any additional speech data nor external speaker population, as opposed to the state-ofthe-art approaches. The D-Norm is based on the use of Kullback-Leibler (KL) distances in an ASV context. In a first step, we estimate the KL distances with a MonteCarlo method and we experimentally show that they are correlated with the verifi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2005
2005
2013
2013

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 20 publications
(17 citation statements)
references
References 4 publications
0
17
0
Order By: Relevance
“…In this respect, the technique is related to the 'Dnorm' (distance normalization) class of normalizing methods (Ben et al, 2002;Yuan et al, 2009). In these works, the distance between each speaker GMM and a UBM is estimated via approximations of the Kullback-Leibler (KL) divergence.…”
Section: Quality Measures For Speechmentioning
confidence: 99%
“…In this respect, the technique is related to the 'Dnorm' (distance normalization) class of normalizing methods (Ben et al, 2002;Yuan et al, 2009). In these works, the distance between each speaker GMM and a UBM is estimated via approximations of the Kullback-Leibler (KL) divergence.…”
Section: Quality Measures For Speechmentioning
confidence: 99%
“…A MonteCarlo based symmetric Kullback-Leibler (KL) distance is used to obtain a set of client and impostor data using client model and UBM respectively. However, results presented in [13] show that Z-norm always outperformed D-norm, particularly at low miss-detection rates.…”
Section: λ(X Testmentioning
confidence: 96%
“…As opposed to Z-norm, D-norm does not require any speech data to estimate the normalization parameters [13]. Here, pseudo-impostor data is generated using the UBM.…”
Section: λ(X Testmentioning
confidence: 99%
“…In the terms used in [3,13], Z-Norm [13] is impostor-centric (i.e, normalisation is carried out with respect to the impostor distributions calculated "offline" by using additional data), T-Norm [13] is also impostor-centric (but with respect to a given utterance calculated "online" by using additional cohort impostor models). DNorm [14] is neither client-nor impostor-centric; it is specific to the Gaussian Mixture Model (GMM) architecture and is based on Kullback-Leibler distance between two GMM models. In [2], a client-centric version of Z-Norm was proposed.…”
Section: Deriving Client-dependent Informationmentioning
confidence: 99%