2014
DOI: 10.1186/preaccept-1667880097114310
|View full text |Cite
|
Sign up to set email alerts
|

PLDA in the i-supervector space for text-independent speaker verification

Abstract: In this paper, we advocate the use of the uncompressed form of i-vector and depend on subspace modeling using probabilistic linear discriminant analysis (PLDA) in handling the speaker and session (or channel) variability. An i-vector is a low-dimensional vector containing both speaker and channel information acquired from a speech segment. When PLDA is used on an i-vector, dimension reduction is performed twice: first in the i-vector extraction process and second in the PLDA model. Keeping the full dimensional… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2015
2015
2018
2018

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 9 publications
0
3
0
Order By: Relevance
“…The Integer Linear Programming I-Vector (ILP IV) clustering [15] extracts an i-vector for each cluster and computes the distances among all of them (PLDA [16], cosine [17] or Mahalanobis [18]). ILP clustering was inspired by the k-medoids algorithm which choose k observations as class centers.…”
Section: Ilp IVmentioning
confidence: 99%
“…The Integer Linear Programming I-Vector (ILP IV) clustering [15] extracts an i-vector for each cluster and computes the distances among all of them (PLDA [16], cosine [17] or Mahalanobis [18]). ILP clustering was inspired by the k-medoids algorithm which choose k observations as class centers.…”
Section: Ilp IVmentioning
confidence: 99%
“…where N (x|µ, Σ) represents a Gaussian in x with mean µ and covariance Σ. Here it's worth to notice that the mathematical relationship between DoJoBa and joint Bayesian [15] is analogous (not exactly) to that between joint PLDA [11] and PLDA [16]. Compared to joint PLDA, DoJoBa allows the data to determine the appropriate dimensionality of the low-rank speaker and text subspaces for maximal discrimination, as opposed to requiring heuristic manual selections.…”
Section: Double Joint Bayesianmentioning
confidence: 99%
“…Secondly, dimensionality reduction is not possible. As D is a supervector-size diagonal matrix, the latent variable has to be of supervector length as well [14].…”
Section: The I-vector Paradigmmentioning
confidence: 99%