2014
DOI: 10.1109/taslp.2013.2290505
|View full text |Cite
|
Sign up to set email alerts
|

Factorized Sub-Space Estimation for Fast and Memory Effective I-vector Extraction

Abstract: Torino, where he leads the speech technology research group. He has published over 120 papers in the area of pattern recognition, artificial intelligence, and spoken language processing. His current research interests include all aspects of automatic speech recognition and its applications, in particular speaker and spoken language recognition.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
21
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 14 publications
(21 citation statements)
references
References 26 publications
0
21
0
Order By: Relevance
“…The slowness of training has often forced many researchers to limit their experimental validation, for example by limiting the number of training iterations, or by relaying on the results from a single run with random initialization. In addition, simplifications and approximations of the model have been proposed to reduce the computational load [6,7,8].…”
Section: Introductionmentioning
confidence: 99%
“…The slowness of training has often forced many researchers to limit their experimental validation, for example by limiting the number of training iterations, or by relaying on the results from a single run with random initialization. In addition, simplifications and approximations of the model have been proposed to reduce the computational load [6,7,8].…”
Section: Introductionmentioning
confidence: 99%
“…SPAM models are Gaussian mixture models with a subspace constraint, where each covariance matrix is represented as a weighted sum of globally shared full-rank matrices. In 2014, Sandro Cumani proposed an i-vector extractor factorization [16], for faster i-vector extraction and smaller memory footprint, where each row of the i-vector extractor matrix is represented as a linear combination of the atoms of a common dictionary with the assumption that it is not necessary to store all rows this matrix to perform ivector extraction.…”
Section: Introductionmentioning
confidence: 99%
“…In our approach to factorization, we were inspired by [16], but instead of factorizing each row, we perform factorization on the level submatrices of the i-vector extractor that represent individual GMM-UBM components. Also, our motivation is different, as we aim to greatly decrease the memory footprint and therefore substantially speedup the discriminative training.…”
Section: Introductionmentioning
confidence: 99%
“…Previous studies on rapid i-vector extraction have primarily optimized computations in the standard front-end factor analysis (FEFA) approach [1,2] by adopting new computational algorithms, often approximative in nature [6,8,9]. In this study, however, we focus on an alternative and straightforward compression of classic maximum a posteriori (MAP) [3] adapted GMM supervectors with a goal of obtaining fast execution times without compromising on ASV accuracy.…”
Section: Introductionmentioning
confidence: 99%