2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472726
|View full text |Cite
|
Sign up to set email alerts
|

Content-aware local variability vector for speaker verification with short utterance

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
9
0

Year Published

2017
2017
2018
2018

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(9 citation statements)
references
References 12 publications
0
9
0
Order By: Relevance
“…Speaker verification performs almost ideally in neutral talking environment, while it performs poorly in emotional talking environments. There are many studies that study "speaker verification in neutral environment" [2][3][4][5][6], while few studies spotlight on "speaker verification in emotional environments" [7][8][9][10][11].…”
Section: Prior Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Speaker verification performs almost ideally in neutral talking environment, while it performs poorly in emotional talking environments. There are many studies that study "speaker verification in neutral environment" [2][3][4][5][6], while few studies spotlight on "speaker verification in emotional environments" [7][8][9][10][11].…”
Section: Prior Workmentioning
confidence: 99%
“…There are many research [2][3][4][5][6] that study speaker verification in neutral environments. The authors of [2] aimed in one of their work at addressing the long-term speaker variability problem in the feature domain in which they extracted more exact speaker-specific and time-insensitive information.…”
Section: Prior Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Test utterances are then scored with subregion models. In Chen et al (2016), the authors use the local session variability vectors estimated from certain phonetic components instead of computing the i-vector from the whole utterance. Phonetic classes are obtained by clustering similar senones (group of triphones with similar acoustic properties) that are estimated from posterior probabilities of a DNN trained for phone state classification.…”
Section: Introductionmentioning
confidence: 99%