2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2014
DOI: 10.1109/icassp.2014.6854056
|View full text |Cite
|
Sign up to set email alerts
|

Contrastive auto-encoder for phoneme recognition

Abstract: Speech data typically contains task irrelevant information lying within features. Specifically, phonetic information, speaker characteristic information, emotional information and noise are always mixed together and tend to impair one another for certain task. We propose a new type of auto-encoder for feature learning called contrastive auto-encoder. Unlike other variants of auto-encoders, contrastive auto-encoder is able to leverage class labels in constructing its representation layer. We achieve this by mod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 9 publications
0
9
0
Order By: Relevance
“…We also introduce an improved cAE architecture and training method that reduces the number of hyperparameters to be tuned, and show that narrow architectures work better, with reduced error rates on a zero-resource language after tuning on English. Unlike similar previous work [9,10,11] our cAE-based systems are fully unsupervised, train on individual frames without context, and use a loss function in the input vector space instead of the representation vector space.…”
Section: Introductionmentioning
confidence: 99%
“…We also introduce an improved cAE architecture and training method that reduces the number of hyperparameters to be tuned, and show that narrow architectures work better, with reduced error rates on a zero-resource language after tuning on English. Unlike similar previous work [9,10,11] our cAE-based systems are fully unsupervised, train on individual frames without context, and use a loss function in the input vector space instead of the representation vector space.…”
Section: Introductionmentioning
confidence: 99%
“…No Rifai et al (2011a) Higher order Contractive autoencoder: CAE + Hessian of the output wrt the input. No Zheng et al (2014) Contrastive autoencoder: A term to reduce the intra-class variations between the learned representation of samples belonging to the same class is added at the final layer.…”
Section: Nomentioning
confidence: 99%
“…Contrastive Autoencoder (CsAE) proposed by Zheng et al (2014), is another variant of supervised autoencoder which uses the class label information during training. The loss function of the model is the difference between the output of two sub-autoencoders trained simultaneously on samples belonging to the same class, along with the loss function of each subautoencoder.…”
Section: Nomentioning
confidence: 99%
“…Recently, supervised extensions of traditional unsupervised model of the autoencoder have also been proposed [17], [18], [19], [20]. Most of these algorithms incorporate class information at the time of feature extraction with an aim to reduce only the intra-class variations.…”
Section: B Proposed Class Representative Autoencodermentioning
confidence: 99%