Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-1393
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Representation Learning Using Convolutional Restricted Boltzmann Machine for Spoof Speech Detection

Abstract: Speech Synthesis (SS) and Voice Conversion (VC) presents a genuine risk of attacks for Automatic Speaker Verification (ASV) technology. In this paper, we use our recently proposed unsupervised filterbank learning technique using Convolutional Restricted Boltzmann Machine (ConvRBM) as a frontend feature representation. ConvRBM is trained on training subset of ASV spoof 2015 challenge database. Analyzing the filterbank trained on this dataset shows that ConvRBM learned more low-frequency subband filters compared… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 27 publications
0
5
0
Order By: Relevance
“…Before the recent boom of self-supervised models, there is a pioneering work using conditional Restricted Bolzman Machine (ConvRBM) as a trainable front end [76]. It showed competitive performance on the ASVspoof 2015 database.…”
Section: Front End: Dnn-based Self-supervised Training Approachmentioning
confidence: 99%
“…Before the recent boom of self-supervised models, there is a pioneering work using conditional Restricted Bolzman Machine (ConvRBM) as a trainable front end [76]. It showed competitive performance on the ASVspoof 2015 database.…”
Section: Front End: Dnn-based Self-supervised Training Approachmentioning
confidence: 99%
“…Deep learning methods were also frequently applied in PAD tasks. Unlike conventional classifiers, deep learning is one of the machine learning that is composed of networks capable of learning without supervision from labeled data [39,104]. From recent works, it is found that deep learning classifiers such as DNN [120], RNN [30], and CNN [88] are capable of automatic feature abstraction in which more informative features can be identified.…”
Section: Classifiersmentioning
confidence: 99%
“…Bottleneck features extracted from the DNN hidden layers were also used with GMM classifier in [102]. In [103], the Convolutional Restricted Boltzmann Machine (ConvRBM) is used for auditory filterbank learning that performed better than traditionally handcrafted filterbanks. The study [103] shows that ConvRBM learns better low-frequency subband filters on ASVspoof 2015 dataset than on TIMIT.…”
Section: ) Representation Learning Approachesmentioning
confidence: 99%
“…In [103], the Convolutional Restricted Boltzmann Machine (ConvRBM) is used for auditory filterbank learning that performed better than traditionally handcrafted filterbanks. The study [103] shows that ConvRBM learns better low-frequency subband filters on ASVspoof 2015 dataset than on TIMIT. Supervised auditory filterbank learning using DNN was also studied in [104].…”
Section: ) Representation Learning Approachesmentioning
confidence: 99%