2018
DOI: 10.1016/j.neunet.2017.11.010
|View full text |Cite
|
Sign up to set email alerts
|

Towards understanding sparse filtering: A theoretical perspective

Abstract: In this paper we present a theoretical analysis to understand sparse filtering, a recent and effective algorithm for unsupervised learning. The aim of this research is not to show whether or how well sparse filtering works, but to understand why and when sparse filtering does work. We provide a thorough theoretical analysis of sparse filtering and its properties, and further offer an experimental validation of the main outcomes of our theoretical analysis. We show that sparse filtering works by explicitly maxi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
19
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 34 publications
(19 citation statements)
references
References 37 publications
0
19
0
Order By: Relevance
“…For deep learning methods, for instance, deep belief networks [5], sparse filtering [6], and autoencoders (AEs) [7,8], the main assumption is that the datasets applied to train and test the model have the same feature distribution. Unfortunately, the raw vibration signals are usually obtained under variable working cases in practical applications, which show deviation from the assumption [9,10].…”
Section: Introductionmentioning
confidence: 99%
“…For deep learning methods, for instance, deep belief networks [5], sparse filtering [6], and autoencoders (AEs) [7,8], the main assumption is that the datasets applied to train and test the model have the same feature distribution. Unfortunately, the raw vibration signals are usually obtained under variable working cases in practical applications, which show deviation from the assumption [9,10].…”
Section: Introductionmentioning
confidence: 99%
“…SF is just a two-layer neural network, which is much simpler than DBN and AEs. Therefore, SF can overcome the parameter tuning difficulty and converge easily to an optimal solution [28]. Meanwhile, SF attempts to learn discriminative features by optimizing the sparsity of latent features, instead of learning the principal components of the input data [29].…”
Section: Introductionmentioning
confidence: 99%
“…Classical unsupervised learning methods include kernel principle component analysis (KPCA) [30], restricted Boltzmann machine (RBM) [31], sparse autoencoder (SAE) [32], and sparse filtering (SF) [33]. Compared with other approaches, SF optimizes feature sparsity distribution instead of modeling data distribution [34]. It is different from other conventional or novel neural network (NN) structures such as variable NNs in fault diagnosis application [35].…”
Section: Introductionmentioning
confidence: 99%