2017
DOI: 10.48550/arxiv.1708.03735
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sparse Coding and Autoencoders

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
11
0

Year Published

2018
2018
2019
2019

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(11 citation statements)
references
References 0 publications
0
11
0
Order By: Relevance
“…We conclude with the case of training two-layer autoencoders whose weights are shared (i.e., A = W ). This is a common architectural choice in practice, and indeed previous theoretical analysis for autoencoders [Rangamani et al, 2017, Nguyen et al, 2019, Li and Phan-Minh, 2019 have focused on this setting. We will show that somewhat surprisingly, allowing the network to be over-parameterized in this setting leads to certain degeneracies.…”
Section: Weight-tied Autoencodersmentioning
confidence: 99%
See 1 more Smart Citation
“…We conclude with the case of training two-layer autoencoders whose weights are shared (i.e., A = W ). This is a common architectural choice in practice, and indeed previous theoretical analysis for autoencoders [Rangamani et al, 2017, Nguyen et al, 2019, Li and Phan-Minh, 2019 have focused on this setting. We will show that somewhat surprisingly, allowing the network to be over-parameterized in this setting leads to certain degeneracies.…”
Section: Weight-tied Autoencodersmentioning
confidence: 99%
“…None of the above papers have rigorously studied the training dynamics of autoencoder models. The loss surface of autoencoder training was first characterized in Rangamani et al [2017]. Subsequently, Nguyen et al [2019] proved that under-parameterized (and suitably initialized) autoencoders performed (approximate) proper parameter learning in the regime of asymptotically many samples, building upon techniques in provable dictionary learning; cf.…”
Section: Introductionmentioning
confidence: 99%
“…Our focus here is on shallow two-layer autoencoder architectures with shared weights. Conceptually, we build upon previous theoretical results on learning autoencoder networks [Arora et al, 2014a, 2015a, Rangamani et al, 2017, and we elaborate on the novelty of our work in the discussion on prior work below.…”
Section: Our Contributionsmentioning
confidence: 99%
“…samples from a high-dimensional distribution parameterized by a generative model, and we train the weights of the autoencoder using ordinary (batch) gradient descent. We consider three families of generative models that are commonly adopted in machine learning: (i) the Gaussian mixture model with well-separated centers [Arora and Kannan, 2005]; (ii) the k-sparse model, specified by sparse linear combination of atoms [Spielman et al, 2012]; and (iii) the non-negative k-sparse model [Rangamani et al, 2017]. While these models are traditionally studied separately depending on the application, all of these model families can be expressed via a unified, generic form:…”
Section: Our Contributionsmentioning
confidence: 99%
See 1 more Smart Citation