The 2011 International Joint Conference on Neural Networks 2011
DOI: 10.1109/ijcnn.2011.6033359
|View full text |Cite
|
Sign up to set email alerts
|

Modular deep belief networks that do not forget

Abstract: Abstract-Deep belief networks (DBNs) are popular for learning compact representations of highdimensional data. However, most approaches so far rely on having a single, complete training set. If the distribution of relevant features changes during subsequent training stages, the features learned in earlier stages are gradually forgotten. Often it is desirable for learning algorithms to retain what they have previously learned, even if the input distribution temporarily changes. This paper introduces the M-DBN, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 19 publications
0
7
0
Order By: Relevance
“…Introducing modularization to prevent forgetting dates back to [42] on the training of deep belief networks (DBN) [27]. Recently, [54] suggests a modular solution by identifying the trained modules (groups of neurons) to be re-used and extending the network with new modules for each new task.…”
Section: Regularization-based Continual Learning and Network Modulari...mentioning
confidence: 99%
“…Introducing modularization to prevent forgetting dates back to [42] on the training of deep belief networks (DBN) [27]. Recently, [54] suggests a modular solution by identifying the trained modules (groups of neurons) to be re-used and extending the network with new modules for each new task.…”
Section: Regularization-based Continual Learning and Network Modulari...mentioning
confidence: 99%
“…One variation of DBN, called modular DBN (M-DBN), trains different parts of the network separately, while adjusting the learning rate as training progresses (Pape et al 2011), as opposed to using one training set for the whole network. This allows M-DBN to avoid forgetting features learned early in training, a weakness of DBN that can hinder its performance in online learning applications in which the data distribution changes dynamically over time.…”
Section: Deep Network Similar To Dbnmentioning
confidence: 99%
“…However, some works have started addressing this issue in the context of deep networks. First, [67] proposes an organization in which new data are affected to the training of different subnetworks according to their current capacities. Another interesting approach was proposed by [8]: since deep networks are trained to generate their input data, these generated data can be used to extend the dataset.…”
Section: Perspectives and Future Workmentioning
confidence: 99%