2020
DOI: 10.1109/access.2020.3003748
|View full text |Cite
|
Sign up to set email alerts
|

Additive Ensemble Neural Networks

Abstract: Deep neural networks (DNNs) have been making progress in many ways. DNNs are typically used to model complex nonlinearity of high-dimensional data in regression or classification problems. As DNNs contain additional hidden layers, they generally improve performance but increase the number of parameters to train, thereby extending the learning time. Many studies, such as those employing Dropout and regularization methods, have undertaken to solve these problems. The method proposed in this paper is an additive … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
6
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(6 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…There are several aggregation strategies, and boosting is one of the most important obtaining state-of-the-art estimators. Bringing together boosting and deep learning has shown very good results in other classification/regression problems [ 9 , 16 ]. The additive ensemble models considered in this work will follow the gaNet architecture [ 9 ], a deep learning boosting ensemble model specifically intended for time-series forecasting.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…There are several aggregation strategies, and boosting is one of the most important obtaining state-of-the-art estimators. Bringing together boosting and deep learning has shown very good results in other classification/regression problems [ 9 , 16 ]. The additive ensemble models considered in this work will follow the gaNet architecture [ 9 ], a deep learning boosting ensemble model specifically intended for time-series forecasting.…”
Section: Methodsmentioning
confidence: 99%
“…The only requirement for a base model is to be trainable end-to-end by gradient descent and support the addition of a final layer in both the training and prediction stages. Thus, we have considered as base models several configurations of 1D and 2D convolutional neural networks (CNN) [ 13 , 14 ], long short-term memory (LSTM) [ 15 ] networks and their combination, as well as several additive ensembles (AE) deep learning models especially suitable for time-series forecasting [ 9 , 16 ]. We do not include sequence-to-sequence (Seq2seq) models as a base model since the forward pass for the training, and test stages are different with added complexity for the proposed extension.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Examples of leveraging model complexity for better generalisation abound in the machine learning community -the most relatable being better generalisation through efficient dataset utilisation by using an ensemble of less and more complex models on different subsets of data points, each subset characterised by different difficulty of learning (Maini et al, 2022). Historically, the commonly used techniques to improve model generalisability, such as dropout (Srivastava et al, 2014) and pruning (Han et al, 2015), were inspired by reducing model complexity (Park et al, 2020). Thus, a model complexity for traffic prediction tasks furthers the research into better generalisable DL models for traffic prediction.…”
Section: Introductionmentioning
confidence: 99%
“…Bagging is the most famous representative of the parallel ensemble learning strategy [20]. Bagging can be combined with almost any learning algorithm to form an ensemble learning system, such as a neural network [21] or decision tree [22].…”
mentioning
confidence: 99%