2019
DOI: 10.1007/978-3-030-30967-1_9
|View full text |Cite
|
Sign up to set email alerts
|

Optimizing Ensemble Weights for Machine Learning Models: A Case Study for Housing Price Prediction

Abstract: Designing ensemble learners has been recognized as one of the significant trends in the field of data knowledge especially in data science competitions. Building models that are able to outperform all individual models in terms of bias, which is the error due to the difference in the average model predictions and actual values, and variance, which is the variability of model predictions, has been the main goal of the studies in this area. An optimization model has been proposed in this paper to design ensemble… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
36
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
3

Relationship

2
6

Authors

Journals

citations
Cited by 26 publications
(36 citation statements)
references
References 16 publications
0
36
0
Order By: Relevance
“…Based on bias and variance tradeoff, the objective function of the optimization problem can be mean squared error (MSE) of out-of-bag predictions for the ensemble ( Hastie et al., 2009 ). The out-of-bag predictions matrix created previously can be used as an emulator of unseen test observations ( Shahhosseini et al., 2019b ). Using the out-of-bag predictions, we propose an optimization problem which is a nonlinear convex optimization problem as follows.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Based on bias and variance tradeoff, the objective function of the optimization problem can be mean squared error (MSE) of out-of-bag predictions for the ensemble ( Hastie et al., 2009 ). The out-of-bag predictions matrix created previously can be used as an emulator of unseen test observations ( Shahhosseini et al., 2019b ). Using the out-of-bag predictions, we propose an optimization problem which is a nonlinear convex optimization problem as follows.…”
Section: Methodsmentioning
confidence: 99%
“…Studies show that a single machine learning model can be outperformed by a “committee” of individual models, which is called a machine learning ensemble ( Zhang and Ma, 2012 ). Ensemble learning is proved to be effective as it can reduce bias, variance, or both and is able to better capture the underlying distribution of the data in order to make better predictions, if the base learners are diverse enough ( Dietterich, 2000 ; Pham and Olafsson, 2019a ; Pham and Olafsson, 2019b ; Shahhosseini et al., 2019a ; Shahhosseini et al., 2019b ). The usage of ensemble learning in ecological problems is becoming more widespread; for instance, bagging and specifically random forest ( Vincenzi et al., 2011 ; Mutanga et al., 2012 ; Fukuda et al., 2013 ; Jeong et al., 2016 ), boosting ( De'ath, 2007 ; Heremans et al., 2015 ; Belayneh et al., 2016 ; Stas et al., 2016 ; Sajedi-Hosseini et al., 2018 ), and stacking ( Conţiu and Groza, 2016 ; Cai et al., 2017 ; Shahhosseini et al., 2019a ), are some of the ensemble learning applications in agriculture.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, an average weighted ensemble that assigns equal weights to all base learners is the simplest ensemble model created. Additionally, optimized weighted ensemble method proposed in Shahhosseini et al 60 was applied here to test its predictive performance. Several two-level stacking ensembles, namely stacked regression, stacked LASSO, stacked random forest, and stacked LightGBM, were built, which are expected to demonstrate excellent performance.…”
Section: Methodsmentioning
confidence: 99%
“…An optimization model was proposed in Shahhosseini et al 60 , which accounts for the tradeoff between bias and variance of the predictions, as it uses mean squared error (MSE) to form the objective function for the optimization problem 68 . In addition, out-of-bag predictions generated by -fold cross-validation are used as emulators of unseen test observations to create the input matrices of the optimization problem, which are out-of-bag predictions made by each base learner.…”
Section: Methodsmentioning
confidence: 99%
“…Shahhosseini et al [41] compare the behaviour of several ensemble models for the prediction of dwelling prices using two databases, widely cited in the relevant literature, the Boston metropolitan area dataset [42] and the sales database of residential homes in Ames (Iowa) presented in [43]. To demonstrate the validity of the ensemble models, they use the following algorithms: multiple learners including lasso regression, random forest, deep neural networks, extreme gradient boosting (XGBoost), and support vector machines with three kernels (polynomial, RBF, and sigmoid).…”
Section: Literature Reviewmentioning
confidence: 99%