Proceedings of the 15th Annual Conference on Genetic and Evolutionary Computation 2013
DOI: 10.1145/2463372.2463506
|View full text |Cite
|
Sign up to set email alerts
|

Learning regression ensembles with genetic programming at scale

Abstract: In this paper we examine the challenge of producing ensembles of regression models for large datasets. We generate numerous regression models by concurrently executing multiple independent instances of a genetic programming learner. Each instance may be configured with different parameters and a different subset of the training data. Several strategies for fusing predictions from multiple regression models are compared. To overcome the small memory size of each instance, we challenge our framework to learn fro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
11
1

Year Published

2014
2014
2021
2021

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 17 publications
(12 citation statements)
references
References 23 publications
0
11
1
Order By: Relevance
“…Another important issue is the comparison against ensembles of trees [14]. Ensembles usually rely on the model performance, therefore, the best models are selected to generate a fused prediction.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Another important issue is the comparison against ensembles of trees [14]. Ensembles usually rely on the model performance, therefore, the best models are selected to generate a fused prediction.…”
Section: Discussionmentioning
confidence: 99%
“…The present work may be more related to ensembles of trees [14], a method in which several models (trees) are independently evolved, and then combined to provide a better forecast. The final fused prediction may be the average of the individual predictions, a weighted average, or another statistic.…”
Section: Related Workmentioning
confidence: 99%
“…Experiments showed the validity of the approach when compared to standard techniques for the task at hand. Other applications of ensemble methods to GP includes the use of querying-by-committee methods [26,2] and of a divide-andconquer strategy, in which ax solution need to work well only on a subset of the entire training set [31,1] With respect to ensembles of regression models, a quite recent contribution was proposed in [38]. The idea explored by the authors was to generate several regression models by concurrently executing multiple independent instances of a GP and, subsequently to analyze several strategies for fusing predictions from the multiple regression models.…”
Section: Related Workmentioning
confidence: 99%
“…The study considered only small datasets due to memory constraints, but authors were able to draw interesting conclusions about the suitability of their approach in producing accurate predictions. Our study will differ from the one described in [38] in several ways: we do not put any constraint on the size of the datasets, we will consider models produced by different GP algorithms (blend of STGP and GSGP) and we define and use different similarity-based criteria that, by taking into account the information related to all the populations evolved, aim at improving the generalization ability of the final ensemble as well as reducing the computational effort. Hence, in the experiments described in this contribution and as explained in Section 3, the populations evolved are not independent of each other.…”
Section: Related Workmentioning
confidence: 99%
“…A windowing method that divides the training dataset into non-overlapping strata preserving the same class distribution, named ILAS, was presented in [11]. Note that other types of techniques such as subgroup discovery [20], frequent patterns mining in data streams [24] or regression models [62] have been also tackled by windowing mechanisms.…”
Section: Related Workmentioning
confidence: 99%