2014
DOI: 10.1007/s10723-014-9320-9
|View full text |Cite
|
Sign up to set email alerts
|

FlexGP

Abstract: We describe FlexGP, the first Genetic Programming system to perform symbolic regression on large-scale datasets on the cloud via massive data-parallel ensemble learning. FlexGP provides a decentralized, fault tolerant parallelization framework that runs many copies of Multiple Regression Genetic Programming, a sophisticated symbolic regression algorithm, on the cloud. Each copy executes with a different sample of the data and different parameters. The framework can create a fused model or ensemble on demand as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2014
2014
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(10 citation statements)
references
References 35 publications
0
10
0
Order By: Relevance
“…Instead of using one candidate, GP uses a group of individuals (known as population), formed by randomly combining mathematical building blocks such as constants, mathematical operators, analytic functions, state variables, and genetic operators, to make new individuals (generations) guided by fitness and complexity as objective functions that are meant to gauge the quality of each individual. The regression model 15,16 in Algorithm 1 (flowchart available in Reference 17) is implemented as an MOGP approach based on the work of Deb et al on nondominated sorting genetic algorithm II (NSGA-II). 18 Instead of fitness sharing (which was used in NSGA-I), NSGA-II uses the concept of crowding distance.…”
Section: Multiobjective Genetic Programmingmentioning
confidence: 99%
See 1 more Smart Citation
“…Instead of using one candidate, GP uses a group of individuals (known as population), formed by randomly combining mathematical building blocks such as constants, mathematical operators, analytic functions, state variables, and genetic operators, to make new individuals (generations) guided by fitness and complexity as objective functions that are meant to gauge the quality of each individual. The regression model 15,16 in Algorithm 1 (flowchart available in Reference 17) is implemented as an MOGP approach based on the work of Deb et al on nondominated sorting genetic algorithm II (NSGA-II). 18 Instead of fitness sharing (which was used in NSGA-I), NSGA-II uses the concept of crowding distance.…”
Section: Multiobjective Genetic Programmingmentioning
confidence: 99%
“…There are two values that the algorithm will minimize; first, the error, which can be the mean squared error (MSE) or mean absolute error (MAE), and second, the subtree complexity measure. 15,16,19 GP has great potential in predicting complex patterns. [20][21][22] It is also a suitable approach to be implemented using common parallelization frameworks such as message passing interface (MPI) and open multiprocessing (OpenMP).…”
Section: Multiobjective Genetic Programmingmentioning
confidence: 99%
“… EC has proven to work well in combination with many other AI techniques, including artificial neural networks [ 7 ] and other machine learning algorithms [ 8 ]. EC algorithms are inherently distributed, and are ripe for running in parallel on multi-core or distributed cloud-computing systems [ 9 ]. EC algorithms are anytime algorithms, meaning that they can provide a reasonable solution to a problem even when prematurely interrupted.…”
Section: Editorialmentioning
confidence: 99%
“…EC algorithms are inherently distributed, and are ripe for running in parallel on multi-core or distributed cloud-computing systems [ 9 ].…”
Section: Editorialmentioning
confidence: 99%
“…Finally we publish two papers based on the same ideas although with different implementations of the evolutionary algorithm [14]. The method is called Multiple Regression Genetic Programming (MRGP) which decouples and linearly combines a program's subexpressions via multiple regression on the target variable.…”
Section: Description Of the Papersmentioning
confidence: 99%