2019
DOI: 10.1080/0952813x.2019.1647562
|View full text |Cite
|
Sign up to set email alerts
|

Population subset selection for the use of a validation dataset for overfitting control in genetic programming

Abstract: Genetic Programming (GP) is a technique which is able to solve different problems through the evolution of mathematical expressions. However, in order to be applied, its tendency to overfit the data is one of its main issues. The use of a validation dataset is a common alternative to prevent overfitting in many Machine Learning (ML) techniques, including GP. But, there is one key point which differentiates GP and other ML techniques: instead of training a single model, GP evolves a population of models. Theref… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…However, the modelling of CS in previous studies is solely based on survey data. Moreover, the CS models developed based on conventional GP based methods could not adapt to new datasets due to overfitting (Chen et al 2019;Rivero et al 2019). On the other hand, the solutions generated based on chaos-based optimisation could be unstable due to the likelihood of the unstable structures generated for CS models.…”
Section: Related Literaturementioning
confidence: 99%
“…However, the modelling of CS in previous studies is solely based on survey data. Moreover, the CS models developed based on conventional GP based methods could not adapt to new datasets due to overfitting (Chen et al 2019;Rivero et al 2019). On the other hand, the solutions generated based on chaos-based optimisation could be unstable due to the likelihood of the unstable structures generated for CS models.…”
Section: Related Literaturementioning
confidence: 99%
“…How can we select models according to both the training performance and the complexity measure? Researchers have been working on measuring and controlling model complexity in GP for decades [5], [6], [7], [8], [9]. Existing complexity measures in GP can be classified into two groups, which are structural/genotypic complexity measures and behavioural/phenotypic complexity measures.…”
Section: Introductionmentioning
confidence: 99%