2006
DOI: 10.1007/11729976_10
|View full text |Cite
|
Sign up to set email alerts
|

Genetic Programming, Validation Sets, and Parsimony Pressure

Abstract: Fitness functions based on test cases are very common in Genetic Programming (GP). This process can be assimilated to a learning task, with the inference of models from a limited number of samples. This paper is an investigation on two methods to improve generalization in GP-based learning: 1) the selection of the best-of-run individuals using a three data sets methodology, and 2) the application of parsimony pressure in order to reduce the complexity of the solutions. Results using GP in a binary classificati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
54
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 60 publications
(56 citation statements)
references
References 17 publications
2
54
0
Order By: Relevance
“…To determine loss for a build script for example, the value may be determined by counting the number of actions that execute successfully and dividing by the total number of steps. A further consideration in semantic evaluation is parsimony, which is the general expectation that the shortest adequate solution is to be preferred (Gagne et al, 2006). To incorporate parsimony in the evaluation we can add a measure(s) of the solution's cost(s), such as the size of the label y and/or execution resources consumed, to L.…”
Section: Loss Function Variationsmentioning
confidence: 99%
“…To determine loss for a build script for example, the value may be determined by counting the number of actions that execute successfully and dividing by the total number of steps. A further consideration in semantic evaluation is parsimony, which is the general expectation that the shortest adequate solution is to be preferred (Gagne et al, 2006). To incorporate parsimony in the evaluation we can add a measure(s) of the solution's cost(s), such as the size of the label y and/or execution resources consumed, to L.…”
Section: Loss Function Variationsmentioning
confidence: 99%
“…While generalisation has traditionally been underexplored in the GP literature, there have been a number of recent papers examining this important issue [3,6,9,11,19,22,25]. Among the techniques proposed to counteract overfitting, popular examples include the use of parsimony constraints, and the use of a validation set.…”
Section: Causes Of Overfittingmentioning
confidence: 99%
“…Gagné and co-authors [9] also investigate the use of three datasets (training data, validation data, and test data). They evolve solutions to binary classification problems using Genetic Programming.…”
Section: Validation Sets and Parsimonymentioning
confidence: 99%
See 2 more Smart Citations