Medical Imaging 2018: Image Perception, Observer Performance, and Technology Assessment 2018
DOI: 10.1117/12.2293818
|View full text |Cite
|
Sign up to set email alerts
|

Test data reuse for evaluation of adaptive machine learning algorithms: over-fitting to a fixed 'test' dataset and a potential solution

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
9
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 20 publications
0
9
0
Order By: Relevance
“…If the hypothesis is not powerful enough to describe the data, then there will be an issue of underfitting . On the other hand, if the selected hypothesis is over complicated, then the model will learn from not only the inherent trend of the data but also the noise, which end up with overfitting . The model should be carefully selected considering three factors: the complexity of the hypothesis, the complexity of the training data, and the generalization performance on new examples .…”
Section: Approaches In Computational Materials Sciencementioning
confidence: 99%
See 1 more Smart Citation
“…If the hypothesis is not powerful enough to describe the data, then there will be an issue of underfitting . On the other hand, if the selected hypothesis is over complicated, then the model will learn from not only the inherent trend of the data but also the noise, which end up with overfitting . The model should be carefully selected considering three factors: the complexity of the hypothesis, the complexity of the training data, and the generalization performance on new examples .…”
Section: Approaches In Computational Materials Sciencementioning
confidence: 99%
“…[24][25][26] On the other hand, if the selected hypothesis is over complicated, then the model will learn from not only the inherent trend of the data but also the noise, which end up with overfitting. [27][28][29] The model should be carefully selected considering three factors: the complexity of the hypothesis, the complexity of the training data, and the generalization performance on new examples. [30][31][32] A set of assumptions that work well in one domain may work poorly in another.…”
Section: Approaches In Computational Materials Sciencementioning
confidence: 99%
“…Lockbox data site D lb : Lockbox [13] data site refers to data sites which the analyst from the openbox side can not access by any means. In practice, lockbox correspond to data sites that could not contribute in the process of building a machine learning model due to various reasons, but are likely to participate in the future or simply benefit from the model built.…”
Section: Terminology and Notationmentioning
confidence: 99%
“…Thresholdout Family: [7] showes that differential privacy is deeply associated with model generalization and propose the Thresholdout algorithm to avoid overfitting on the validation set due to repetitive usage. [13] extends the instance wise Thresholdout to AUC measures. However, these methods rely on the i.i.d assumption of data which does not fit our scenario here.…”
Section: Related Workmentioning
confidence: 99%
“…Acquiring enough independent test data to validate each model, although the ideal, can become infeasible. An important and ongoing research area with many open questions is whether and how some test data can best be reused for validating future models (Dwork et al ., 2015; Gossmann et al ., 2018; Lee and Lee, 2020; Roelofs et al ., 2019; U.S. Food and Drug Administration, 2020a).…”
mentioning
confidence: 99%