2017
DOI: 10.1007/s13755-017-0023-z
|View full text |Cite
|
Sign up to set email alerts
|

Progressive sampling-based Bayesian optimization for efficient and automatic machine learning model selection

Abstract: Purpose: Machine learning is broadly used for clinical data analysis. Before training a model, a machine learning algorithm must be selected. Also, the values of one or more model parameters termed hyper-parameters must be set. Selecting algorithms and hyper-parameter values requires advanced machine learning knowledge and many laborintensive manual iterations. To lower the bar to machine learning, miscellaneous automatic selection methods for algorithms and/or hyper-parameter values have been proposed. Existi… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
57
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 71 publications
(57 citation statements)
references
References 30 publications
0
57
0
Order By: Relevance
“…Instead of switching between lower fidelities and the highest fidelity, it is possible to perform HPO on a subset of the original data and extract the best-performing configurations in order to use them as an initial design for HPO on the full dataset [152]. To speed up solutions to the CASH problem, it is also possible to iteratively remove entire algorithms (and their hyperparameters) from the configuration space based on poor performance on small dataset subsets [159].…”
Section: Bandit-based Algorithm Selection Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Instead of switching between lower fidelities and the highest fidelity, it is possible to perform HPO on a subset of the original data and extract the best-performing configurations in order to use them as an initial design for HPO on the full dataset [152]. To speed up solutions to the CASH problem, it is also possible to iteratively remove entire algorithms (and their hyperparameters) from the configuration space based on poor performance on small dataset subsets [159].…”
Section: Bandit-based Algorithm Selection Methodsmentioning
confidence: 99%
“…1.4), Bayesian optimization with meta-learning (see Chap. 2), and Bayesian optimization taking the pipeline structure into account [159,160]. Furthermore, many recent developments in Bayesian optimization do not directly target HPO, but can often be readily applied to HPO, such as new acquisition functions, new models and kernels, and new parallelization schemes.…”
Section: Bayesian Optimizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Sequential model-based optimization methods [7,8] and particularly Bayesian optimization methods [9][10][11][12][13][14] were also used in the hyper-parameter optimization. Bayesian search uses a completely different approach compared to the grid search and random search, as the latter ignores the information of the previous search point in the search process, but Bayesian search makes full use of the previous search point information.…”
Section: Introductionmentioning
confidence: 99%
“…An example hyper-parameter is the number of hidden layers in a deep neural network. Different combinations of algorithms and hyper-parameter values often impact model accuracy by 40% or more [71] and model building cost by several orders of magnitude [80]. According to the “no free lunch” theorem [76], no single combination performs well on model accuracy for every modeling problem.…”
Section: Introductionmentioning
confidence: 99%