2017
DOI: 10.1214/17-ejs1335si
|View full text |Cite
|
Sign up to set email alerts
|

Fast Bayesian hyperparameter optimization on large datasets

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
118
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 107 publications
(120 citation statements)
references
References 12 publications
2
118
0
Order By: Relevance
“…are the holdout and crossvalidation error for a user-given loss function (such as misclassification rate); see Bischl et al [16] for an overview of validation protocols. Several strategies for reducing the evaluation time have been proposed: It is possible to only test machine learning algorithms on a subset of folds [149], only on a subset of data [78,102,147], or for a small amount of iterations; we will discuss some of these strategies in more detail in Sect. 1.4.…”
Section: Popular Choices For the Validation Protocol V(• • • •)mentioning
confidence: 99%
See 3 more Smart Citations
“…are the holdout and crossvalidation error for a user-given loss function (such as misclassification rate); see Bischl et al [16] for an overview of validation protocols. Several strategies for reducing the evaluation time have been proposed: It is possible to only test machine learning algorithms on a subset of folds [149], only on a subset of data [78,102,147], or for a small amount of iterations; we will discuss some of these strategies in more detail in Sect. 1.4.…”
Section: Popular Choices For the Validation Protocol V(• • • •)mentioning
confidence: 99%
“…Multi-task Bayesian optimization (and the methods presented in the previous subsection) requires an upfront specification of a set of fidelities. This can be suboptimal since these can be misspecified [74,78] and because the number of fidelities that can be handled is low (usually five or less). Therefore, and in order to exploit the typically smooth dependence on the fidelity (such as, e.g., size of the data subset used), it often yields better results to treat the fidelity as continuous (and, e.g., choose a continuous percentage of the full data set to evaluate a configuration on), trading off the information gain and the time required for evaluation [78].…”
Section: Adaptive Choices Of Fidelitiesmentioning
confidence: 99%
See 2 more Smart Citations
“…Therefore, BOA can overcome the disruption of building blocks in genetic algorithms. The BOA has advantages in the optimization of machine learning algorithm hyperparameters, because of its faster search speed and fewer iteration compared to traditional search algorithms [28][29][30]. In this study, the BOA is employed to optimize the parameters of Random Forest (which is the basic model, see Section 2.3 for details) for traffic incident duration prediction, in order to achieve better prediction results.…”
Section: Methodsmentioning
confidence: 99%