Proceedings of the 26th Annual International Conference on Machine Learning 2009
DOI: 10.1145/1553374.1553415
|View full text |Cite
|
Sign up to set email alerts
|

A majorization-minimization algorithm for (multiple) hyperparameter learning

Abstract: We present a general Bayesian framework for hyperparameter tuning in L 2 -regularized supervised learning models. Paradoxically, our algorithm works by first analytically integrating out the hyperparameters from the model. We find a local optimum of the resulting nonconvex optimization problem efficiently using a majorization-minimization (MM) algorithm, in which the non-convex problem is reduced to a series of convex L 2 -regularized parameter estimation tasks. The principal appeal of our method is its simpli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2010
2010
2023
2023

Publication Types

Select...
6
1
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 23 publications
(23 citation statements)
references
References 18 publications
0
23
0
Order By: Relevance
“…We use the Bayesian regularization method in [11] for the setting of the regularization parameters (i.e., λ 1 and λ 2 ) in all models. The core idea of [11] is to first place a Gamma prior on each of the regularization parameters and then integrate out the regularization parameters.…”
Section: Experimental Settingsmentioning
confidence: 99%
See 1 more Smart Citation
“…We use the Bayesian regularization method in [11] for the setting of the regularization parameters (i.e., λ 1 and λ 2 ) in all models. The core idea of [11] is to first place a Gamma prior on each of the regularization parameters and then integrate out the regularization parameters.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…The core idea of [11] is to first place a Gamma prior on each of the regularization parameters and then integrate out the regularization parameters. By using the majorization-minimization (MM) algorithm [14], [15], [22], the objective function in each iteration is similar to the original problem with the regularization parameters inversely depending on the solution in the previous iteration.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…In the field of machine learning, MM algorithm has been applied to parameters selection for bayesian classification [23]. In the area of signal processing, MM algorithm leads to a number of interesting applications, including wavelet-based processing [24] and total variation (TV) minimization [25].…”
Section: B MM Algorithm and Reweighted Approachesmentioning
confidence: 99%
“…We note that, to preserve the posterior covariance matrix, these parameters are not integrated out as discussed in [10].…”
Section: E Learning Additional Model Parametersmentioning
confidence: 99%