2010
DOI: 10.1002/cjs.10083
|View full text |Cite
|
Sign up to set email alerts
|

New estimation and feature selection methods in mixture‐of‐experts models

Abstract: We study estimation and feature selection problems in mixture‐of‐experts models. An $l_2$‐penalized maximum likelihood estimator is proposed as an alternative to the ordinary maximum likelihood estimator. The estimator is particularly advantageous when fitting a mixture‐of‐experts model to data with many correlated features. It is shown that the proposed estimator is root‐$n$ consistent, and simulations show its superior finite sample behaviour compared to that of the maximum likelihood estimator. For feature … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
43
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 40 publications
(44 citation statements)
references
References 31 publications
(35 reference statements)
1
43
0
Order By: Relevance
“…It is well documented that maximum likelihood (ML) estimation can be used to obtain consistent estimates of the mean function, when the number of components g is known; see [18] for results regarding the ML estimator for MoE models with normal component PDFs, [9] for results regarding the ML estimator when the component PDFs are from the oneparameter exponential family, and [14] for the case of Laplace component PDFs. Results regarding regularized ML estimation of MoE models were obtained in [11]. In [14] and [5], the Bayesian information criterion [15] is demonstrated to be effective for the determination of an unknown g.…”
Section: Discussionmentioning
confidence: 99%
“…It is well documented that maximum likelihood (ML) estimation can be used to obtain consistent estimates of the mean function, when the number of components g is known; see [18] for results regarding the ML estimator for MoE models with normal component PDFs, [9] for results regarding the ML estimator when the component PDFs are from the oneparameter exponential family, and [14] for the case of Laplace component PDFs. Results regarding regularized ML estimation of MoE models were obtained in [11]. In [14] and [5], the Bayesian information criterion [15] is demonstrated to be effective for the determination of an unknown g.…”
Section: Discussionmentioning
confidence: 99%
“…Furthermore, the added estimation of the covariate-dependent class probabilities introduces excess variability that is likely to adversely effect the performance of these models under the assumptions of the problem. Similarly to the regular FMR, a penalized HME using generalized linear models as experts (Khalili, 2010) requires the definition of the correct structure for the linear predictor.…”
Section: Introductionmentioning
confidence: 99%
“…The L-MLR was then generalized to the divergent number of variables setting in Khalili and Lin (2013), and to the mixture of experts setting in Khalili (2010).…”
Section: Introductionmentioning
confidence: 99%