2008
DOI: 10.1007/s10479-008-0412-4
|View full text |Cite
|
Sign up to set email alerts
|

Quadratic mixed integer programming and support vectors for deleting outliers in robust regression

Abstract: We consider the problem of deleting bad influential observations (outliers) in linear regression models. The problem is formulated as a Quadratic Mixed Integer Programming (QMIP) problem, where penalty costs for discarding outliers are used into the objective function. The optimum solution defines a robust regression estimator called penalized trimmed squares (PTS). Due to the high computational complexity of the resulting QMIP problem, the proposed robust procedure is computationally suitable for small sample… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(14 citation statements)
references
References 16 publications
0
14
0
Order By: Relevance
“…From another point of view, datasets with outliers pose a serious challenge in regression analysis and many solution techniques have been proposed (e.g. Panagopoulos et al 2019, andZioutas et al 2009). Mielke and Berry (1997) have proposed L 1 regression when errors are generated from fat-tailed or outlier-producing distributions, which are common in operations research.…”
Section: Introductionmentioning
confidence: 99%
“…From another point of view, datasets with outliers pose a serious challenge in regression analysis and many solution techniques have been proposed (e.g. Panagopoulos et al 2019, andZioutas et al 2009). Mielke and Berry (1997) have proposed L 1 regression when errors are generated from fat-tailed or outlier-producing distributions, which are common in operations research.…”
Section: Introductionmentioning
confidence: 99%
“…We impose two separate integer constraints on 尾 and 蠁 in (1), combining in a single framework the use of L 0 constraints for feature selection (Bertsimas et al 2016;Kenney et al 2018;Bertsimas and Van Parys 2020) and outlier detection (Bertsimas and Mazumder 2014;Zioutas et al 2009). In particular, we consider the following MIP formulation:…”
Section: Mip Formulationmentioning
confidence: 99%
“…In contrast to existing methodologies which extensively rely on heuristics, we propose a discrete and provably optimal approach to perform SFSOD, highlighting its connections with other approaches. The L 0 constraint has been used separately in the context of feature selection (Bertsimas et al 2016;Bertsimas and Van Parys 2020) and robust estimation (Zioutas et al 2009;Bertsimas and Mazumder 2014) -both of which can be formulated as a Mixed-Integer Program (MIP) and solved with optimality guarantees. We combine these two approaches into a novel formulation and take advantage of existing heuristics to provide effective big-M bounds and warm-starts to reduce the computational burden of MIP.…”
Section: Introductionmentioning
confidence: 99%
“…Interestingly enough, the nonsparse case of the robust subsets problem (1.2) (i.e., the standard low-dimensional LTS estimator) was formulated and solved as a mixed-integer program in Zioutas et al (2009), and their formulation can be considered (with certain small modifications) as a special case of our own mixed-integer program. The robust subsets estimator also bears close similarities to the high-dimensional robust regression estimator that appeared in Bhatia et al (2015).…”
Section: Robust Regressionmentioning
confidence: 99%