2007
DOI: 10.1007/s10439-007-9317-7
|View full text |Cite
|
Sign up to set email alerts
|

Large-Scale Optimization-Based Classification Models in Medicine and Biology

Abstract: Abstract-We present novel optimization-based classification models that are general purpose and suitable for developing predictive rules for large heterogeneous biological and medical data sets. Our predictive model simultaneously incorporates (1) the ability to classify any number of distinct groups; (2) the ability to incorporate heterogeneous types of attributes as input; (3) a high-dimensional data transformation that eliminates noise and errors in biological data; (4) the ability to incorporate constraint… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
39
0

Year Published

2008
2008
2018
2018

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 60 publications
(39 citation statements)
references
References 67 publications
0
39
0
Order By: Relevance
“…Since 1996, Lee and her medical colleagues have explored and demonstrated the capability of DAMIP in classifying various types of data arising from real biological and medical problems. DAMIP has been able to consistently maximize the correct classification rate (80%-100% correct rates were obtained) while satisfying pre-set limits on inter-group misclassifications (Gallagher et al 1996(Gallagher et al , 1997Feltus et al 2003Feltus et al , 2006Lee et al 2002Lee et al , 2004Lee 2007aLee , 2007bLee and Wu 2007). In these real applications, beyond reporting the tenfold cross-validation results, the resulting classification rule was also blind tested against new data of unknown group identity and resulted in remarkable rates of correct prediction.…”
Section: Estimating the Anderson Optimal Rule Via A Mixed Integer Promentioning
confidence: 86%
See 2 more Smart Citations
“…Since 1996, Lee and her medical colleagues have explored and demonstrated the capability of DAMIP in classifying various types of data arising from real biological and medical problems. DAMIP has been able to consistently maximize the correct classification rate (80%-100% correct rates were obtained) while satisfying pre-set limits on inter-group misclassifications (Gallagher et al 1996(Gallagher et al , 1997Feltus et al 2003Feltus et al , 2006Lee et al 2002Lee et al , 2004Lee 2007aLee , 2007bLee and Wu 2007). In these real applications, beyond reporting the tenfold cross-validation results, the resulting classification rule was also blind tested against new data of unknown group identity and resulted in remarkable rates of correct prediction.…”
Section: Estimating the Anderson Optimal Rule Via A Mixed Integer Promentioning
confidence: 86%
“…DAMIP consistently returned superior classification rates compared to other methods (80-100% correct classification rates) while satisfying pre-set limits on inter-group misclassifications (Gallagher et al 1996(Gallagher et al , 1997Feltus et al 2003Feltus et al , 2006Lee et al 2002Lee et al , 2004Lee 2007a, 2007b, Lee and Wu 2007. In these applications, the DAMIP model was solved using a specialized solver, MIPSOL, developed by Lee.…”
Section: Performance On Real-world Datamentioning
confidence: 99%
See 1 more Smart Citation
“…This approach enabled us to identify 131 differentially methylated gene promoters. In stage 2, a binary particle swarm optimization (PSO) algorithm combined with DAMIP was used to identify genes that displayed changes in promoter methylation (11,12,24,34). We reported the results with 100% 10-fold cross validation accuracy for both DCM and NF groups.…”
Section: Methodsmentioning
confidence: 99%
“…. ; ng: There has recently been considerable interest in solving instances of (1) of large dimensions (i.e., large values of n), motivated by the emergence of applications in bio-computing, web and data mining (Lee 2007;Nasiri et al 2009). Many current optimization techniques can efficiently solve instances of (1) of small to moderate dimension, involving up to 50 variables.…”
Section: Introductionmentioning
confidence: 99%