2013
DOI: 10.1007/s10472-013-9367-5
|View full text |Cite
|
Sign up to set email alerts
|

Multiprobabilistic prediction in early medical diagnoses

Abstract: This paper describes the methodology of providing multiprobability predictions for proteomic mass spectrometry data. The methodology is based on a newly developed machine learning framework called Venn machines. Is allows to output a valid probability interval. The methodology is designed for mass spectrometry data. For demonstrative purposes, we applied this methodology to MALDI-TOF data sets in order to predict the diagnosis of heart disease and early diagnoses of ovarian cancer and breast cancer. The experi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
11
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
6
2

Relationship

3
5

Authors

Journals

citations
Cited by 10 publications
(11 citation statements)
references
References 11 publications
0
11
0
Order By: Relevance
“…Nevertheless, the large number of samples that do not offer details about the reliability of individual prediction [ 26 ] lead PAC to be not so appropriate for the herbal medicine classification problem. On the other hand, although methods such as Bayesian learning, logistic regression [ 27 ], and Platt’s method [ 28 ] do associate individual prediction with additional reliability information, they are usually based on stringent distribution assumptions. Notwithstanding, since data gathered from the E-nose is usually influenced by sensor drifts due the variations in the surrounding environment, the distribution assumptions cannot be readily satisfied.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Nevertheless, the large number of samples that do not offer details about the reliability of individual prediction [ 26 ] lead PAC to be not so appropriate for the herbal medicine classification problem. On the other hand, although methods such as Bayesian learning, logistic regression [ 27 ], and Platt’s method [ 28 ] do associate individual prediction with additional reliability information, they are usually based on stringent distribution assumptions. Notwithstanding, since data gathered from the E-nose is usually influenced by sensor drifts due the variations in the surrounding environment, the distribution assumptions cannot be readily satisfied.…”
Section: Introductionmentioning
confidence: 99%
“…Conformal prediction was issued and improved by Vladimir Vovk and his co-workers [ 26 , 27 , 29 , 30 ]. It is based on the identical and independent distribution assumptions which state that all samples and labels are generated from the same identical and independent distributions which is a weaker assumption when compared with the methods mentioned above.…”
Section: Introductionmentioning
confidence: 99%
“…In addition, PAC learning does not provide any information on the reliability of individual prediction [10]. In contrast, Bayesian learning and other probability algorithms, such as logistic regression [11] and Platt’s method [12], can complement every individual prediction with information of probability to indicate how every potential label is correct. However, the main disadvantage of these algorithms is that they depend on a strong statistical assumption for the model.…”
Section: Introductionmentioning
confidence: 99%
“…Conformal prediction was proposed and developed by Vladimir Vovk and his co-workers since 2005 [10,11,19,20]. It is based on a consistent and well-defined mathematical framework and measures how well the new example is conformed to the group of observations.…”
Section: Introductionmentioning
confidence: 99%
“…In this work, a flexible machine learning framework, Venn machine [ 12 , 13 , 14 ] was introduced to make valid probabilistic prediction for different species of ginseng samples, which means the estimated probability is unbiased. The only assumption required for Venn machine is that the example distribution is an independent identical distribution (I.I.D assumption), which can be easily satisfied by E-nose data, and do not need specific distribution of E-nose data Once the I.I.D assumption is satisfied, the validity of predicted probabilities is guaranteed.…”
Section: Introductionmentioning
confidence: 99%