2023
DOI: 10.3390/diagnostics13142383
|View full text |Cite
|
Sign up to set email alerts
|

Application of Machine Learning Models for Early Detection and Accurate Classification of Type 2 Diabetes

Orlando Iparraguirre-Villanueva,
Karina Espinola-Linares,
Rosalynn Ornella Flores Castañeda
et al.

Abstract: Early detection of diabetes is essential to prevent serious complications in patients. The purpose of this work is to detect and classify type 2 diabetes in patients using machine learning (ML) models, and to select the most optimal model to predict the risk of diabetes. In this paper, five ML models, including K-nearest neighbor (K-NN), Bernoulli Naïve Bayes (BNB), decision tree (DT), logistic regression (LR), and support vector machine (SVM), are investigated to predict diabetic patients. A Kaggle-hosted Pim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(8 citation statements)
references
References 38 publications
0
8
0
Order By: Relevance
“…As a more sophisticated imputation technique, k-NN was applied to predict missing values based on similar data points. This method takes into account the relationships between attributes, ensuring that the imputed value is consistent with other attributes of the dataset ( 32 ). We opted for k-Nearest Neighbors (k-NN) imputation due to its effectiveness in handling datasets where similarity between instances suggests a correlation, as is common in medical data.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…As a more sophisticated imputation technique, k-NN was applied to predict missing values based on similar data points. This method takes into account the relationships between attributes, ensuring that the imputed value is consistent with other attributes of the dataset ( 32 ). We opted for k-Nearest Neighbors (k-NN) imputation due to its effectiveness in handling datasets where similarity between instances suggests a correlation, as is common in medical data.…”
Section: Methodsmentioning
confidence: 99%
“…This visual validation confirms that the median imputation has addressed the issue of zero values in the specified columns, resulting in a dataset that likely better represents the real-world distribution of these attributes. Algorithm 1 offers a structured representation of the imputation process, starting with identifying zeros, performing median imputation, and then using k-NN for a more advanced imputation ( 32 , 33 ).…”
Section: Methodsmentioning
confidence: 99%
“…A DT is a nonparametric supervised learning algorithm that can be used for both classification and regression tasks. It has a hierarchical tree structure consisting of root nodes, branches, internal nodes, and leaf nodes [32], [33], [34]. Depending on the available features, both types of nodes perform evaluations by forming homogeneous subsets represented, by leaf nodes or end nodes.…”
Section: A Decision Treementioning
confidence: 99%
“…The following are nonlinear activation functions such as ReLU. These layers learn to recognize simple image patterns and features such as edges, texture, and color [24].  Reduction layer: a reduction layer is usually added after the convolution layer to improve computational efficiency and reduce the size of the resulting features.…”
Section: Figure 1 Basic Cnn Architecturementioning
confidence: 99%