2019
DOI: 10.1590/0104-6632.20190361s20170455
|View full text |Cite
|
Sign up to set email alerts
|

K-Rank: An Evolution of Y-Rank for Multiple Solutions Problem

Abstract: Y-rank can present faults when dealing with non-linear problems. A methodology is proposed to improve the selection of data in situations where y-rank is fragile. The proposed alternative, called k-rank, consists of splitting the data set into clusters using the k-means algorithm, and then apply y-rank to the generated clusters. Models were calibrated and tested with subsets split by y-rank and k-rank. For the Heating Tank case study, in 59% of the simulations, models calibrated with k-rank subsets achieved be… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 18 publications
0
2
0
Order By: Relevance
“…If the test metrics are consistent with the training metrics, the ability of the model to deal with new data is confirmed, and the models can be considered more robust and reliable. The train/test split was based on the y-rank methodology: the samples were sorted in ascending order of the output of interest (in this case, the percentage of lipid content), and then the samples were allocated to the train and test subsets following the pattern 2 (train) to 1 (test) [24,25]. The split resulted in a training subset with 66% of the total samples (8 samples) and a test subset with 34% (4 samples).…”
Section: Lipids Determination: Fluorescencementioning
confidence: 99%
“…If the test metrics are consistent with the training metrics, the ability of the model to deal with new data is confirmed, and the models can be considered more robust and reliable. The train/test split was based on the y-rank methodology: the samples were sorted in ascending order of the output of interest (in this case, the percentage of lipid content), and then the samples were allocated to the train and test subsets following the pattern 2 (train) to 1 (test) [24,25]. The split resulted in a training subset with 66% of the total samples (8 samples) and a test subset with 34% (4 samples).…”
Section: Lipids Determination: Fluorescencementioning
confidence: 99%
“…The methodology for splitting the data into calibration (cal), validation (val), and testing (test) subsets is the one implemented by Santos et al [45]. This methodology is especially useful when dealing with multiple solutions problems: situations where standardizes combinations of the input variables can yield the same output y.…”
Section: Data Set Splitting Based On a Modified Version Of K-rankmentioning
confidence: 99%