2021
DOI: 10.1145/3429447
|View full text |Cite
|
Sign up to set email alerts
|

Mitigating Class-Boundary Label Uncertainty to Reduce Both Model Bias and Variance

Abstract: The study of model bias and variance with respect to decision boundaries is critically important in supervised learning and artificial intelligence. There is generally a tradeoff between the two, as fine-tuning of the decision boundary of a classification model to accommodate more boundary training samples (i.e., higher model complexity) may improve training accuracy (i.e., lower bias) but hurt generalization against unseen data (i.e., higher variance). By focusing on just classification boundary fine-tuning a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(9 citation statements)
references
References 22 publications
0
9
0
Order By: Relevance
“…This can imply that by employing weighted-down labels (discussed in Section 3.3), we can achieve a reduction in both model bias and variance, hence, leading to an improved performance. Although the performance gain achieved using weighted-down labels technique from the study by [6] is marginal, for further studies involving data labels beyond the three categories used in this paper, the gains achieved using the weighted-down labels technique could be significant.…”
Section: Resultsmentioning
confidence: 91%
See 4 more Smart Citations
“…This can imply that by employing weighted-down labels (discussed in Section 3.3), we can achieve a reduction in both model bias and variance, hence, leading to an improved performance. Although the performance gain achieved using weighted-down labels technique from the study by [6] is marginal, for further studies involving data labels beyond the three categories used in this paper, the gains achieved using the weighted-down labels technique could be significant.…”
Section: Resultsmentioning
confidence: 91%
“…This could be due to its poor generalization to new unseen data. Based on the segmentation results obtained in Figures 10-13, we can note that using the random forests for all filter sizes with mtry = 5 and weighted-down labels technique from the study by Almeida et al [6], it is possible to segment the data into the three different categories of interest. Furthermore, mtry = 5 is in line with the study by Probst et al [18], which suggests that the recommended number of predictors is approximately equal to the square root value of the total number of features, which is 35 in our case.…”
Section: Discussionmentioning
confidence: 96%
See 3 more Smart Citations