2021
DOI: 10.1007/s10489-021-02623-9
|View full text |Cite
|
Sign up to set email alerts
|

BBW: a batch balance wrapper for training deep neural networks on extremely imbalanced datasets with few minority samples

Abstract: In recent years, Deep Neural Networks (DNNs) have achieved excellent performance on many tasks, but it is very difficult to train good models from imbalanced datasets. Creating balanced batches either by majority data down-sampling or by minority data up-sampling can solve the problem in certain cases. However, it may lead to learning process instability and overfitting. In this paper, we propose the Batch Balance Wrapper (BBW), a novel framework which can adapt a general DNN to be well trained from extremely … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 30 publications
0
2
0
Order By: Relevance
“…According to the findings of our previous studies [8,13], the proportion of patients with distant metastases is relatively small. To ensure a balanced data set, we applied a batch balance wrapper during training [19]. Batch balance was also used to ensure that the samples in each batch were always balanced during the learning process (Appendix 1, Supplemental digital content 1, http://links.lww.com/NMC/A273).…”
Section: Methodsmentioning
confidence: 99%
“…According to the findings of our previous studies [8,13], the proportion of patients with distant metastases is relatively small. To ensure a balanced data set, we applied a batch balance wrapper during training [19]. Batch balance was also used to ensure that the samples in each batch were always balanced during the learning process (Appendix 1, Supplemental digital content 1, http://links.lww.com/NMC/A273).…”
Section: Methodsmentioning
confidence: 99%
“…Li et al [36] proposed a method called HM-loss cost, which pays more attention to the misclassifed samples in minority classes when calculating the loss. Hu et al [37] proposed a method of batch balancing datasets based on deep learning to ensure that the number of samples in each class is equal in each batch, so as to achieve the balance of datasets.…”
Section: Balanced Datasetmentioning
confidence: 99%