2018
DOI: 10.1109/tnnls.2017.2732482
|View full text |Cite
|
Sign up to set email alerts
|

Cost-Sensitive Learning of Deep Feature Representations From Imbalanced Data

Abstract: Abstract-Class imbalance is a common problem in the case of real-world object detection and classification tasks. Data of some classes is abundant making them an over-represented majority, and data of other classes is scarce, making them an underrepresented minority. This imbalance makes it challenging for a classifier to appropriately learn the discriminating boundaries of the majority and minority classes. In this work, we propose a cost-sensitive deep neural network which can automatically learn robust feat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
83
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 687 publications
(83 citation statements)
references
References 46 publications
0
83
0
Order By: Relevance
“…If the training set is generated by randomly sampling from the complete dataset, it is likely that two-thirds of those samples are steady-state class. This class imbalance problem has been well documented in the literature [55][56][57]. Failure cases tend to be considerably less abundant than steady wear cases.…”
Section: Experiments and Resultsmentioning
confidence: 93%
“…If the training set is generated by randomly sampling from the complete dataset, it is likely that two-thirds of those samples are steady-state class. This class imbalance problem has been well documented in the literature [55][56][57]. Failure cases tend to be considerably less abundant than steady wear cases.…”
Section: Experiments and Resultsmentioning
confidence: 93%
“…The loss function acts an important role in deep convolutional tracker by solving the data imbalance problem [ 26 ], though little attention had been paid to this kind of issue [ 27 ]. So far, the cost-sensitive loss [ 28 ] is proven to be an effective approach when suffering data imbalance. When pre-training the Siamese networks, Bertinetto et al [ 29 ] proposed to balance the loss of positives and negatives in order to improve the discriminative ability of the network.…”
Section: Related Workmentioning
confidence: 99%
“…Nonetheless, over-sampling increases the probability of overfitting, and down-sampling risks from losing useful information. Algorithm-level methods focus on weighting the training samples, including loss-based methods and cost-sensitive methods [ 25 , 26 , 27 , 28 , 29 , 30 , 31 ]. For example, Lin et al [ 26 ] proposed focal loss to weight the hard samples.…”
Section: Related Workmentioning
confidence: 99%
“…In References [ 15 , 16 ], data imbalance is addressed with the cost-sensitive loss and Area Under the receiver operating characteristics Curve (AUC) statistics, which is directly derived from the statistic information of the input dataset. In References [ 20 , 21 , 22 , 23 , 24 , 25 , 26 , 27 , 28 , 29 , 30 , 31 , 32 , 33 , 34 , 35 , 36 ], data-level methods, algorithm-level methods, and hybrid methods were employed to address the data imbalance, boosting the performance to some extent. Hybrid methods inherit the advantages of data-level and algorithm-level methods [ 20 ].…”
Section: Introductionmentioning
confidence: 99%