2019
DOI: 10.48550/arxiv.1903.01182
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Complement Objective Training

Abstract: Learning with a primary objective, such as softmax cross entropy for classification and sequence generation, has been the norm for training deep neural networks for years. Although being a widely-adopted approach, using cross entropy as the primary objective exploits mostly the information from the ground-truth class for maximizing data likelihood, and largely ignores information from the complement (incorrect) classes. We argue that, in addition to the primary objective, training also using a complement objec… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(7 citation statements)
references
References 14 publications
0
7
0
Order By: Relevance
“…Existing Methods. We compare the proposed method (CCE) with the following techniques: (i) empirical risk minimization (ERM): we train models with only standard softmax cross entropy; (ii) complement objective training (COT): there are two optimizers for training: one for softmax cross entropy and the other one for softmax complement entropy [8]; (iii) focal loss (FL): it uses sigmoid cross entropy with a modulating factor to concentrate on hard samples [24].…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Existing Methods. We compare the proposed method (CCE) with the following techniques: (i) empirical risk minimization (ERM): we train models with only standard softmax cross entropy; (ii) complement objective training (COT): there are two optimizers for training: one for softmax cross entropy and the other one for softmax complement entropy [8]; (iii) focal loss (FL): it uses sigmoid cross entropy with a modulating factor to concentrate on hard samples [24].…”
Section: Methodsmentioning
confidence: 99%
“…It means that ̂ ( ) [ ≠ ] in the cross entropy is totally ignored, so inaccurately predicted probabilities may produce a cumulative error. To avoid such error, complement objective training (COT) was proposed by Chen et al, where the core idea is evenly suppressing softmax probabilities on incorrect classes during training [8].…”
Section: Introductionmentioning
confidence: 99%
“…Differently, our PC loss with logit constraints is motivated by the predictive behavior of CNN on adversarial samples from the probability perspective, which avoids this issue by learning probabilistically compact features without geometric assumptions. The work that is closest to ours is (Chen et al 2019b) that encourages the predicted probabilities of false classes to be equally distributed, whereas our PC loss directly enlarges the gap of probabilities between true class and the first several most probable false classes.…”
Section: Related Workmentioning
confidence: 96%
“…In addition, Chen proposed complement entropy loss [32], which suppresses misclassification and achieves error neutralization by exploiting the complement information of samples in this class, thus improving the confidence level. Because the minority class is richer in error information, Kim et al [33] used complement entropy to improve the class imbalance problem.…”
Section: Manuscript Clear Copymentioning
confidence: 99%