2017
DOI: 10.1609/aaai.v31i1.10894
|View full text |Cite
|
Sign up to set email alerts
|

Robust Loss Functions under Label Noise for Deep Neural Networks

Abstract: In many applications of classifier learning, training data suffers from label noise. Deep networks are learned using huge training data where the problem of noisy labels is particularly relevant. The current techniques proposed for learning deep networks under label noise focus on modifying the network architecture and on algorithms for estimating true labels from noisy labels. An alternate approach would be to look for loss functions that are inherently noise-tolerant. For binary classification there exist th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
158
0
2

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 591 publications
(161 citation statements)
references
References 9 publications
1
158
0
2
Order By: Relevance
“…In machine learning, such mismatch between labels and predictive data is called label noise (Frénay and Verleysen, 2013) and is actively studied (Ghosh et al, 2017;Lee et al, 2018). The strength of our dataset in counteracting this noise is its very large size, as demonstrated by Rolnick et al (2017).…”
Section: On Temporal and Spatial Biasesmentioning
confidence: 99%
“…In machine learning, such mismatch between labels and predictive data is called label noise (Frénay and Verleysen, 2013) and is actively studied (Ghosh et al, 2017;Lee et al, 2018). The strength of our dataset in counteracting this noise is its very large size, as demonstrated by Rolnick et al (2017).…”
Section: On Temporal and Spatial Biasesmentioning
confidence: 99%
“…Major existing robust losses include: mean absolute error (MAE), shown to be theoretically robust in [10], but hard to train in [54]; generalised cross-entropy (GCE) which attempts to be robust yet easy to train [54]; bi-temper [1], two-temperature [2], and Huber [15] motivated by heavy tailed outlier robustness; symmetric cross-entropy [45] motivated by reducing overfitting; and active-passive loss (APL) [31] which aims to balance over-and-underfitting for robust losses. These losses are all hand-designed based on various good motivations, but (as we will see in our evaluation) none provide reliably high performance empirically.…”
Section: Learning With Label Noisementioning
confidence: 99%
“…We compare our ARL with the standard cross-entropy (CE) baseline, as well as several strong alternative losses hand-designed for label-noise robustness: MAE: Mean Absolute Error was theoretically shown to be robust in [10]. GCE: [54] analysed MAE as hard to train, and proposed generalised cross-entropy to provide the best of CE and MAE; FW: [33] iteratively estimates the label noise transfer matrix, and trains the model corrected by the label noise estimate; SCE: [45] argued that symmetrising cross-entropy by adding reverse cross-entropy (RCE) improves label-noise robustness; Bootstrap: A classic method of replacing the noisy labels in training by the convex combination of the prediction and the given labels [37].…”
Section: Competitorsmentioning
confidence: 99%
See 2 more Smart Citations