Proceedings of the Canadian Conference on Artificial Intelligence 2021
DOI: 10.21428/594757db.da1a3d44
|View full text |Cite
|
Sign up to set email alerts
|

Class-wise Calibration: A Case Study on COVID-19 Hate Speech

Abstract: Proper calibration of deep-learning models is critical for many high-stakes problems. In this paper, we show that existing calibration metrics fail to pay attention to miscalibration on individual classes, hence overlooking minority classes and causing significant issues on imbalanced classification problems. Using a COVID-19 hate-speech dataset, we first discover that in imbalanced datasets, miscalibration error on an individual class varies greatly, and error on minority classes can be magnitude times worse … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 20 publications
0
5
0
Order By: Relevance
“…Unlike vector and matrix scaling, RD-TS cannot change the relative ranking of logits, and therefore model accuracy is retained (in single-label settings). One line of future work could be to apply RD-TS on top of weighted temperature scaling, a method known to decrease variance in calibration error among classes (Obadinma et al, 2021). Another line of work would be to investigate whether improved certainty estimates can increase model accuracy (in multi-label settings where predictions are applied by meeting a certainty threshold), especially in out-of-domain problems.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Unlike vector and matrix scaling, RD-TS cannot change the relative ranking of logits, and therefore model accuracy is retained (in single-label settings). One line of future work could be to apply RD-TS on top of weighted temperature scaling, a method known to decrease variance in calibration error among classes (Obadinma et al, 2021). Another line of work would be to investigate whether improved certainty estimates can increase model accuracy (in multi-label settings where predictions are applied by meeting a certainty threshold), especially in out-of-domain problems.…”
Section: Discussionmentioning
confidence: 99%
“…Weighted temperature scaling (WTS): TS using a class-weighted NLL loss during convergence (Obadinma et al, 2021).…”
Section: Matrix Scaling (Ms)mentioning
confidence: 99%
“…To prevent the model trained on a long-tailed dataset from being skewed to high-frequency classes, Islam et al proposed a class distribution-aware temperature scaling, which uses class frequency information to get the temperature value. Similarly, Weighted temperature scaling is proposed by Qkadinma et al [149], which re-weights the loss function by the inverse class count to tune the scaling parameter.…”
Section: E Parametric Methodsmentioning
confidence: 99%
“…T ransf er Knowledge f rom Head to T ail [16] ⃝ ✕ ⃝ △ ✕ ⃝ ⃝ Scipy 20 Class-wise Loss Scaling [150] △ ✕ △ △ △ ⃝ ⃝ PyTorch 21 Region-dependent T emperature Scaling [143] ⃝ ✕ ⃝ △ ✕ ⃝ ⃝ URL invalid 22 Class-wise T emperature Scaling [144] ⃝ ✕ △ △ △ ⃝ ⃝ N/A N ormalized Calibration [17] △ △ △ ⃝ ✕ ⃝ △ PyTorch 23 Gaussian and Gamma Calibration [151] ⃝ ✕ △ ⃝ ✕ ✕ ⃝ URL invalid 24 CS-T S-AT C [145] △ ✕ △ △ △ ⃝ ⃝ PyTorch 25 Dual-Branch T emperature Scaling [148] △ ✕ △ △ △ ⃝ ⃝ N/A Class-distribution-aware T S [21] △ ✕ △ ⃝ △ ⃝ ⃝ URL invalid 14 W eighted T emperature Scaling [149] △…”
Section: Parametric Methods (Section Iv-e)mentioning
confidence: 99%
See 1 more Smart Citation