2020
DOI: 10.48550/arxiv.2010.11506
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Calibrated Language Model Fine-Tuning for In- and Out-of-Distribution Data

Abstract: Fine-tuned pre-trained language models can suffer from severe miscalibration for both in-distribution and out-of-distribution (OOD) data due to over-parameterization. To mitigate this issue, we propose a regularized fine-tuning method. Our method introduces two types of regularization for better calibration: (1) On-manifold regularization, which generates pseudo on-manifold samples through interpolation within the data manifold. Augmented training with these pseudo samples imposes a smoothness regularization t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
11
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(11 citation statements)
references
References 24 publications
0
11
0
Order By: Relevance
“…Kong et al [25] encourage the model to output uniform distributions on pseudo off-manifold samples to alleviate the over-confidence in OOD regions. On the contrary, we apply off-manifold samples by enforcing the model to predict high vacuity:…”
Section: Utilizing Off-manifold Samplesmentioning
confidence: 99%
See 4 more Smart Citations
“…Kong et al [25] encourage the model to output uniform distributions on pseudo off-manifold samples to alleviate the over-confidence in OOD regions. On the contrary, we apply off-manifold samples by enforcing the model to predict high vacuity:…”
Section: Utilizing Off-manifold Samplesmentioning
confidence: 99%
“…The embeddings of pre-trained transformers contain rich features that benefit the generated adversarial examples. Thus following [25], we evaluate off-manifold regularization on BERT [8].…”
Section: Utilizing Off-manifold Samplesmentioning
confidence: 99%
See 3 more Smart Citations