Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.71
|View full text |Cite
|
Sign up to set email alerts
|

Towards Interpreting and Mitigating Shortcut Learning Behavior of NLU models

Abstract: Recent studies indicate that NLU models are prone to rely on shortcut features for prediction, without achieving true language understanding. As a result, these models fail to generalize to real-world out-of-distribution data. In this work, we show that the words in the NLU training set can be modeled as a longtailed distribution. There are two findings: 1) NLU models have strong preference for features located at the head of the long-tailed distribution, and 2) Shortcut features are picked up during very earl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2021
2021
2025
2025

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 40 publications
(35 citation statements)
references
References 40 publications
0
35
0
Order By: Relevance
“…This suggests that naively applying debiasing techniques may incur unexpected negative impacts on other aspects of the moderation system. Further research is needed into modeling approaches that can achieve robust performance both in prediction and in uncertainty calibration under data bias and distributional shift (Nam et al, 2020;Utama et al, 2020;Du et al, 2021;Yaghoobzadeh et al, 2021;Bao et al, 2021;Karimi Mahabadi et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…This suggests that naively applying debiasing techniques may incur unexpected negative impacts on other aspects of the moderation system. Further research is needed into modeling approaches that can achieve robust performance both in prediction and in uncertainty calibration under data bias and distributional shift (Nam et al, 2020;Utama et al, 2020;Du et al, 2021;Yaghoobzadeh et al, 2021;Bao et al, 2021;Karimi Mahabadi et al, 2020).…”
Section: Discussionmentioning
confidence: 99%
“…UP only requires the pre-trained model to restore the order of ut- terances in noisy dialogues rather than predicting new words. Such NLU-style task would make it easier for models to learn shortcuts, i.e., models tend to exploit dataset biases as shortcuts to make predictions, rather than learn the semantic understanding and reasoning (Geirhos et al, 2020;Du et al, 2021). For ablation studies of combining two or three tasks, please refer to Appendix C.…”
Section: Ablation Studymentioning
confidence: 99%
“…Alternatively, one can make strong assumptions about what a ground truth should be like and compare salience rankings with what is expected to be the ground truth. In this vein human reasoning (Poerner et al, 2018;, gradient information (Du et al, 2021), aggregated model internal representations (Atanasova et al, 2020), changes in predicted probabilities (DeYoung et al, 2020; or surrogate models (Ding and Koehn, 2021) all have been taken as a proxy for the ground truth when evaluating salience methods. Unfortunately, they also resulted in divergent recommendations so the question of what the ground truth is and which method to use remains open.…”
Section: Compute Rankingsmentioning
confidence: 99%