2021
DOI: 10.48550/arxiv.2109.12393
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sorting through the noise: Testing robustness of information processing in pre-trained language models

Abstract: Pre-trained LMs have shown impressive performance on downstream NLP tasks, but we have yet to establish a clear understanding of their sophistication when it comes to processing, retaining, and applying information presented in their input. In this paper we tackle a component of this question by examining robustness of models' ability to deploy relevant context information in the face of distracting content. We present models with cloze tasks requiring use of critical context information, and introduce distrac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 4 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?