2020
DOI: 10.1162/tacl_a_00342
|View full text |Cite
|
Sign up to set email alerts
|

oLMpics-On What Language Model Pre-training Captures

Abstract: Recent success of pre-trained language models (LMs) has spurred widespread interest in the language capabilities that they possess. However, efforts to understand whether LM representations are useful for symbolic reasoning tasks have been limited and scattered. In this work, we propose eight reasoning tasks, which conceptually require operations such as comparison, conjunction, and composition. A fundamental challenge is to understand whether the performance of a LM on a task should be attributed to the pre-t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
97
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 177 publications
(98 citation statements)
references
References 36 publications
1
97
0
Order By: Relevance
“…These works cover a wide variety of topics, including grammatical generalization(Goldberg, 2019;Warstadt et al, 2019), syntax(Tenney et al, 2019b;Lin et al, 2019;Reif et al, 2019;Hewitt and Manning, 2019;Liu et al, 2019a), world knowledge(Petroni et al, 2019;Jiang et al, 2020), reasoning(Talmor et al, 2019), and common sense(Forbes et al, 2019;Zhou et al, 2019;Weir et al, 2020). 169 Downloaded from http://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00359/1894330/tacl_a_00359.pdf by guest on 10 May 2021…”
mentioning
confidence: 99%
“…These works cover a wide variety of topics, including grammatical generalization(Goldberg, 2019;Warstadt et al, 2019), syntax(Tenney et al, 2019b;Lin et al, 2019;Reif et al, 2019;Hewitt and Manning, 2019;Liu et al, 2019a), world knowledge(Petroni et al, 2019;Jiang et al, 2020), reasoning(Talmor et al, 2019), and common sense(Forbes et al, 2019;Zhou et al, 2019;Weir et al, 2020). 169 Downloaded from http://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00359/1894330/tacl_a_00359.pdf by guest on 10 May 2021…”
mentioning
confidence: 99%
“…Understanding what information NLP models encode has attracted great interest in recent years (Rogers et al, 2020). From factual (Petroni et al, 2019;Jawahar et al, 2019;Roberts et al, 2020) to linguistic (Conneau et al, 2018;Liu et al, 2019a;Talmor et al, 2019) and commonsense (Forbes et al, 2019) knowledge, a wide set of properties have been previously analysed. We refer to Belinkov and Glass (2019) and Rogers et al (2020) for a more comprehensive literature review.…”
Section: Related Workmentioning
confidence: 99%
“…If a model predicts the correct attribute in the top-5 position, then we infer that the model representations have captured the corresponding affect association. Additionally, to understand the behavior after fine-tuning, we introduce MLP with a 1-hidden layer to the MLM setup to train the LMs as discussed in (Talmor et al, 2019); we call this setup MLP-MLM.…”
Section: Probing Tasksmentioning
confidence: 99%
“…Dataset and the BCD datasets (RQ1). Further, to understand the impact of fine-tuning, we employ techniques proposed by (Talmor et al, 2019) to measure the language mismatch. In this exercise, we fine-tune the LM with examples from Ads.…”
Section: Impact Of Fine-tuningmentioning
confidence: 99%
See 1 more Smart Citation