Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.494
|View full text |Cite
|
Sign up to set email alerts
|

ReasonBERT: Pre-trained to Reason with Distant Supervision

Abstract: We present ReasonBERT, a pre-training method that augments language models with the ability to reason over long-range relations and multiple, possibly hybrid, contexts. Unlike existing pre-training methods that only harvest learning signals from local contexts of naturally occurring texts, we propose a generalized notion of distant supervision to automatically connect multiple pieces of text and tables to create pre-training examples that require long-range reasoning. Different types of reasoning are simulated… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 23 publications
0
12
0
Order By: Relevance
“…It consists of various sub-skills including commonsense reasoning (Zellers et al, 2018;Talmor et al, 2019;Bhagavatula et al, 2019), numerical reasoning (Dua et al, 2019), arithmetic reasoning (Koncel-Kedziorski et al, 2015;Roy and Roth, 2016;Miao et al, 2020;Cobbe et al, 2021), logical reasoning (Yu et al, 2020), tabular reasoning (Zhu et al, 2021), and so on. Previous efforts in machine learning exploited symbolic systems (Mihaylov and Frank, 2018;Ding et al, 2019;Wang et al, 2022b,a) and pre-training strategies (Deng et al, 2021;Asai and Hajishirzi, 2020;Pi et al, 2022). Recently, large language models with chain-of-thought prompting (Wei et al, 2022b;Wang et al, 2022c;Zhou et al, 2022;Zhang et al, 2022b) demonstrate promising reasoning abilities with appropriately designed prompts, achieving competitive performance on several benchmarks.…”
Section: Reasoning Abilitymentioning
confidence: 99%
“…It consists of various sub-skills including commonsense reasoning (Zellers et al, 2018;Talmor et al, 2019;Bhagavatula et al, 2019), numerical reasoning (Dua et al, 2019), arithmetic reasoning (Koncel-Kedziorski et al, 2015;Roy and Roth, 2016;Miao et al, 2020;Cobbe et al, 2021), logical reasoning (Yu et al, 2020), tabular reasoning (Zhu et al, 2021), and so on. Previous efforts in machine learning exploited symbolic systems (Mihaylov and Frank, 2018;Ding et al, 2019;Wang et al, 2022b,a) and pre-training strategies (Deng et al, 2021;Asai and Hajishirzi, 2020;Pi et al, 2022). Recently, large language models with chain-of-thought prompting (Wei et al, 2022b;Wang et al, 2022c;Zhou et al, 2022;Zhang et al, 2022b) demonstrate promising reasoning abilities with appropriately designed prompts, achieving competitive performance on several benchmarks.…”
Section: Reasoning Abilitymentioning
confidence: 99%
“…For example, 1. Smarter span selection: We only consider spans selected uniformly at random for generality, but mixing in semantically or syntactically meaningful spans [Donahue et al, 2020, Joshi et al, 2020, Deng et al, 2021 can considerably improve infilling performance. In Section 4.5, we see that training on line-level spans instead of character-level spans improves line-based infilling results.…”
Section: Future Directionsmentioning
confidence: 99%
“…(2) The latter is task-oriented, which includes several well designed pre-training or fine-tuning tasks (Jiao et al, 2021;Deng et al, 2021). However, none of these works target cross-lingual areas, especially for xSL tasks.…”
Section: Related Workmentioning
confidence: 99%