Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP 2020
DOI: 10.18653/v1/2020.blackboxnlp-1.3
|View full text |Cite
|
Sign up to set email alerts
|

Examining the rhetorical capacities of neural language models

Abstract: Recently, neural language models (LMs) have demonstrated impressive abilities in generating high-quality discourse. While many recent papers have analyzed the syntactic aspects encoded in LMs, to date, there has been no analysis of the inter-sentential, rhetorical knowledge. In this paper, we propose a method that quantitatively evaluates the rhetorical capacities of neural LMs. We examine the capacities of neural LMs understanding the rhetoric of discourse by evaluating their abilities to encode a set of ling… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
9
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 39 publications
0
9
0
Order By: Relevance
“…At the base of our work are two of the most popular and frequently used PLMs: BERT (Devlin et al, 2018) and BART (Lewis et al, 2020). We choose these two popular approaches in our study due to their complementary nature (encoder-only vs. encoder-decoder) and based on previous work by Zhu et al (2020) and Koto et al (2021a), showing the effectiveness of BERT and BART models for discourse related tasks.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…At the base of our work are two of the most popular and frequently used PLMs: BERT (Devlin et al, 2018) and BART (Lewis et al, 2020). We choose these two popular approaches in our study due to their complementary nature (encoder-only vs. encoder-decoder) and based on previous work by Zhu et al (2020) and Koto et al (2021a), showing the effectiveness of BERT and BART models for discourse related tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Compared to our work, their tree inference approach is however computationally expensive and only explores the outputs of the BERT model. Further, Zhu et al (2020) use 24 hand-crafted rhetorical features to execute three different supervised probing tasks, showing promising performance of the BERT model. Similarly, Pandia et al (2021) aim to infer pragmatics through the prediction of discourse connectives by analyzing the model inputs and outputs and Koto et al (2021a) analyze discourse in PLMs through seven supervised probing tasks, finding that BART and BERT contain Looking at all these prior works analyzing the amount of discourse in PLMs, structures are solely explored through the use of proxy tasks, such as connective prediction (Pandia et al, 2021), relation classification (Kurfalı and Östling, 2021) and others (Koto et al, 2021a).…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Upadhye et al (2020) analyzed how well two pretrained models capture referential biases of different classes of English verbs. Zhu et al (2020) applied the model of Feng and Hirst (2014) to parse IMDB documents (Maas et al, 2011) into discourse trees. Using this (potentially noisy) data, probing tasks were conducted by mapping attention layers into single vectors of document-level rhetorical features.…”
Section: Introductionmentioning
confidence: 99%