Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.418
|View full text |Cite
|
Sign up to set email alerts
|

Rethinking Self-Supervision Objectives for Generalizable Coherence Modeling

Abstract: Given the claims of improved text generation quality across various pre-trained neural models, we consider the coherence evaluation of machine generated text to be one of the principal applications of coherence models that needs to be investigated. Prior work in neural coherence modeling has primarily focused on devising new architectures for solving the permuted document task. We instead use a basic model architecture and show significant improvements over state of the art within the same training regime. We … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 30 publications
0
5
0
Order By: Relevance
“…Our analyses reveal that Likely-AI complaints were significantly different from Likely-Human complaints across various linguistic features. The first column of Table 2 shows that Likely-AI complaints were more coherent (i.e., show higher coherence scores) than Likely-Human complaints based on the recently developed coherence metric (Jwalapuram, Joty, and Lin, 2022). Moreover, as shown in the second column of Table 2, Likely-AI complaints were easier to read than Likely-Human complaints, as lower readability scores (Kincaid, Fishburne, Rogers, and Chissom, 1975) indicate texts that are easier to read.…”
Section: Observation 3: Linguistic Feature Improvementsmentioning
confidence: 98%
See 1 more Smart Citation
“…Our analyses reveal that Likely-AI complaints were significantly different from Likely-Human complaints across various linguistic features. The first column of Table 2 shows that Likely-AI complaints were more coherent (i.e., show higher coherence scores) than Likely-Human complaints based on the recently developed coherence metric (Jwalapuram, Joty, and Lin, 2022). Moreover, as shown in the second column of Table 2, Likely-AI complaints were easier to read than Likely-Human complaints, as lower readability scores (Kincaid, Fishburne, Rogers, and Chissom, 1975) indicate texts that are easier to read.…”
Section: Observation 3: Linguistic Feature Improvementsmentioning
confidence: 98%
“…We compared these two groups in terms of linguistic features proposed by prior computational linguistics research (Kincaid, Fishburne, Rogers, and Chissom, 1975;Jwalapuram, Joty, and Lin, 2022;Danescu-Niculescu-Mizil, Sudhof, Jurafsky, Leskovec, and Potts, 2013;Huang, Wang, and Yang, 2023). Each column in Table 2 presents results from a regression analysis in which the dependent variable was each computational linguistic metric estimated for the complaints and the independent variable was the binary variable of whether the complaint was Likely-AI (1) or Likely-Human (0) as in Table 1.…”
Section: Observation 3: Linguistic Feature Improvementsmentioning
confidence: 99%
“…On the other hand, the paragraph-level coherence score is determined by the normalized score of a discourse coherence model (Jwalapuram et al, 2021), denoted as S coh_para (k):…”
Section: Intrinsic Evaluationmentioning
confidence: 99%
“…text summarization, text classification, and text quality assessment. It is also utilized to model the coherence representation (Jwalapuram et al, 2022), because it takes advantage of both the autoregressive model and the BERT (Devlin et al, 2019). Formally, for a sequence x of the length I, there are I!…”
Section: Representation Learning With Xlnetmentioning
confidence: 99%
“…Document-Level Coherence with Contrastive Learning. Following Jwalapuram et al (2022), to learn robust coherence representations, we adopt the sentence ordering task by using contrastive learning. It enforces that the coherence score of the positive sample (original document) should be higher than that of the negative sample (disorder document).…”
Section: Coherence Modelingmentioning
confidence: 99%