Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.294
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised Reference-Free Summary Quality Evaluation via Contrastive Learning

Abstract: Evaluation of a document summarization system has been a critical factor to impact the success of the summarization task. Previous approaches, such as ROUGE, mainly consider the informativeness of the assessed summary and require human-generated references for each test summary. In this work, we propose to evaluate the summary qualities without reference summaries by unsupervised contrastive learning. Specifically, we design a new metric which covers both linguistic qualities and semantic informativeness based… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 32 publications
(10 citation statements)
references
References 22 publications
0
10
0
Order By: Relevance
“…C) Reference-free Approach Rather than measuring the quality of candidate summary based on a ground truth summary, reference-free metrics for relevance measure the quality of candidate summary based on pseudo-reference summaries that are generated from source documents. Wu et al [117]'s proposed metric requires training samples of high-quality summaries for model supervision while other reference-free metrics can generate metric scores without the use of high-quality summaries as supervisory signals [14,35]. In section 3 of benchmark datasets, we see that the information covered by a reference summary depends on the data annotation approach as well as the intent of the original authors.…”
Section: A) Hard Lexical Overlapmentioning
confidence: 99%
“…C) Reference-free Approach Rather than measuring the quality of candidate summary based on a ground truth summary, reference-free metrics for relevance measure the quality of candidate summary based on pseudo-reference summaries that are generated from source documents. Wu et al [117]'s proposed metric requires training samples of high-quality summaries for model supervision while other reference-free metrics can generate metric scores without the use of high-quality summaries as supervisory signals [14,35]. In section 3 of benchmark datasets, we see that the information covered by a reference summary depends on the data annotation approach as well as the intent of the original authors.…”
Section: A) Hard Lexical Overlapmentioning
confidence: 99%
“…More recently, MatchSum (Zhong et al 2020) formulates extractive summarization as a semantic text matching problem using contrastive learning. Wu et al (2020) measures the summary qualities without reference summaries by contrasting the document with the summaries using a ranking model. GSum (Dou et al 2021) takes different kinds of external guidance as additional input to the document and advances summarization performance significantly.…”
Section: Related Workmentioning
confidence: 99%
“…(Shi et al, 2019) randomly replaces a sentence in the ground-truth summary with a random sentence to form the negative sample. Wu et al (2020) constructs negative samples on different aspects of summary qualities and propose a new summary evaluation method by contrastive learning. Zhong et al (2020) use a pre-trained extractive model to select several candidates as negative samples and take the groundtruth as the positive.…”
Section: Contrastive Learning In Summarizationmentioning
confidence: 99%