2021
DOI: 10.48550/arxiv.2109.09209
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

CLIFF: Contrastive Learning for Improving Faithfulness and Factuality in Abstractive Summarization

Abstract: We study generating abstractive summaries that are faithful and factually consistent with the given articles. A novel contrastive learning formulation is presented, which leverages both reference summaries, as positive training data, and automatically generated erroneous summaries, as negative training data, to train summarization systems that are better at distinguishing between them. We further design four types of strategies for creating negative samples, to resemble errors made commonly by two state-of-the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 43 publications
0
7
0
Order By: Relevance
“…Entity Replacing for Hard Negative: We choose T5 (Raffel et al, 2019) to generate the most similar word to head or tail entity, and then replace the head or tail entity with its augmented word to obtain the hard negative instance, which possesses the similar entity and context to original instance. Entity Swap for Semi-Hard Negative: To con-struct a semi-hard negative instance for the original instance, we follow the setting of Entity Swap (Cao and Wang, 2021), which swaps the target entities with other randomly selected entities of the same entity type in the original instance.…”
Section: Data Augmentationmentioning
confidence: 99%
“…Entity Replacing for Hard Negative: We choose T5 (Raffel et al, 2019) to generate the most similar word to head or tail entity, and then replace the head or tail entity with its augmented word to obtain the hard negative instance, which possesses the similar entity and context to original instance. Entity Swap for Semi-Hard Negative: To con-struct a semi-hard negative instance for the original instance, we follow the setting of Entity Swap (Cao and Wang, 2021), which swaps the target entities with other randomly selected entities of the same entity type in the original instance.…”
Section: Data Augmentationmentioning
confidence: 99%
“…We will introduce effective approaches that use alternative supervision sources for IE, that is, to use supervision signals from related tasks to make up for the lack of quantity and comprehensiveness in IE-specific training data. This includes indirect supervision sources such as question answering and reading comprehension Lyu et al, 2021;Levy et al, 2017;Li et al, 2019b;Du and Cardie, 2020), natural language inference (Li et al, 2022a;Yin et al, 2020) and generation (Lu et al, 2021;. We will also cover the use of weak supervision sources such as structural texts (e.g., Wikipedia) (Ji et al, 2017;Zhou et al, 2018) and global biases (Ning et al, 2018b).…”
Section: Minimally and Indirectly Supervised Ie [35min]mentioning
confidence: 99%
“…Shu et al (2021) propose to improve logicto-text generation models by designing rule-based data augmentation to create contrastive examples to cover variations of logic forms paired with diverse natural language expressions to improve the generalizability. CLIFF (Cao and Wang, 2021) propose to improve faithful and factual consistency for abstractive summarization by contrasting reference summaries as positive training data and automatically generated erroneous summaries as negative training data. Wu et al (2020a) also propose to use contrastive learning for unsupervised referencefree summary quality evaluation.…”
Section: Contrastive Learning For Nlpmentioning
confidence: 99%
See 1 more Smart Citation
“…To improve the robustness of the graph structure, refs. [ 26 , 27 , 28 , 29 , 30 ] introduced contrast learning in the model for pulling positive samples close to push away negative samples. Most of the encoders of existing models are based on sentences or words with low reliability, and the joint summary model proposed in [ 31 ] improves the performance of summary generation.…”
Section: Introductionmentioning
confidence: 99%