Proceedings of the 21st Conference on Computational Natural Language Learning (CoNLL 2017) 2017
DOI: 10.18653/v1/k17-1045
|View full text |Cite
|
Sign up to set email alerts
|

Graph-based Neural Multi-Document Summarization

Abstract: We propose a neural multi-document summarization (MDS) system that incorporates sentence relation graphs. We employ a Graph Convolutional Network (GCN) on the relation graphs, with sentence embeddings obtained from Recurrent Neural Networks as input node features. Through multiple layer-wise propagation, the GCN generates high-level hidden sentence features for salience estimation. We then use a greedy heuristic to extract salient sentences while avoiding redundancy. In our experiments on DUC 2004, we consider… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
200
0

Year Published

2019
2019
2020
2020

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 209 publications
(200 citation statements)
references
References 30 publications
0
200
0
Order By: Relevance
“…A neural extractive summarizer learns to predict a binary label for each source sentence indicating if it is to be included in the summary. These studies build distributed sentence representations using neural networks (Cheng and Lapata, 2016;Yasunaga et al, 2017) and use reinforcement learning to optimize the evaluation metric (Narayan et al, 2018b) and improve summary coherence (Wu and Hu, 2018). However, sentence extraction can be coarse and in many cases, only a part of the sentence is worthy to be added to the summary.…”
Section: Related Workmentioning
confidence: 99%
“…A neural extractive summarizer learns to predict a binary label for each source sentence indicating if it is to be included in the summary. These studies build distributed sentence representations using neural networks (Cheng and Lapata, 2016;Yasunaga et al, 2017) and use reinforcement learning to optimize the evaluation metric (Narayan et al, 2018b) and improve summary coherence (Wu and Hu, 2018). However, sentence extraction can be coarse and in many cases, only a part of the sentence is worthy to be added to the summary.…”
Section: Related Workmentioning
confidence: 99%
“…All of these models are significantly underperforming compared to SemSentSum. In addition, we include state-ofthe-art models : RegSum (Hong and Nenkova, 2014) and GCN+PADG (Yasunaga et al, 2017). We outperform both in terms of ROUGE-1.…”
Section: Summarization Performancementioning
confidence: 99%
“…More specifically, extractive summarization systems output summaries in two steps : via sentence ranking, where an importance score is assigned to each sentence, and via the subsequent sentence selection, where the most appropriate sentence is chosen, by considering 1) their importance and 2) their frequency among all documents. Due to data sparcity, models heavily rely on well-designed features at the word level (Hong and Nenkova, 2014;Cao et al, 2015;Christensen et al, 2013;Yasunaga et al, 2017) or take advantage of other large, manually annotated datasets and then apply transfer learning (Cao et al, 2017). Additionally, most of the time, all sentences in the same collection of documents are processed independently and therefore, their relationships are lost.…”
Section: Introductionmentioning
confidence: 99%
“…For extractive methods, Nallapati et al [22] use recurrent neural networks (RNNs) to read the article and get the representations of the sentences and select important sentences. Yasunaga et al [23] combine RNNs with graph convolutional networks (CNNs) to compute the salience of each sentence. Narayan et al [24] propose a framework composed of a hierarchical encoder based on CNNs and an attention-based extractor with attention over external information.…”
Section: Related Workmentioning
confidence: 99%