Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.748
|View full text |Cite
|
Sign up to set email alerts
|

On Extractive and Abstractive Neural Document Summarization with Transformer Language Models

Abstract: We present a method to produce abstractive summaries of long documents that exceed several thousand words via neural abstractive summarization. We perform a simple extractive step before generating a summary, which is then used to condition the transformer language model on relevant information before being tasked with generating a summary. We also show that this approach produces more abstractive summaries compared to prior work that employs a copy mechanism while still achieving higher ROUGE scores. We provi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
92
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 119 publications
(93 citation statements)
references
References 27 publications
0
92
0
1
Order By: Relevance
“…By setting γ=0.0, our method is comparable to the attention-based method in Manakul and Gales (2020). By setting γ=1.0, our method is similar to the extractive models in Hsu et al (2018); Pilault et al (2020). In Table 4, we show that when coupled with BART, MCS yields better summarization performance than both Attn-only and Ext-only baselines.…”
Section: Multitask Content Selection (Mcs)mentioning
confidence: 84%
See 1 more Smart Citation
“…By setting γ=0.0, our method is comparable to the attention-based method in Manakul and Gales (2020). By setting γ=1.0, our method is similar to the extractive models in Hsu et al (2018); Pilault et al (2020). In Table 4, we show that when coupled with BART, MCS yields better summarization performance than both Attn-only and Ext-only baselines.…”
Section: Multitask Content Selection (Mcs)mentioning
confidence: 84%
“…Alternatively, earlier methods show that good content selection helps abstractive news summarization systems (Chen and Bansal, 2018;Gehrmann et al, 2018;Hsu et al, 2018). Hybrid systems that select sentences and generate an abstractive summary have been proposed such as extractive system + TLM for scientific articles (Pilault et al, 2020), simple selection + BART for podcasts (Manakul and Gales, 2020;Song et al, 2020), and guided summarization by BERT-based keyword/sentence extraction + BART for news and scientific articles (He et al, 2020;Dou et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Work in this area has mostly used used abstracts or peer reviews as targets (Cachola et al, 2020;Cohan et al, 2018;Jaidka et al, 2017). In particular, Pilault et al (2020) show that using a simple extractive summary as input for abstractive summarization of scholarly texts work well. Researchers have also used citing sentences as part of the input for summarization, recognizing the explanatory power of these texts (Nakov et al, 2004;Cohan and Goharian, 2017;Yasunaga et al, 2019).…”
Section: Related Workmentioning
confidence: 96%
“…How abstractive are the summaries? Abstractive summarizers generate surprisingly extractive summaries, copying large fragments unmodified from the input documents into the summaries (Weber et al, 2018;Pilault et al, 2020). We hypothesize that providing graph representations of the input can help the model abstract away from the specific lexical content of the input and generate summaries that are more abstractive.…”
Section: Ablations and Analysesmentioning
confidence: 99%