Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning 2016
DOI: 10.18653/v1/k16-1028
|View full text |Cite
|
Sign up to set email alerts
|

Abstractive Text Summarization using Sequence-to-sequence RNNs and Beyond

Abstract: In this work, we model abstractive text summarization using Attentional EncoderDecoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora. We propose several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling key-words, capturing the hierarchy of sentence-toword structure, and emitting words that are rare or unseen at training time. Our work shows that many of our propo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
1,609
3
6

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 1,810 publications
(1,621 citation statements)
references
References 20 publications
3
1,609
3
6
Order By: Relevance
“…7 Preliminary experiments training the models proposed by Rush et al (2015) and Nallapati et al (2016) on our dataset have been promising: by manual inspection of individual samples, they produce useful summaries for many Reddit posts; we leave a quantitative evaluation for future work.…”
Section: Resultsmentioning
confidence: 99%
“…7 Preliminary experiments training the models proposed by Rush et al (2015) and Nallapati et al (2016) on our dataset have been promising: by manual inspection of individual samples, they produce useful summaries for many Reddit posts; we leave a quantitative evaluation for future work.…”
Section: Resultsmentioning
confidence: 99%
“…In Table 4 we compare our model with the abstractive attentional encoder-decoder models in ( Nallapati et al, 2016), which leverage several effective techniques and achieve state-of-the-art performance on sentence abstractive summarization tasks. The words-lvt2k and words-lvt2k-ptr are flat models and words-lvt2k-hieratt is a hierarchical extension.…”
Section: Discussionmentioning
confidence: 99%
“…Cheng and Lapata (2016) also adopt a word extraction model, which is restricted to use the words of the source document to generate a summary, although the performance is much worse than the sentence extractive model. Nallapati et al (2016) extend the sentence summarization model by trying a hierarchical attention architecture and a limited vocabulary during the decoding phase. However these models still investigate few properties of the document summarization task.…”
Section: Abstractive Summarization Methodsmentioning
confidence: 99%
“…In the fields of NLP, the popular Seq2Seq [16] model is based on RNN/LSTM, in which a multi-layer of an LSTM network is used as an encoder and another multi-layer of the LSTM network is used as a decoder. This kind of Seq2Seq model has many variants and has been applied in many applications, such as machine translation [17], text summary generation [18], and Chinese poetry generation [19], among others.…”
Section: Recurrent Neural Network In Nlpmentioning
confidence: 99%