Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1318
|View full text |Cite
|
Sign up to set email alerts
|

Deep Copycat Networks for Text-to-Text Generation

Abstract: Most text-to-text generation tasks, for example text summarisation and text simplification, require copying words from the input to the output. We introduce copycat, a transformerbased pointer network for such tasks which obtains competitive results in abstractive text summarisation and generates more abstractive summaries. We propose a further extension of this architecture for automatic post-editing, where generation is conditioned over two inputs (source language and machine translation), and the model is c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(9 citation statements)
references
References 19 publications
0
9
0
Order By: Relevance
“…-like our Graformer -instead average attention weights over all heads. Other approaches compute their copy scores independently from encoder-decoder attention weights -either based on the output (Cai and Lam, 2020) or an intermediate representation of the last decoder layer (Ive et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…-like our Graformer -instead average attention weights over all heads. Other approaches compute their copy scores independently from encoder-decoder attention weights -either based on the output (Cai and Lam, 2020) or an intermediate representation of the last decoder layer (Ive et al, 2019).…”
Section: Related Workmentioning
confidence: 99%
“…In Table 5.2, the results of our systems and other state-of-the-art systems for the CNN/DailyMail corpus are shown 13 . It can be seen how our systems obtain better results than other widely used systems PGen+Cov [291], CopyCat [317] and SummaRunner [290]. The obtained results are worse in comparison to other extractive systems that use oracles, especially in the case of BertSumEXT [294], in spite of our systems share the same backbone architecture (transformer encoders).…”
Section: Discussionmentioning
confidence: 76%
“…Regarding the abstractive systems, all of them are based on encoder-decoder models with attention mechanisms [95,96,269,289,291,296,297,312,317]. First approaches suffered from known problems of the traditional sequence-to-sequence approaches: repetitions, grammatically incorrect generations, lack of coherence, coverage, hallucination (especially factual inconsistency), and the inability of producing words out of the training vocabulary [269,289,312].…”
Section: State Of the Artmentioning
confidence: 99%
See 2 more Smart Citations