Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.111
|View full text |Cite
|
Sign up to set email alerts
|

FELIX: Flexible Text Editing Through Tagging and Insertion

Abstract: We present FELIX -a flexible text-editing approach for generation, designed to derive maximum benefit from the ideas of decoding with bi-directional contexts and self-supervised pretraining. In contrast to conventional sequenceto-sequence (seq2seq) models, FELIX is efficient in low-resource settings and fast at inference time, while being capable of modeling flexible input-output transformations. We achieve this by decomposing the text-editing task into two sub-tasks: tagging to decide on the subset of input t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
85
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 43 publications
(85 citation statements)
references
References 35 publications
0
85
0
Order By: Relevance
“…Pointer-generator style models (See et al, 2017;Xu et al, 2020) can accurately generate mostly extractive summaries by copying words from the source text via pointing. Text editing models Dong et al, 2019b;Mallinson et al, 2020) cast text generation as a sequence tagging problem with carefully selected edit operations required for the task. Others focus on improving content selection to better constrain the model to likely input phrases (Gehrmann et al, 2018) or by improving the representation of relevant input tokens .…”
Section: Related Workmentioning
confidence: 99%
“…Pointer-generator style models (See et al, 2017;Xu et al, 2020) can accurately generate mostly extractive summaries by copying words from the source text via pointing. Text editing models Dong et al, 2019b;Mallinson et al, 2020) cast text generation as a sequence tagging problem with carefully selected edit operations required for the task. Others focus on improving content selection to better constrain the model to likely input phrases (Gehrmann et al, 2018) or by improving the representation of relevant input tokens .…”
Section: Related Workmentioning
confidence: 99%
“…To model the length, it is possible to use an autoregressive decoder or a separate model (Mansimov et al, 2019). Instead, we use an efficient non-autoregressive padded MLM approach by Mallinson et al (2020) which enables BERT to predict [PAD] symbols when infilling a fixed-length spans of n p [MASK] tokens.…”
Section: Padded Masked Language Modelsmentioning
confidence: 99%
“…Text-editing methods (Dong et al, 2019;Awasthi et al, 2019;Mallinson et al, 2020), that target monolingual sequence transduction tasks like sentence fusion, grammar correction, and text simplification, are typically more dataefficient than the traditional sequence-to-sequence methods, but they still require substantial amounts of parallel training examples to work well. When parallel source-target training pairs are difficult to obtain, it is often still possible to collect nonparallel examples for the source and the target domain separately.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…We hypothesize that BERT2BERT's strategy (Botha et al, 2018). 32.31 Dong et al (2019) 34.94 Xu et al (2016) 37.94 Mallinson et al (2020) 38.13 This work (no tags)…”
Section: Sentence Fusionmentioning
confidence: 99%