Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-tutorials.1
|View full text |Cite
|
Sign up to set email alerts
|

Text Generation with Text-Editing Models

Abstract: Text-editing models have recently become a prominent alternative to seq2seq models for monolingual text-generation tasks such as grammatical error correction, simplification, and style transfer. These tasks share a common trait -they exhibit a large amount of textual overlap between the source and target texts. Text-editing models take advantage of this observation and learn to generate the output by predicting edit operations applied to the source sequence. In contrast, seq2seq models generate outputs word-by… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 22 publications
0
3
0
Order By: Relevance
“…For directly predicting the target corrections from corresponding input tokens, Omelianchuk et al (2020) and Malmi et al (2022) regarded the encoder of the transformer model as a nonautoregressive GEC sequence tagger. The experimental results of Omelianchuk et al (2020) showed that, compared with the randomly initialized LSTM (Hochreiter and Schmidhuber 1997), the pre-trained models, such as RoBERTa (Liu et al 2019), GPT-2 (Radford et al 2019), and ALBERT (Lan et al 2020), can achieve higher F 0.5 scores as a tagger.…”
Section: Related Workmentioning
confidence: 99%
“…For directly predicting the target corrections from corresponding input tokens, Omelianchuk et al (2020) and Malmi et al (2022) regarded the encoder of the transformer model as a nonautoregressive GEC sequence tagger. The experimental results of Omelianchuk et al (2020) showed that, compared with the randomly initialized LSTM (Hochreiter and Schmidhuber 1997), the pre-trained models, such as RoBERTa (Liu et al 2019), GPT-2 (Radford et al 2019), and ALBERT (Lan et al 2020), can achieve higher F 0.5 scores as a tagger.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, a line of work has emerged examining how large language models (LLMs) can serve as collaborative writing/coding assistants. Because of their remarkable ability to generate coherent texts over a wide range of domains and topics, LLMs have proven surprisingly effective for editing, elaboration, infilling, etc., across a wide range of domains (Malmi et al, 2022;Bavarian et al, 2022;Donahue et al, 2020). Though our system also makes use of LLMs, it supports a different mode of editing than these prior works.…”
Section: Background and Related Workmentioning
confidence: 99%
“…This paper studies a multiagent collaborative framework where one language model can generate critiques to improve its peer's performance. 2020; Malmi et al, 2022), grammatical (Lichtarge et al, 2019) or factual error correction (Mitchell et al, 2022b), debiasing and detoxification (Schick et al, 2021). Unlike humans who can understand natural language feedback and improve using the information, most of the previous work relied on sequence tagging (Reid and Neubig, 2022), retraining from scratch (Sun et al, 2019) or parameter editing (Mitchell et al, 2022a) to repair model predictions.…”
Section: Introductionmentioning
confidence: 99%