Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications 2019
DOI: 10.18653/v1/w19-4412
|View full text |Cite
|
Sign up to set email alerts
|

The Unreasonable Effectiveness of Transformer Language Models in Grammatical Error Correction

Abstract: Recent work on Grammatical Error Correction (GEC) has highlighted the importance of language modeling in that it is certainly possible to achieve good performance by comparing the probabilities of the proposed edits. At the same time, advancements in language modeling have managed to generate linguistic output, which is almost indistinguishable from that of human-generated text. In this paper, we up the ante by exploring the potential of more sophisticated language models in GEC and offer some key insights on … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
18
1
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(20 citation statements)
references
References 20 publications
0
18
1
1
Order By: Relevance
“…Recent work leverages the success of large pretrained language models to generate long texts such as stories (Rashkin et al, 2020), reviews (Cho et al, 2019a and fake news (Zellers et al, 2019). Most end-user applications for assisting user writing, however, are confined to sentence-level generation Kannan et al, 2016;Alikaniotis and Raheja, 2019;Prabhumoye et al, 2019;Faltings et al, 2021). Our work focuses on document-level writing assistance in which a document sketch is constructed from a set of similar documents.…”
Section: Document Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Recent work leverages the success of large pretrained language models to generate long texts such as stories (Rashkin et al, 2020), reviews (Cho et al, 2019a and fake news (Zellers et al, 2019). Most end-user applications for assisting user writing, however, are confined to sentence-level generation Kannan et al, 2016;Alikaniotis and Raheja, 2019;Prabhumoye et al, 2019;Faltings et al, 2021). Our work focuses on document-level writing assistance in which a document sketch is constructed from a set of similar documents.…”
Section: Document Generationmentioning
confidence: 99%
“…These risks have ensured that end-user applications involving text generation (e.g., Smart Compose, Smart Reply, Grammarly) still require a human to remain in control of content and are restricted to individual sentences or even smaller segments of text Figure 1: The right side shows a sketch for writing the report of a future democratic national convention, generated from a pile of previous reports. Kannan et al, 2016;Alikaniotis and Raheja, 2019;Prabhumoye et al, 2019;Faltings et al, 2021).…”
Section: Introductionmentioning
confidence: 99%
“…The authors use an n-gram language model, which we replace with GPT-2 (Radford et al, 2019) to see how a strong neural language model performsthis approach is similar to Alikaniotis and Raheja (2019). Hyperparameters are tuned for each dataset (see Appendix C for details).…”
Section: False Positive Examplesmentioning
confidence: 99%
“…There have been recent attempts to eliminate the time-consuming pre-training step by employing pre-trained transformer models. Alikaniotis and Raheja (2019) used pre-trained transformer models in a language-model setting, and fine-tuned BART (Lewis et al, 2020) with a small corpus of annotated sentences. Both approaches achieved results comparable to models trained with millions of sentences.…”
Section: Introductionmentioning
confidence: 99%