Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1297
|View full text |Cite
|
Sign up to set email alerts
|

Neural Sequence-Labelling Models for Grammatical Error Correction

Abstract: We propose an approach to N -best list reranking using neural sequence-labelling models. We train a compositional model for error detection that calculates the probability of each token in a sentence being correct or incorrect, utilising the full sentence as context. Using the error detection model, we then re-rank the N best hypotheses generated by statistical machine translation systems. Our approach achieves state-of-the-art results on error correction for three different datasets, and it has the additional… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 41 publications
(37 citation statements)
references
References 30 publications
0
37
0
Order By: Relevance
“…Models can also be partially initialized by pre-training monolingual language models (Ramachandran et al, 2017) or only word-embeddings (Gangi and Federico, 2017). In GEC, Yannakoudakis et al (2017) apply pretrained monolingual word-embeddings as initializations for error-detection models to re-rank SMT n-best lists. Approaches based on pre-training with monolingual data appear to be particularly wellsuited to the GEC task.…”
Section: Transfer Learning For Gecmentioning
confidence: 99%
See 2 more Smart Citations
“…Models can also be partially initialized by pre-training monolingual language models (Ramachandran et al, 2017) or only word-embeddings (Gangi and Federico, 2017). In GEC, Yannakoudakis et al (2017) apply pretrained monolingual word-embeddings as initializations for error-detection models to re-rank SMT n-best lists. Approaches based on pre-training with monolingual data appear to be particularly wellsuited to the GEC task.…”
Section: Transfer Learning For Gecmentioning
confidence: 99%
“…For the CoNLL 2014 benchmark on grammatical error correction , Junczys-Dowmunt and Grundkiewicz (2016) established a set of methods for GEC by SMT that remain state-of-the-art. Systems (Chollampatt and Ng, 2017;Yannakoudakis et al, 2017) that improve on results by Junczys-Dowmunt and Grundkiewicz (2016) use their set-up as a backbone for more complex systems.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…This improved the scores by about 0.3 F 0.5 in CoNLL-2014 and FCE-test Table 3: Our LM-based approach is compared against several state-of-the-art results. AMU16 SM T +LSTM and CAMB16 SM T +LSTM were both originally reported by Yannakoudakis et al (2017), while Lee and Lee (2014) is the system entered by POST in CoNLL-2014. Only our approach does not use annotated training data.…”
Section: Resultsmentioning
confidence: 99%
“…These approaches have since come to dominate the field, and a lot of recent research has focused on fine-tuning SMT systems (JunczysDowmunt and Grundkiewicz, 2016), reranking SMT output , combining SMT and classifier systems Rozovskaya and Roth, 2016), and developing various neural architectures Xie et al, 2016;Chollampatt and Ng, 2017;Yannakoudakis et al, 2017).…”
Section: Introductionmentioning
confidence: 99%