2020
DOI: 10.1609/aaai.v34i05.6389
|View full text |Cite
|
Sign up to set email alerts
|

Lexical Simplification with Pretrained Encoders

Abstract: Lexical simplification (LS) aims to replace complex words in a given sentence with their simpler alternatives of equivalent meaning. Recently unsupervised lexical simplification approaches only rely on the complex word itself regardless of the given sentence to generate candidate substitutions, which will inevitably produce a large number of spurious candidates. We present a simple LS approach that makes use of the Bidirectional Encoder Representations from Transformers (BERT) which can consider both the given… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
57
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 62 publications
(57 citation statements)
references
References 12 publications
0
57
0
Order By: Relevance
“…Our approach to the MWLS problem, called the Plainifier, is an extension of the unsupervised method for single word lexical simplification by Qiang et al (2020). Following their work, we generate candidate replacements using BERT predictions for a given context and rank them according to language-model probability, simplicity and similarity of meaning to the original text.…”
Section: Simplification Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Our approach to the MWLS problem, called the Plainifier, is an extension of the unsupervised method for single word lexical simplification by Qiang et al (2020). Following their work, we generate candidate replacements using BERT predictions for a given context and rank them according to language-model probability, simplicity and similarity of meaning to the original text.…”
Section: Simplification Methodsmentioning
confidence: 99%
“…In step (a), a gap created by removing the replaced token is filled with a single [MASK], for which the predictions are acquired from Terse-BERT. This method is used by Qiang et al (2020) to obtain one-word candidates, such as The cat sleeps on the mat, and their likelihoods. The multi-words setting in Plainifer requires step (b), in which the gap is filled with two [MASK] elements and the best (according to ranking described in section 4.3) K predictions for the first position are obtained.…”
Section: Candidate Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…They also tend to change the meaning of the sentence and have problems dealing with ambiguous words [30] [31] [32]. However, recently, unsupervised approaches have been improved in this regard by allowing more detailed context information to be obtained [33]. On the other hand, hybrid strategies employ methods from both the previous two approaches, such as [34], which uses a corpus-based approach and a combination of a free lexicon, decision trees, and context-based rules.…”
Section: Nlp Approaches To Lexical Simplificationmentioning
confidence: 99%
“…1 (Petersen and Ostendorf 2007) ( 3 http://www.dianamccarthy.co.uk/task10index.html 4 https://www.cs.york.ac.uk/semeval-2012/task1/ 5 http://people.cs.kuleuven.be/˜jan.debelder/lseval.zip 6 https://simple.wikipedia.org/wiki/Wikipedia:Basic English combined wordlist 7 https://cs.pomona.edu/˜dkauchak/simplification/ 1 (Zhu et al 2010) PWKP / WikiSmall 108K Wikipedia (Coster and Kauchak 2011) 137K Wikipedia (Xu, Callison-Burch, and Napoles 2015) Newsela 96K (Hwang, Hajishirzi, Ostendorf, and Wu 2015) 392K Wikipedia (Kajiwara and Komachi 2016) 493K Wikipedia (Zhang and Lapata 2017) WikiLarge 286K Wikipedia (Maruyama and Yamamoto 2018) SNOW T15 50K (Katsuta and Yamamoto 2018) SNOW T23 35K (Paetzold and Specia 2017b;Qiang, Li, Zhu, Yuan, and Wu 2020) LSeval (3)…”
mentioning
confidence: 99%