“…Within traditional machine learning, NLI is usually approached as a multi-class classification problem of assigning class labels representing L1s to texts written in L2, where the main focus is to design features that capture the systematic fingerprints of the first language in the second language writing (native language interference (Odlin, 1989)). These features include: spelling errors (Koppel et al, 2005;Chen et al, 2017); lexical features, e.g., word and lemma n-grams (Jarvis et al, 2013), cognates (Markov et al, 2019), etymologically-related words ; syntactic features, e.g., context-free grammar features (Wong and Dras, 2011), Stanford parser dependency features (Tetreault et al, 2012); stylometric features, e.g., punctuation (Markov et al, 2018a), character n-gram features (Kulmizev et al, 2017); emotion-based features (Markov et al, 2018b), etc. The combination of such features provides the best results for NLI, as shown by the two shared tasks organized in 2013 and 2017 (Malmasi et al, 2017), where the two top-ranked systems (Cimino and Dell'Orletta, 2017;Markov et al, 2017) used Support Vector Machines (SVM) with a variety of engineered features.…”