“…Our hypothesis is that, although each language is unique, different languages manifest similar characteristics (e.g., morphological, lexical, syntactic) which can be exploited by training a single model with data from multiple languages (Ammar, 2016). Previous work has shown this to be true to some degree in the context of syntactic dependency parsing , semantic role labeling (Mulcaire et al, 2018), named entity recognition (Xie et al, 2018), and language modeling for phonetic sequences (Tsvetkov et al, 2016) and for speech recognition (Ragni et al, 2016). Recently, de Lhoneux et al (2018) showed that parameter sharing between languages can improve performance in dependency parsing, but the effect is variable, depending on the language pair and the parameter sharing strategy.…”