“…In terms of models , pre-trained multilingual word embeddings (MUSE ( Conneau et al, 2017 )) and language models (mBERT ( Devlin et al, 2019 ), XLM-R ( Conneau et al, 2020 )) are frequently chosen as baselines. They are an intuitive and easily accessible starting point for cross-lingual experiments, but their limitations are also clear—the “curse of multilinguality” trades off single-language performance for its broad language coverage, as displayed in the results of the cross-lingual generalisation studies mentioned above ( Pamungkas & Patti, 2019 ; Pamungkas, Basile & Patti, 2020 ; Glavaš, Karan & Vulić, 2020 ; Arango, Prez & Poblete, 2020 ; Fortuna, Soler-Company & Wanner, 2021 ) and in other tasks ( Conneau et al, 2020 ). Similarly to the monolingual case, there are cases where traditional machine learning models outperform deep learning ones, such as SVM ( Pamungkas, Basile & Patti, 2020 ) and GBDT ( Arango, Prez & Poblete, 2020 ) compared to LSTM.…”