“…Fine-tuning pre-trained language models like BERT (Devlin et al, 2019) has achieved great success in a wide variety of natural language processing (NLP) tasks, e.g., sentiment analysis (Abu Farha et al, 2021), question answering (Antoun et al, 2020), named entity recognition (Ghaddar et al, 2022), and dialect identification (Abdelali et al, 2021). Pre-trained LMs have also been used for enabling technologies such as part-of-speech (POS) tagging (Lan et al, 2020;Khalifa et al, 2021; to produce features for downstream processes. Previous POS tagging results using pre-trained LMs focused on core POS tagsets; however, it is still not clear how these models perform on the full morphosyntactic tagging task of very morphologically rich languages, where the size of the full tagset can be in the thousands.…”