“…With the development of large pre-trained transformer models such as BERT and XLNET (Devlin et al, 2019;Yang et al, 2019), several studies have explored the use of general pre-trained transformers in offensive language identification (Liu et al, 2019;Bucur et al, 2021) as well retrained or fine-tuned models on offensive language corpora such as HateBERT (Caselli et al, 2020). While the vast majority of studies address offensive language identification using English data (Yao et al, 2019;Ridenhour et al, 2020), several recent studies have created new datasets for various languages and applied computational models to identify such content in Arabic (Mubarak et al, 2021), Dutch (Tulkens et al, 2016), French (Chiril et al, 2019), German (Wiegand et al, 2018), Greek (Pitenis et al, 2020), Hindi (Bohra et al, 2018), Italian (Poletto et al, 2017), Portuguese (Fortuna et al, 2019), Slovene (Fišer et al, 2017), Spanish (Plazadel Arco et al, 2021), and Turkish (C ¸öltekin, 2020. A recent trend is the use of pre-trained multilingual models such as XLM-R (Conneau et al, 2019) to leverage available English resources to make predictions in languages with less resources (Plaza-del Arco et al, 2021;Zampieri, 2020, 2021c,b;Sai and Sharma, 2021).…”