“…Similarly to how long existing stereotypes are deep-rooted in word embeddings (Papakyriakopoulos et al, 2020;Garg et al, 2018), PTLMs have also been shown to recreate stereotypical content due to the nature of their training data (Sheng et al, 2019) Different probing experiments have been proposed to study the drawbacks of PTLMs in areas such as the biomedical domain (Jin et al, 2019), syntax (Hewitt and Manning, 2019;Marvin and Linzen, 2018), semantic and syntactic sentence structures (Tenney et al, 2019), prenomial anaphora (Sorodoc et al, 2020), common-sense (Petroni et al, 2019), gender bias (Kurita et al, 2019), and typicality in judgement (Misra et al, 2021). Except for Hutchinson et al (2020) who examine what words BERT generate in some fill-in-the-blank experiments with regard to people with disabilities, and more recently Nozza et al (2019) who assess hurtful auto-completion by multilingual PTLMs, we are not aware of other strategies designed to estimate toxic content in PTLMs with regard to several social groups.…”