“…Recent works have witnessed prominent performances of multilingual pre-trained language models (PrLMs) (Devlin et al, 2019;Conneau et al, 2020) on cross-lingual tasks, including machine translation (Lin et al, 2020;Chen et al, 2021), semantic role labeling (SRL) (Conia and Navigli, 2020;Conia et al, 2021) and semantic parsing (Fei et al, 2020b;Sherborne and Lapata, 2021). How-ever, cross-lingual CSRL, as a combination of three challenging tasks (i.e., cross-lingual task, dialogue task and SRL task), suffers three outstanding difficulties: 1) latent space alignment -how to map word representations of different languages into an overlapping space; 2) conversation structure encoding -how to capture high-level dialogue features such as speaker dependency and temporal dependency; and 3) semantic arguments identification -how to highlight the relations between the predicate and its arguments, wherein PrLMs can only partially encode multilingual inputs to an overlapping vector space.…”