“…Conventional cross-lingual summarization methods mainly focus on incorporating bilingual information into the pipeline methods (Leuski et al, 2003;Ouyang et al, 2019;Orȃsan and Chiorean, 2008;Wan et al, 2010;Wan, 2011;Yao et al, 2015;Zhang et al, 2016b), i.e., translation and then summarization or summarization and then translation. Due to the difficulty of acquiring cross-lingual summarization dataset, some previous researches focus on constructing datasets (Ladhak et al, 2020;Scialom et al, 2020;Yela-Bello et al, 2021;Zhu et al, 2019;Hasan et al, 2021;Perez-Beltrachini and Lapata, 2021;Varab and Schluter, 2021), mixed-lingual pre-training (Xu et al, 2020), knowledge distillation (Nguyen and Tuan, 2021), contrastive learning (Wang et al, 2021) or zero-shot approaches (Ayana et al, 2018;Duan et al, 2019;Dou et al, 2020), i.e., using machine translation (MT) or monolingual summarization (MS) or both to train the CLS system. Among them, Zhu et al (2019) propose to use roundtrip translation strategy to obtain large-scale CLS datasets and then present two multi-task learning methods for CLS.…”