“…Self-supervised learning (SSL) is a method for building models where the output labels are already included in the input data, eliminating the need for additional labeled data (Liu et al, 2021;Hu et al, 2021a,b;Liu et al, 2022d,c). SSL has been widely used in NLP domains such as sentence generation (West et al, 2019;Yan et al, 2021), document processing (You et al, 2021;Ginzburg et al, 2021), natural language inference (Li et al, , 2023, and text reasoning (Klein and Nabi, 2020;Fu et al, 2020;Chen et al, 2022). BERT (Devlin et al, 2019) is one of the most eminent SSL methods which exploit self-supervisions from corpus with next sentence prediction and masked language modeling tasks.…”