“…(2) Self-training is to first train a model with manually labeled data, then use the model to automatically label unlabeled data, and finally leverage the manually and automatically labeled data to enhance itself (Xie et al, 2019(Xie et al, , 2020. It shows promising results in many SpanID tasks, including NER (Wang et al, 2020), propaganda detection (Hou et al, 2021) (Seo et al, 2016;Chen et al, 2017), while recent trends have shown great advantages of formulating NLP tasks as MRC problems. In the con- (Li et al, 2019a), event detection ), and summarization (McCann et al, 2018 are also reported to benefit from the MRC paradigm.…”