“…Consequently, human-in-the-loop solutions [14,26,27,29,32,34,37,41,58], such as crowdsourcing, were developed to enable cheaper acquisition of labels from a large number of annotators in a short time. However, the increasing availability of human-labeled data may also yield variation in the reliability and proficiency of these human annotators [3,12,39,44,46,56], which, in turn, impair the quality of the generated labels. This variation may arise due to different expertise levels [12,46,56], mood changes [43], various cognitive biases [3,44] and more.…”