Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval 2021
DOI: 10.1145/3404835.3463093
|View full text |Cite
|
Sign up to set email alerts
|

A Systematic Evaluation of Transfer Learning and Pseudo-labeling with BERT-based Ranking Models

Abstract: Due to high annotation costs, making the best use of existing human-created training data is an important research direction. We, therefore, carry out a systematic evaluation of transferability of BERT-based neural ranking models across five English datasets. Previous studies focused primarily on zero-shot and few-shot transfer from a large dataset to a dataset with a small number of queries. In contrast, each of our collections has a substantial number of queries, which enables a full-shot evaluation mode and… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(20 citation statements)
references
References 30 publications
2
18
0
Order By: Relevance
“…In Table 2, we report the BEIR results for our two CoCondenser scenarios, reaching state-of-the-art results on zero-shot evaluation, strengthening the observation that sparse retrieval models seem to be better able to generalize [5,17,26]. We also report in Table 3 the results of the two previous approaches combined with BM25 (sum); additional gains can be obtained, showing that pure lexical approaches are still somehow complementary to sparse neural models, especially in a zero-shot setting.…”
Section: Modelsupporting
confidence: 62%
“…In Table 2, we report the BEIR results for our two CoCondenser scenarios, reaching state-of-the-art results on zero-shot evaluation, strengthening the observation that sparse retrieval models seem to be better able to generalize [5,17,26]. We also report in Table 3 the results of the two previous approaches combined with BM25 (sum); additional gains can be obtained, showing that pure lexical approaches are still somehow complementary to sparse neural models, especially in a zero-shot setting.…”
Section: Modelsupporting
confidence: 62%
“…Humor (Forough and Momtazi, 2021) , label smoothing (Müller et al, 2019), and pseudo labeling (Mokrii et al, 2021). However, we argue that SAM is the main module to attain the exceptional performance.…”
Section: Methodsmentioning
confidence: 67%
“…There is limited evidence regarding the use of machine learning to score free‐text medical professionalism questions 3 . With the development of modern machine learning techniques, such as Bidirectional Encoder Representations from Transformers (BERT), it is possible that machine learning may be able to assist with this task 8–10 …”
Section: Number Of Marks Model Professionalism Mean Classification Ac...mentioning
confidence: 99%