Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval 2022
DOI: 10.1145/3477495.3531772
|View full text |Cite
|
Sign up to set email alerts
|

Pre-train a Discriminative Text Encoder for Dense Retrieval via Contrastive Span Prediction

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 18 publications
(6 citation statements)
references
References 20 publications
0
6
0
Order By: Relevance
“…In experiments, we conduct extensive evaluations of our method on the two long-document retrieval benchmark datasets: MS-Marco Doc (Nguyen et al 2016) and TREC Deep Learning 2019 document retrieval (TREC 2019) (Craswell et al 2020). Following previous works (Ma et al 2022), we use official metrics MRR@100 and Recall@100 (R@100) to report evaluation results on MS-Marco dev, while using nDCG@10 and Recall@100 for TREC 2019.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…In experiments, we conduct extensive evaluations of our method on the two long-document retrieval benchmark datasets: MS-Marco Doc (Nguyen et al 2016) and TREC Deep Learning 2019 document retrieval (TREC 2019) (Craswell et al 2020). Following previous works (Ma et al 2022), we use official metrics MRR@100 and Recall@100 (R@100) to report evaluation results on MS-Marco dev, while using nDCG@10 and Recall@100 for TREC 2019.…”
Section: Methodsmentioning
confidence: 99%
“…In particular, to produce consistent embeddings, previous methods directly apply mean-pooling over contextual embeddings for different granularity, which however becomes inferior when the document length goes extremely long and has proven less effective in our pilot experiments. This is the reason why most previous document retrieval works rely on [CLS] embedding paradigm (Ma et al 2022;Xiong et al 2021;Zhan et al 2021b).…”
Section: Global-consistent Granularity Embeddingmentioning
confidence: 99%
See 2 more Smart Citations
“…Pre-trained Bi-encoder. Researchers have explored pre-training models for retrieval with the bi-encoder architecture [6,7,14,18,20,30]. For example, Gao and Callan [6] added extra head layers atop the Transformer, with shortcut connections between early outputs and the head, enhancing the CLS embedding of the encoder.…”
Section: Related Workmentioning
confidence: 99%