Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &Amp; Data Mining 2021
DOI: 10.1145/3447548.3467149
|View full text |Cite
|
Sign up to set email alerts
|

Pre-trained Language Model for Web-scale Retrieval in Baidu Search

Abstract: Search engine plays a crucial role in satisfying users' diverse information needs. Recently, Pretrained Language Models (PLMs) based text ranking models have achieved huge success in web search. However, many state-of-the-art text ranking approaches only focus on core relevance while ignoring other dimensions that contribute to user satisfaction, e.g., document quality, recency, authority, etc. In this work, we focus on ranking user satisfaction rather than relevance in web search, and propose a PLM-based fram… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4
1

Relationship

1
8

Authors

Journals

citations
Cited by 39 publications
(10 citation statements)
references
References 76 publications
0
10
0
Order By: Relevance
“…In more detail, there can be a stack of complex re-rankers after the efficient first-stage retriever. The multi-stage cascaded architecture is very common and practical both in the industry (Yin et al, 2016;Liu et al, 2021d;Li and Xu, 2014) and the ranking leaderboard in the academia (Craswell et al, 2021). Considering the large computational cost of Transformer-based pre-trained models, they are often employed to model the last stage re-ranker whose goal is to re-rank a small set of documents provided by previous stage.…”
Section: Pre-training Methods Applied In Re-ranking Componentmentioning
confidence: 99%
See 1 more Smart Citation
“…In more detail, there can be a stack of complex re-rankers after the efficient first-stage retriever. The multi-stage cascaded architecture is very common and practical both in the industry (Yin et al, 2016;Liu et al, 2021d;Li and Xu, 2014) and the ranking leaderboard in the academia (Craswell et al, 2021). Considering the large computational cost of Transformer-based pre-trained models, they are often employed to model the last stage re-ranker whose goal is to re-rank a small set of documents provided by previous stage.…”
Section: Pre-training Methods Applied In Re-ranking Componentmentioning
confidence: 99%
“…In this section, we introduce recent works designing PTMs tailored for IR (Lee et al, 2019b;Chang et al, 2019;Ma et al, 2021b;Ma et al, 2021c;Boualili et al, 2020;Ma et al, 2021d;Zou et al, 2021;Liu et al, 2021d). General pre-trained models like BERT have achieved great success when applied to IR tasks on both the firststage retrieval and the re-ranking stage.…”
Section: Keyphrase Extractionmentioning
confidence: 99%
“…However, the research on combining language models and user behavior data remains less developed. There are industrial works that fine-tune pretrained language models, for example based on BERT [33] or ERNIE [31,69], to produce representations for text contents and search queries. Yet these works are limited in terms of supported domains and tasks, i.e., they target a single use case such as web searches and concern only scoring tasks, with no intention to support diverse domains or generation tasks.…”
Section: Language Models As Foundationsmentioning
confidence: 99%
“…Embedding based retrieval has been widely applied in practice, such as search engines [30,39], question answering [33,34,54], online advertising [26,31], and content-based recommender systems [27,44]. Knowing that the documents need to be retrieved from a large-scale corpus, where brute-force linear scan will be temporally infeasible, it calls for approximate nearest neighbour search (ANN) [28] such that documents with high embedding similarities can be efficiently selected.…”
Section: Related Workmentioning
confidence: 99%