Proceedings of the Sixth ACM International Conference on Web Search and Data Mining 2013
DOI: 10.1145/2433396.2433419
|View full text |Cite
|
Sign up to set email alerts
|

Reusing historical interaction data for faster online learning to rank for IR

Abstract: We summarize the findings from Hofmann et al. [6]. Online learning to rank for information retrieval (IR) holds promise for allowing the development of "self-learning" search engines that can automatically adjust to their users. With the large amount of e.g., click data that can be collected in web search settings, such techniques could enable highly scalable ranking optimization. However, feedback obtained from user interactions is noisy, and developing approaches that can learn from this feedback quickly and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
109
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
6
2

Relationship

4
4

Authors

Journals

citations
Cited by 94 publications
(109 citation statements)
references
References 24 publications
0
109
0
Order By: Relevance
“…The advantage of P-MGD is that it can learn faster by having n, the number of candidates, in Algorithm 1 exceed the length of the result list. Candidate Preselection (CPS) [5], unlike MGD, does not alter the number of candidates compared per impression. It speeds up learning by reusing historical data to select…”
Section: Online Learning To Rank Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The advantage of P-MGD is that it can learn faster by having n, the number of candidates, in Algorithm 1 exceed the length of the result list. Candidate Preselection (CPS) [5], unlike MGD, does not alter the number of candidates compared per impression. It speeds up learning by reusing historical data to select…”
Section: Online Learning To Rank Methodsmentioning
confidence: 99%
“…We show experimentally that P-MGD significantly outperforms state-of-the-art online learning to rank methods in terms of online performance, without sacrificing offline performance and at greater learning speed than those methods. In particular, we include comparisons between P-MGD on the one hand and multiple types of DBGD and multileaved gradient descent methods [14,MGD] and candidate preselection [5,CPS] on the other. We answer the following research questions: (RQ1) Does P-MGD convergence on a ranker of the same quality as MGD and CPS?…”
Section: Introductionmentioning
confidence: 99%
“…DBGD implements a stochastic gradient descent method to find the best ranker in an infinite space of rankers. This algorithm has been extended before by Hofmann et al [11] such that it would reuse historical interaction data and not just live user interactions. Alternative methods are based on the k-armed bandit formulation by [32] and assume a finite space of possible rankers, the best of which needs to be found.…”
Section: Online Learning To Rankmentioning
confidence: 99%
“…A major advantage of PI is that it can also infer preferences between rankers that were not originally interleaved. This allows one to learn from historical interaction data [11]. However, PI risks showing users poor rankings.…”
Section: Evaluation For Information Retrievalmentioning
confidence: 99%
“…• More advanced methods that build on DBGD such as Probabilistic Multileave Gradient Decent [23,27] and DBGD with Candidate Preselection [12].…”
Section: Part I [10 Minutes] Introduction Aims and Historical Notesmentioning
confidence: 99%