2010
DOI: 10.1007/s10994-010-5176-9
|View full text |Cite
|
Sign up to set email alerts
|

Preference-based learning to rank

Abstract: This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4

Citation Types

1
27
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(28 citation statements)
references
References 22 publications
1
27
0
Order By: Relevance
“…Active preference-based learning has been successfully used in many domains [1,6,7,14], but what makes applying it to learning reward functions difficult is the complexity of the queries, as well as the continuous nature of the underlying hypothesis space of possible reward functions. We focus on dynamical systems with continuous or hybrid discrete-continuous state.…”
Section: Introductionmentioning
confidence: 99%
“…Active preference-based learning has been successfully used in many domains [1,6,7,14], but what makes applying it to learning reward functions difficult is the complexity of the queries, as well as the continuous nature of the underlying hypothesis space of possible reward functions. We focus on dynamical systems with continuous or hybrid discrete-continuous state.…”
Section: Introductionmentioning
confidence: 99%
“…Preference-based models have been considered in various works [17]- [19]. In preference-based models, instead of learning the scoring function for each particular item, a preference function over pairs of items is learned in the training stage.…”
Section: Introductionmentioning
confidence: 99%
“…For instance, [18] proves that a simple approach [17], which is of time complexity quadratic to the number of test instances, is a 2-approximation of the optimal solution for a special ranking task called bipartite ranking. [19], [23] make improvements by using a quick-sort-like approach to achieve 3-approximation within sub-quadratic time complexity, and [24], [25] further achieve (1 + )-approximation within sub-quadratic time. Given the theoretical nature of the works, though many of them have discussed the possibility to employ preference-based LTR [19], [25], [26], few have yet to design algorithms that work well in practice or examine them properly on real-world data.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…There are many known regret reductions for such problems as multiclass classification [33,47], costsensitive classification [12,39], and ranking [2,3]. There is also a rich body of work on so called surrogate regret bounds.…”
mentioning
confidence: 99%