2005
DOI: 10.1145/1105696.1105699
|View full text |Cite
|
Sign up to set email alerts
|

Flexible pseudo-relevance feedback via selective sampling

Abstract: Although Pseudo-Relevance Feedback (PRF) is a widely used technique for enhancing average retrieval performance, it may actually hurt performance for around one-third of a given set of topics. To enhance the reliability of PRF, Flexible PRF has been proposed, which adjusts the number of pseudo-relevant documents and/or the number of expansion terms for each topic. This paper explores a new, inexpensive Flexible PRF method, called Selective Sampling, which is unique in that it can skip documents in the initial … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
28
0

Year Published

2008
2008
2020
2020

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 61 publications
(28 citation statements)
references
References 15 publications
0
28
0
Order By: Relevance
“…A selective sampling method by Sakai et al [22] skips some top-retrieved documents based on a clustering criterion. The cluster is generated not by document similarity but by the same set of query terms.…”
Section: Related Workmentioning
confidence: 99%
“…A selective sampling method by Sakai et al [22] skips some top-retrieved documents based on a clustering criterion. The cluster is generated not by document similarity but by the same set of query terms.…”
Section: Related Workmentioning
confidence: 99%
“…The Robustness Index (−1 ≤ RI(q) ≤ 1), also called Reliability of Improvement Index, of a model with respect to a baseline was formulated by Sakai et al in [18] as in Eq 12:…”
Section: Training and Evaluationmentioning
confidence: 99%
“…Some research even questions the usefulness of BRF in general [9]. Many approaches have been proposed to increase the overall IR performance of BRF, for example by adapting the number of feedback terms and documents per topic [10], by selecting only good feedback terms [11,12], or by increasing diversity of terms in pseudo-relevant documents by skipping feedback documents [13]. TREC experiments with BRF use conservative settings for the number of feedback terms and documents (see [7,14]) using less than 10 documents and 10-30 feedback terms to obtain the best IR effectiveness.…”
Section: Related Workmentioning
confidence: 99%