2006
DOI: 10.1109/icdm.2006.22
|View full text |Cite
|
Sign up to set email alerts
|

Applying Data Mining to Pseudo-Relevance Feedback for High Performance Text Retrieval

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 34 publications
(19 citation statements)
references
References 21 publications
0
19
0
Order By: Relevance
“…It assumes that topranked documents in the first-pass retrieval are relevant, and then used as feedback documents in order to refine the representation of original queries by adding potentially related terms. Although PRF has been shown to be effective in improving IR performance [4,6,9,13,23,26,28,30,36,37,40,42] in a number of IR tasks, traditional PRF can also fail in some cases. For example, when some of the feedback documents have several incoherent topics, terms in the irrelevant contents are likely to misguide the feedback models by importing noisy terms into the queries.…”
Section: Introductionmentioning
confidence: 98%
“…It assumes that topranked documents in the first-pass retrieval are relevant, and then used as feedback documents in order to refine the representation of original queries by adding potentially related terms. Although PRF has been shown to be effective in improving IR performance [4,6,9,13,23,26,28,30,36,37,40,42] in a number of IR tasks, traditional PRF can also fail in some cases. For example, when some of the feedback documents have several incoherent topics, terms in the irrelevant contents are likely to misguide the feedback models by importing noisy terms into the queries.…”
Section: Introductionmentioning
confidence: 98%
“…Most of the work on PRF used the top ranked documents or passages [Ruthven and Lalmas 2003;Xu and Croft 2000] as pseudo-relevant documents (or positive examples). Some researchers chose to include also pseudo-nonrelevant documents (or negative examples), by using the bottom-ranked documents [Huang et al 2006;Yan et al 2003], while others found no improvement in doing so [Buckley and Robertson 2008;Kaptein et al 2008]. We chose to use both positive and negative examples, as the initial query representation includes irrelevant concepts to be removed (for which we believe negative examples will be useful), in addition to missing relevant concepts (for which the positive examples alone are sufficient).…”
Section: Feature Selection Using Pseudo-relevance Feedbackmentioning
confidence: 99%
“…Request permissions from permissions@acm.org. ous studies have shown the benefits brought by transductive learning for IR [24,14,13,9,17,26,31]. Most of these previous approaches assume the top-ranked documents to be highly relevant to the given query, and is closely related to the query topic.…”
Section: Introductionmentioning
confidence: 99%