Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval 2015
DOI: 10.1145/2766462.2767754
|View full text |Cite
|
Sign up to set email alerts
|

Impact of Surrogate Assessments on High-Recall Retrieval

Abstract: We are concerned with the effect of using a surrogate assessor to train a passive (i.e., batch) supervised-learning method to rank documents for subsequent review, where the effectiveness of the ranking will be evaluated using a different assessor deemed to be authoritative. Previous studies suggest that surrogate assessments may be a reasonable proxy for authoritative assessments for this task. Nonetheless, concern persists in some application domains-such as * The views expressed herein are solely those of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 13 publications
(12 citation statements)
references
References 28 publications
0
12
0
Order By: Relevance
“…Its development was informed by experience from TREC Total Recall tracks [24,49] and also adopted by the CLEF Technology Assisted Reviews in Empirical Medicine Tracks [30][31][32]. The π‘™π‘œπ‘ π‘  π‘’π‘Ÿ measure is the sum of two components:…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 2 more Smart Citations
“…Its development was informed by experience from TREC Total Recall tracks [24,49] and also adopted by the CLEF Technology Assisted Reviews in Empirical Medicine Tracks [30][31][32]. The π‘™π‘œπ‘ π‘  π‘’π‘Ÿ measure is the sum of two components:…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…π‘Ÿπ‘’π‘π‘Žπ‘™π‘™ 𝑑 ), but we choose not to adjust the method to simplify comparison with previous work and because the relative error measure already captures information about the difference between the achieved and target recall. The second component, π‘™π‘œπ‘ π‘  𝑒 , motivated by experience from the TREC 2015 Total Recall Track [49], is defined as:…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…Traditionally, relevance feedback has been construed as a method to automate the query-formulation task envisioned by TREC: The user's assessment of the results from an initial query are provided to the search tool, which reformulates the query and presents a new set of results to the user, and so on. More recently, supervised machine-learning methods have been used to harness relevance feedback, with reported effectiveness apparently exceeding Voorhees' "practical upper bound," see e.g., [6,14,25].…”
Section: Theorymentioning
confidence: 99%
“…Previously published results report recall-precision breakeven scores on the order of 80% for BMI and related methods [7,8,11,25]. With one notable exception (discussed below), these results were derived using simulated feedback from an assumed-infallible user, and evaluated with respect to the same assumed-infallible gold standard.…”
Section: Prediction and Rationalementioning
confidence: 99%