2021
DOI: 10.48550/arxiv.2109.00062
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Shallow pooling for sparse labels

Abstract: Recent years have seen enormous gains in core information retrieval tasks, including document and passage ranking. Datasets and leaderboards, and in particular the MS MARCO datasets, illustrate the dramatic improvements achieved by modern neural rankers. When compared with traditional information retrieval test collections, such as those developed by TREC, the MS MARCO datasets employ substantially more queries -thousands vs. dozens -with substantially fewer known relevant items per query -often just one. For … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
15
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(16 citation statements)
references
References 16 publications
1
15
0
Order By: Relevance
“…Shifting our attention to MS MARCO NLGen, we notice that ELBoLoss still outperforms MarginalizedLoss by 1-2 points across our metrics, reflecting smaller-but nonetheless consistentgains when compared with the one-to-many generation task. We also observe that success@1 is lower for both methods, when compared with Wizard of Wikipedia, a fact we largely attributebased on manual inspection-to the presence of many false negatives, i.e., passages that contain the answer but aren't marked as gold passages, which is consistent with the findings from related studies Arabzadeh et al (2021). Overall, we find that ELBOLOSS improves relevance of retrieved passages over MARGINALIZEDLOSS for two qualitatively different tasks, with larger gains for the one-to-many generation task.…”
Section: Relevance Evaluationsupporting
confidence: 89%
“…Shifting our attention to MS MARCO NLGen, we notice that ELBoLoss still outperforms MarginalizedLoss by 1-2 points across our metrics, reflecting smaller-but nonetheless consistentgains when compared with the one-to-many generation task. We also observe that success@1 is lower for both methods, when compared with Wizard of Wikipedia, a fact we largely attributebased on manual inspection-to the presence of many false negatives, i.e., passages that contain the answer but aren't marked as gold passages, which is consistent with the findings from related studies Arabzadeh et al (2021). Overall, we find that ELBOLOSS improves relevance of retrieved passages over MARGINALIZEDLOSS for two qualitatively different tasks, with larger gains for the one-to-many generation task.…”
Section: Relevance Evaluationsupporting
confidence: 89%
“…The work we report here was carried out in the period May-August 2021, and was conceived and executed independently of and concurrently with the complementary work of Arabzadeh et al [1].…”
Section: Introductionmentioning
confidence: 99%
“…Reducing False Negatives False negatives are a common issue existing in the passage retrieval datasets (Qu et al, 2021;Arabzadeh et al, 2021), which denotes there are passages relevant to the query but labelled as negative. In this section, we discuss our strategy for reducing the false negatives in the development and testing sets of DuReader retrieval .…”
Section: Quality Improvementmentioning
confidence: 99%
“…• Arabzadeh et al (2021) and Qu et al (2021) observe that false negatives (i.e. relevant passages but labelled as negatives) are a common issue existing in the passage retrieval datasets due to their large scale but limited human annotation.…”
Section: Introductionmentioning
confidence: 99%