Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval 2021
DOI: 10.1145/3404835.3463262
|View full text |Cite
|
Sign up to set email alerts
|

How Deep is your Learning: the DL-HARD Annotated Deep Learning Dataset

Abstract: Deep Learning Hard (DL-HARD) is a new annotated dataset designed to more effectively evaluate neural ranking models on complex topics. It builds on TREC Deep Learning (DL) topics by extensively annotating them with question intent categories, answer types, wikified entities, topic categories, and result type metadata from a commercial web search engine. Based on this data, we introduce a framework for identifying challenging queries. DL-HARD contains fifty topics from the official DL 2019/2020 evaluation bench… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 26 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…This suggests that location queries are over-represented in MS MARCO, and that, in general, the queries that are discarded come from a different distribution than those that survive. This distribution shift is potentially concerning, given that others have found description queries to be among the more challenging and interesting of the query types [18].…”
Section: Analysis Of Discarded Queriesmentioning
confidence: 99%
“…This suggests that location queries are over-represented in MS MARCO, and that, in general, the queries that are discarded come from a different distribution than those that survive. This distribution shift is potentially concerning, given that others have found description queries to be among the more challenging and interesting of the query types [18].…”
Section: Analysis Of Discarded Queriesmentioning
confidence: 99%
“…those that "require interpretation [...] beyond the immediate meaning of the terms" [1] in the queries). Coincidentally, Mackie et al [25] introduced a framework for identifying hard queries that challenge DRs. Sciavolino et al [34] found that DRs also perform poorly, and worse than traditional bag-of-words methods, on queries that contain entities.…”
Section: Related Workmentioning
confidence: 99%
“…However, MS MARCO's annotation technique means that the queries tend to be artificially easy and exhibit undesirable qualities like the 'maximum passage bias' [20]. TREC Deep Learning [8] extends MS MARCO with dense judgments to provide a more useful benchmark, and DL-HARD [26] develops a more challenging subset with annotations and metadata. CODEC differs from these datasets in terms of length of queries, i.e.…”
Section: Document Rankingmentioning
confidence: 99%