Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1352
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Domain Modeling of Sentence-Level Evidence for Document Retrieval

Abstract: This paper applies BERT to ad hoc document retrieval on news articles, which requires addressing two challenges: relevance judgments in existing test collections are typically provided only at the document level, and documents often exceed the length that BERT was designed to handle. Our solution is to aggregate sentence-level evidence to rank documents. Furthermore, we are able to leverage passage-level relevance judgments fortuitously available in other domains to fine-tune BERT models that are able to captu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
138
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 113 publications
(141 citation statements)
references
References 23 publications
3
138
0
Order By: Relevance
“…The T5-3B results in bold are significantly better (p < 0.05) than T5-large, T5-base, and the corresponding baseline (BM25 or BM25+RM3), based on the Student's paired t-test with Bonferroni corrections. We compare our model with Birch (Yilmaz et al, 2019), BERT-MaxP (Dai and Callan, 2019), and PARADE , which are BERT-based models that represent the state of the art. BERT-MaxP and PARADE results are from fine-tuning on the MS MARCO data and then fine-tuning again on Robust04 (via cross-validation).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…The T5-3B results in bold are significantly better (p < 0.05) than T5-large, T5-base, and the corresponding baseline (BM25 or BM25+RM3), based on the Student's paired t-test with Bonferroni corrections. We compare our model with Birch (Yilmaz et al, 2019), BERT-MaxP (Dai and Callan, 2019), and PARADE , which are BERT-based models that represent the state of the art. BERT-MaxP and PARADE results are from fine-tuning on the MS MARCO data and then fine-tuning again on Robust04 (via cross-validation).…”
Section: Resultsmentioning
confidence: 99%
“…This leads to the standard multi-stage pipeline architecture where first-stage retrieval is followed by reranking using one or more machine learning models (Asadi and Lin, 2013;Nogueira et al, 2019a). This architecture underlies nearly all transformerbased approaches to document retrieval today, for example, CEDR (MacAvaney et al, 2019), BERT-MaxP (Dai and Callan, 2019), Birch (Yilmaz et al, 2019), and PARADE .…”
Section: Introductionmentioning
confidence: 99%
“…We identified numerous works on neural models published in recent years that would benefit from an evaluation and analysis based on FiRA. A common approach to utilizing the large-scale pre-trained BERT model [8] in document ranking is to apply BERT to passages [30], overlapping windows [32], or single sentences [1]. In all these cases, BERT produces partial results that need to be aggregated externally to produce a final ranking score that could be compared with traditional full-document judgements.…”
Section: Related Workmentioning
confidence: 99%
“…Following this observation, our system's architecture consists of modules ("IR primitives") that declare dependencies on other modules. 2 For example, in Figure 1, Searcher depends on an Index (which depends on a Collection), and Reranker depends on a Trainer and Extractor. Dependencies may specify both a module type (e.g., Searcher) and a default module class (e.g., BM25), which can be overridden by the user via the configuration.…”
Section: Architecturementioning
confidence: 99%
“…Each method makes a different efficiency vs. effectiveness trade-off and potentially operates on different features or document representations. With the growing popularity of computationally expensive BERT-based models (e.g., [2,6,13,17]) and substantially more expensive models based on T5 [18], the telescoping approach becomes particularly appealing as a means for reducing the number of documents these models evaluate.…”
Section: Introductionmentioning
confidence: 99%