Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval 2020
DOI: 10.1145/3397271.3401093
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Document Re-Ranking for Transformers by Precomputing Term Representations

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
67
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 85 publications
(67 citation statements)
references
References 19 publications
0
67
0
Order By: Relevance
“…Evaluation on the Robust04 and GOV2 test collections confirms that BERT-QE significantly outperforms BERT-Large with relatively small extra computational cost (up to 30%). In future work, we plan to further im-prove the efficiency of BERT-QE, by combining the proposed BERT-QE with the pre-computation techniques proposed recently (Khattab and Zaharia, 2020;MacAvaney et al, 2020a), wherein most of the computations could be performed offline. There are two hyper-parameters in BERT-QE, namely α and β, both of which are interpolation coefficients.…”
Section: Discussionmentioning
confidence: 99%
“…Evaluation on the Robust04 and GOV2 test collections confirms that BERT-QE significantly outperforms BERT-Large with relatively small extra computational cost (up to 30%). In future work, we plan to further im-prove the efficiency of BERT-QE, by combining the proposed BERT-QE with the pre-computation techniques proposed recently (Khattab and Zaharia, 2020;MacAvaney et al, 2020a), wherein most of the computations could be performed offline. There are two hyper-parameters in BERT-QE, namely α and β, both of which are interpolation coefficients.…”
Section: Discussionmentioning
confidence: 99%
“…Lighter deep LM rankers are developed (MacAvaney et al, 2020;Gao et al, 2020), but their cross attention operations are still too expensive for fullcollection retrieval.…”
Section: Related Workmentioning
confidence: 99%
“…Each module is a stack of transformer layers (Vaswani et al, 2017), initialized with weights from BERT. In a related approach, MacAvaney et al (2020) investigate the relationship between different numbers of dedicated layers of BERT for query-document interactions and measure the resulting speedup that is due to token representation caching, as well as its impact on the end-to-end ranking quality. Khattab and Zaharia (2020) propose a related approach, namely ColBERT.…”
Section: Key Concepts Of Neural Rankingmentioning
confidence: 99%