Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval 2020
DOI: 10.1145/3397271.3401194
|View full text |Cite
|
Sign up to set email alerts
|

MarkedBERT: Integrating Traditional IR Cues in Pre-trained Language Models for Passage Retrieval

Abstract: The Information Retrieval (IR) community has witnessed a flourishing development of deep neural networks, however, only a few managed to beat strong baselines. Among them, models like DRMM and DUET were able to achieve better results thanks to the proper handling of exact match signals. Nowadays, the application of pretrained language models to IR tasks has achieved impressive results exceeding all previous work. In this paper, we assume that established IR cues like exact term-matching, proven to be valuable … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(26 citation statements)
references
References 6 publications
0
26
0
Order By: Relevance
“…For Siamese BERT, we follow the structure described in the paper (Reimers and Gurevych, 2019) for this experiment. Besides, we have tried to implement more recent matching models (Liu et al, 2018b;Hofstätter, 2020) and BERT-based variant models (Boualili et al, 2020; Rudra and Anand, 2020), but we do not obtain the expected excellent results. For fair comparison, these methods are not included in this paper.…”
Section: A3 Comparison Methodsmentioning
confidence: 92%
“…For Siamese BERT, we follow the structure described in the paper (Reimers and Gurevych, 2019) for this experiment. Besides, we have tried to implement more recent matching models (Liu et al, 2018b;Hofstätter, 2020) and BERT-based variant models (Boualili et al, 2020; Rudra and Anand, 2020), but we do not obtain the expected excellent results. For fair comparison, these methods are not included in this paper.…”
Section: A3 Comparison Methodsmentioning
confidence: 92%
“…In this section, we introduce recent works designing PTMs tailored for IR (Lee et al, 2019b;Chang et al, 2019;Ma et al, 2021b;Ma et al, 2021c;Boualili et al, 2020;Ma et al, 2021d;Zou et al, 2021;Liu et al, 2021d). General pre-trained models like BERT have achieved great success when applied to IR tasks on both the firststage retrieval and the re-ranking stage.…”
Section: Keyphrase Extractionmentioning
confidence: 99%
“…As explained earlier, exact matching is an important matching signal in traditional IR models, and relevance matching-based neural ranking models incorporate the exact matching signal to improve retrieval results. In order to directly incorporate exact matching signal in the sentence pair classification setting of BERT for document retrieval, Boualili et al (2020) proposed to mark the start and end of exact matching query tokens in a document with special markers.…”
Section: Deep Contextualized Language Model-based Representationsmentioning
confidence: 99%
“…Second, some BERTbased models, such as MacAvaney et al (2019), are combined with existing relevancebased ranking models that are asymmetric, and others, such as Nogueira et al (2019), include a component for pairwise comparison of documents, so that the joint model is asymmetric in both cases. Third, Boualili et al (2020) include the exact matching of query tokens into the ranking model which leads to an overall asymmetric architecture. In the case of short documents where BERT can accept the full document and only the BERTbased model is used for ranking, the ranking model is symmetric.…”
Section: Exact Matchingmentioning
confidence: 99%