Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.519
|View full text |Cite
|
Sign up to set email alerts
|

End-to-End Training of Neural Retrievers for Open-Domain Question Answering

Abstract: Recent work on training neural retrievers foropen-domain question answering (OpenQA) has employed both supervised and unsupervised approaches. However, it remains unclear how unsupervised and supervised methods can be used most effectively for neural retrievers. In this work, we systematically study retriever pre-training. We first propose an approach of unsupervised pre-training with the Inverse Cloze Task and masked salient spans, followed by supervised finetuning using question-context pairs. This approach … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
57
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 49 publications
(57 citation statements)
references
References 14 publications
0
57
0
Order By: Relevance
“…While this is a valid mechanism, perhaps conditioning on individual passages like we do is more precise for relevance supervision. Indeed, recent work (Sachan et al, 2021) illustrates by using Fusion-in-Decoder during inference but foregoing the decoder's attention weights and using an equivalent version of the MARGINALIZEDLOSS for training the retriever. Furthermore, Fusion-in-Decoder is uniquely useful for QA style tasks, where it has to select the correct answer from many passages.…”
Section: Discussionmentioning
confidence: 99%
“…While this is a valid mechanism, perhaps conditioning on individual passages like we do is more precise for relevance supervision. Indeed, recent work (Sachan et al, 2021) illustrates by using Fusion-in-Decoder during inference but foregoing the decoder's attention weights and using an equivalent version of the MARGINALIZEDLOSS for training the retriever. Furthermore, Fusion-in-Decoder is uniquely useful for QA style tasks, where it has to select the correct answer from many passages.…”
Section: Discussionmentioning
confidence: 99%
“…Pretraining for dense retrieval has recently gained a considerable attention, following the success of self-supervised models in many NLP tasks Liu et al, 2019;Brown et al, 2020). While most works focus on fine-tuning such retrievers on large datasets after pretraining Guu et al, 2020;Sachan et al, 2021;Gao and Callan, 2021a), we attempt to bridge the gap between unsupervised dense models and strong sparse (e.g., BM25; Robertson and Zaragoza 2009) or supervised dense baselines (e.g., DPR; Karpukhin et al 2020). A concurrent work by Oguz et al (2021) presented DPR-PAQ, which shows strong results on NQ after pretraining.…”
Section: Related Workmentioning
confidence: 98%
“…This is done mainly to avoid uninformative recurring words, e.g., verbs or adjectives. Note that as opposed to other approaches for span filtering (Glass et al, 2020;Guu et al, 2020;Sachan et al, 2021), our heuristics do not require any model.…”
Section: Pretraining: Recurring Span Retrievalmentioning
confidence: 99%
“…Most of the OpenQA models also consist of a retriever and a reasoner. The retriever is devised as a sparse term-based method such as BM25 (Robertson and Zaragoza, 2009) or a trainable dense passage retrieval method (Karpukhin et al, 2020;Sachan et al, 2021a), and the reasoner deals with each doc-ument individually (Guu et al, 2020) or fuses all the documents together (Izacard and Grave, 2021). Different from the documents in openQA, the subgraphs in KBQA can be only obtained by multi-hop retrieval and the reasoner should deal with the entire subgraph instead of each individual relation to find the answer.…”
Section: Related Workmentioning
confidence: 99%