Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.73
|View full text |Cite
|
Sign up to set email alerts
|

R2-D2: A Modular Baseline for Open-Domain Question Answering

Abstract: This work presents a novel four-stage opendomain QA pipeline R2-D2 (RANK TWICE, READ TWICE). The pipeline is composed of a retriever, passage reranker, extractive reader, generative reader and a mechanism that aggregates the final prediction from all system's components. We demonstrate its strength across three open-domain QA datasets: Natu-ralQuestions, TriviaQA and EfficientQA, surpassing state-of-the-art on the first two. Our analysis demonstrates that: (i) combining extractive and generative reader yields … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(12 citation statements)
references
References 12 publications
0
12
0
Order By: Relevance
“…They achieve competitive performance across different knowledge-intensive NLP tasks (Izacard and Grave, 2021b;Glass et al, 2021;Paranjape et al, 2021;Park et al, 2021;Borgeaud et al, 2021). Recent work improves the retrieval component (Paranjape et al, 2021;Maillard et al, 2021) or introduces another passage re-ranking modules (Fajcik et al, 2021) for further improvements. Our work focuses on improving the generator component, which has been underexplored in the literature.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…They achieve competitive performance across different knowledge-intensive NLP tasks (Izacard and Grave, 2021b;Glass et al, 2021;Paranjape et al, 2021;Park et al, 2021;Borgeaud et al, 2021). Recent work improves the retrieval component (Paranjape et al, 2021;Maillard et al, 2021) or introduces another passage re-ranking modules (Fajcik et al, 2021) for further improvements. Our work focuses on improving the generator component, which has been underexplored in the literature.…”
Section: Related Workmentioning
confidence: 99%
“…This leverages a trained generator model and evaluates the relevance of each passage to a query through the correctness of the generated output when the passage is removed from the pool of retrieved passages. Unlike prior multi-task learning work in QA relying on available annotated data (Lee et al, 2021;Nishida et al, 2019) or heuristics such as answer string matching to label pseudo evidentiality (Fajcik et al, 2021), our approach is applicable to diverse downstream tasks, where we cannot use additional annotations or heuristics. Our evidentiality mining approach for high-quality silver labels can be applied to diverse NLP tasks, and our auxiliary task has a new purpose of evaluating passage evidentiality suitable for the open-retrieval.…”
Section: Introductionmentioning
confidence: 99%
“…Bi-encoder retrievers are effective in bringing out relevant passages from a large index but sometimes reranking those passages is essential as the downstream reader can only see a limited number of them. (Fajcik et al, 2021) uses reranker after their retriever as a cross-encoder to improve the recall which proves effective in the end-to-end question answering pipeline. Incorporating Wikidata (Hu et al, 2021) in translation for sentences including named entities is common in literature because the pretrained multilingual reader is unable to translate these entities as it has not seen them during training.…”
Section: Related Workmentioning
confidence: 99%
“…This approach was also applied in RocketQA . UnitedQA (Cheng et al, 2021) and R2D2 (Fajcik et al, 2021) combine results from an ensemble of extractive and generative readers, whereas PAQ directly retrieves possible answers with an FiD fallback.…”
Section: E2e Optimization Of Nralmmentioning
confidence: 99%