2020
DOI: 10.48550/arxiv.2009.12756
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval

Abstract: We propose a simple and efficient multi-hop dense retrieval approach for answering complex open-domain questions, which achieves state-of-the-art performance on two multi-hop datasets, HotpotQA and multi-evidence FEVER. Contrary to previous work, our method does not require access to any corpus-specific information, such as inter-document hyperlinks or human-annotated entity markers, and can be applied to any unstructured text corpus. Our system also yields a much better efficiency-accuracy trade-off, matching… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

1
23
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 31 publications
(24 citation statements)
references
References 44 publications
1
23
0
Order By: Relevance
“…More broadly, RD parallels current trends in open-domain multi-hop question answering, where the state of art has moved to iteratively querying batches of source documents from a corpus (Xiong et al, 2020;Qi et al, 2020). While a different domain from grounded language understanding, this reflects an emerging recognition that a single forwardpass is often insufficient for complex, multi-stage reasoning or planning tasks.…”
Section: Related Workmentioning
confidence: 99%
“…More broadly, RD parallels current trends in open-domain multi-hop question answering, where the state of art has moved to iteratively querying batches of source documents from a corpus (Xiong et al, 2020;Qi et al, 2020). While a different domain from grounded language understanding, this reflects an emerging recognition that a single forwardpass is often insufficient for complex, multi-stage reasoning or planning tasks.…”
Section: Related Workmentioning
confidence: 99%
“…With the support of approximate nearest neighbor (ANN) search [11,15], dense retrievers can efficiently retrieve documents by conducting semantic matching in the embedding space. To encode queries and documents as dense representations, dense retrievers usually employ the BERT-Siamese architecture to provide fully learnable, well pretrained, and effective representations for queries and documents to construct embedding space for retrieval [16,40,41].…”
Section: Introductionmentioning
confidence: 99%
“…To optimize the embedding space for document retrieval, existing dense retrieval models usually sample negative documents and use them to contrastively train dense retrievers to learn query and document representations [16,40,41]. Wang et al [36] prove that contrastive representation learning optimizes neural models and keeps two properties of the document embedding space, "alignment" and "uniformity".…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations