Findings of the Association for Computational Linguistics: NAACL 2022 2022
DOI: 10.18653/v1/2022.findings-naacl.115
|View full text |Cite
|
Sign up to set email alerts
|

UniK-QA: Unified Representations of Structured and Unstructured Knowledge for Open-Domain Question Answering

Abstract: We study open-domain question answering with structured, unstructured and semistructured knowledge sources, including text, tables, lists and knowledge bases. Departing from prior work, we propose a unifying approach that homogenizes all sources by reducing them to text and applies the retrieverreader model which has so far been limited to text sources only. Our approach greatly improves the results on knowledge-base QA tasks by 11 points, compared to latest graphbased methods. More importantly, we demonstrate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 40 publications
(23 citation statements)
references
References 41 publications
0
23
0
Order By: Relevance
“…7 We also normalize embeddings before computing ψ. Following recent work (Gillick et al, 2019;Karpukhin et al, 2020), we use hard negative sampling to add the top nearest incorrect entities for each query to the batch. 8 We follow Botha et al (2020) to balance the hard negatives by fixing the ratio of positive to negative examples allowed for each entity, reducing the proportion of hard negatives that are rare entities (see Appendix A.4).…”
Section: Type-enforced Contrastive Lossmentioning
confidence: 99%
See 4 more Smart Citations
“…7 We also normalize embeddings before computing ψ. Following recent work (Gillick et al, 2019;Karpukhin et al, 2020), we use hard negative sampling to add the top nearest incorrect entities for each query to the batch. 8 We follow Botha et al (2020) to balance the hard negatives by fixing the ratio of positive to negative examples allowed for each entity, reducing the proportion of hard negatives that are rare entities (see Appendix A.4).…”
Section: Type-enforced Contrastive Lossmentioning
confidence: 99%
“…Prior work has shown that a hybrid model that combines sparse retrievers (e.g. TF-IDF) and dense retrievers can improve performance (Karpukhin et al, 2020;Luan et al, 2021) and that entity popularity can help disambiguation (Ganea and Hofmann, 2017).…”
Section: Type-enforced Contrastive Lossmentioning
confidence: 99%
See 3 more Smart Citations