Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.240
|View full text |Cite
|
Sign up to set email alerts
|

UnitedQA: A Hybrid Approach for Open Domain Question Answering

Abstract: To date, most of recent work under the retrieval-reader framework for open-domain QA focuses on either extractive or generative reader exclusively. In this paper, we study a hybrid approach for leveraging the strengths of both models. We apply novel techniques to enhance both extractive and generative readers built upon recent pretrained neural language models, and find that proper training methods can provide large improvements over previous state-of-the-art models. We demonstrate that an hybrid approach by c… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
12
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
2

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(13 citation statements)
references
References 27 publications
1
12
0
Order By: Relevance
“…We hypothesize this may be caused by ELECTRA pre-training method, which shows strong performance through variety of tasks and we further show that it is also due to training and inference with large input size of 128 passages and better objective (discussed in Section 4.2 and Appendix G). Only system that matches the performance of our extractive reader is the concurrent work on UnitedQA-E (Cheng et al, 2021), which uses advanced regularization and Har-dEM techniques. We note that these are orthogonal to our approach and could potentially lead to further improvements.…”
Section: Results and Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…We hypothesize this may be caused by ELECTRA pre-training method, which shows strong performance through variety of tasks and we further show that it is also due to training and inference with large input size of 128 passages and better objective (discussed in Section 4.2 and Appendix G). Only system that matches the performance of our extractive reader is the concurrent work on UnitedQA-E (Cheng et al, 2021), which uses advanced regularization and Har-dEM techniques. We note that these are orthogonal to our approach and could potentially lead to further improvements.…”
Section: Results and Analysismentioning
confidence: 99%
“…Extractive BM25+BERT (Mao et al, 2020) 37.7 60.1 110M Hard EM (Min et al, 2019a) 28.1 50.9 110M Path Retriever (Asai et al, 2020) 32.6 -447M Graph Retriever (Min et al, 2019b) 34.5 56.0 110M ORQA 33.3 45.0 220M REALM (Guu et al, 2020) 40.4 -660M ProQA (Xiong et al, 2021) 34.3 -220M DPR 41.5 56.8 220M RDR (Yang and Seo, 2020) 42.1 57.0 110M GAR+DPR (Mao et al, 2020) 43.8 -626M ColBERT (Khattab et al, 2020) 48.2 63.2 − 440M RIDER (GAR+DPR) (Mao et al, 2021) 48.3 -626M UnitedQA-E (Cheng et al, 2021) 51.8 68.9 440M Generative BM25+SSG (Mao et al, 2020) 35.3 58.6 406M T51.1+SSM 35.2 61.6 11B RAG 44.5 56.8 516M DPR+SSG 42.2 -516M FiD-base (Izacard and Grave, 2021) 48.2 65.0 333M FiD-large (Izacard and Grave, 2021) 51.4 67.6 848M FiD-large++ 54 Finally, we find that our R2-D2 system with 21M passages corpus is competitive even with FiD++, which uses DPR retriever improved via knowledge distillation, and 26M passage corpus, which also includes lists. Additionally, we evaluate our model with a better retrieval model (HN-DPR) based on the DPR checkpoint where hard negatives are mined using the retrieval model itself 11 .…”
Section: Methods Nq Tq #θmentioning
confidence: 99%
See 1 more Smart Citation
“…This approach was also applied in RocketQA . UnitedQA (Cheng et al, 2021) and R2D2 (Fajcik et al, 2021) combine results from an ensemble of extractive and generative readers, whereas PAQ directly retrieves possible answers with an FiD fallback.…”
Section: E2e Optimization Of Nralmmentioning
confidence: 99%
“…All these benchmarks either assume that each question has only one answer with several alternative surface forms, or only require a system to predict one valid answer. A typical question answering system is a pipeline as follows: an efficient retriever retrieves relevant passages using sparse (Mao et al, 2021; or dense Xiong et al, 2021;Izacard and Grave, 2021a;Khattab et al, 2021) representations; an optional passage reranker (Asadi and Lin, 2013;Nogueira and Cho, 2019;Nogueira et al, 2020) further narrows down the evidence; an extractive or generative reader (Izacard and Grave, 2021b;Cheng et al, 2021) predicts an answer conditioned on retrieved or top-ranking passages. Nearly all previous work focused on locating passages covering at least one answer, or tried to predict one answer precisely.…”
Section: Related Workmentioning
confidence: 99%