Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1599
|View full text |Cite
|
Sign up to set email alerts
|

Multi-passage BERT: A Globally Normalized BERT Model for Open-domain Question Answering

Abstract: BERT model has been successfully applied to open-domain QA tasks. However, previous work trains BERT by viewing passages corresponding to the same question as independent training instances, which may cause incomparable scores for answers from different passages. To tackle this issue, we propose a multi-passage BERT model to globally normalize answer scores across all passages of the same question, and this change enables our QA model find better answers by utilizing more passages. In addition, we find that sp… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
137
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 188 publications
(139 citation statements)
references
References 20 publications
1
137
1
Order By: Relevance
“…We follow the preprocessing approached proposed by Wang et al (2019) and split passages into 100-word long chunks with 50-word long strides. We use a BM25 retriever to retrieve the top n passages for each question as inputs to the reader and the Wikipedia dump provided by Chen et al ( 2017) as source corpus.…”
Section: Methodsmentioning
confidence: 99%
“…We follow the preprocessing approached proposed by Wang et al (2019) and split passages into 100-word long chunks with 50-word long strides. We use a BM25 retriever to retrieve the top n passages for each question as inputs to the reader and the Wikipedia dump provided by Chen et al ( 2017) as source corpus.…”
Section: Methodsmentioning
confidence: 99%
“…BERT-Base currently supports 104 languages including Vietnamese. BERT is used in many different applications, especially in the question answering system [24] [25] [26]. We use the fine-tuned BERT model for multilingual Q&A created by Manuel Romero because this model has dataset that is compatible with original dataset SQuAD v1.1 but also extends more eleven languages including Vietnamese language [27].…”
Section: F Bert Modulementioning
confidence: 99%
“…In this study, the BiLSTM-CRF model, known for its high performance, is used; the performance was improved by various feature extensions [21][22][23]. Recently, BERT has shown state-of-the-art performance in sequence labeling and classification NLP tasks [24].…”
Section: B Named Entity Recognition (Ner)mentioning
confidence: 99%