2020
DOI: 10.1007/978-3-030-51310-8_5
|View full text |Cite
|
Sign up to set email alerts
|

Investigating Query Expansion and Coreference Resolution in Question Answering on BERT

Abstract: The Bidirectional Encoder Representations from Transformers (BERT) model produces state-of-the-art results in many question answering (QA) datasets, including the Stanford Question Answering Dataset (SQuAD). This paper presents a query expansion (QE) method that identifies good terms from input questions, extracts synonyms for the good terms using a widely-used language resource, WordNet, and selects the most relevant synonyms from the list of extracted synonyms. The paper also introduces a novel QE method tha… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(13 citation statements)
references
References 20 publications
0
13
0
Order By: Relevance
“…Same-language MT has been successfully used in many NLP applications, e.g. text-to-speech synthesis for creating alternative target sequences (Cahill et al, 2009), translation between varieties of the same language (Brazilian Portuguese to European Portuguese) (Fancellu et al, 2014), paraphrase generation (Plachouras et al, 2018), and producing many alternative sequences of a given input question in question answering (Bhattacharjee et al, 2020). In our case, we developed Portuguese-to-Portuguese MT systems that were able to generate n-best (same-language) alternative sentences of an input Portuguese sentence.…”
Section: Same-language Mtmentioning
confidence: 99%
See 1 more Smart Citation
“…Same-language MT has been successfully used in many NLP applications, e.g. text-to-speech synthesis for creating alternative target sequences (Cahill et al, 2009), translation between varieties of the same language (Brazilian Portuguese to European Portuguese) (Fancellu et al, 2014), paraphrase generation (Plachouras et al, 2018), and producing many alternative sequences of a given input question in question answering (Bhattacharjee et al, 2020). In our case, we developed Portuguese-to-Portuguese MT systems that were able to generate n-best (same-language) alternative sentences of an input Portuguese sentence.…”
Section: Same-language Mtmentioning
confidence: 99%
“…dcu (Haque et al, 2020) compared both phrasebased and neural models by extending the STAPLE data with additional corpora (selected for similarity to the task data under a language model), with the neural model performing better. They generate sets of high-scoring predictions according to beam searches, majority voting, and other techniques, and also run these initial translations through an additional paraphrasing model, placing third in the Portuguese track.…”
Section: Baselinesmentioning
confidence: 99%
“…Information Extraction (IE) plays a fundamental role as a backbone component in many downstream applications. For example, an application such as question answering may be improved by relying on relation extraction (RE) (Hu et al, 2019;Yu et al, 2017), coreference resolution (Bhattacharjee et al, 2020;Gao et al, 2019), named entity recognition (NER) (Molla et al, 2006;Singh et al, 2018), and entity linking (EL) (Broscheit, 2019;Chen et al, 2017) components. This also holds for other applications such as personalized news recommendation (Karimi et al, 2018;Wang et al, 2018Wang et al, , 2019, fact checking (Thorne & Vlachos, 2018;Zhang & Ghorbani, 2020), opinion mining (Sun et al, 2017), semantic search (Cifariello et al, 2019), and conversational agents (Roller et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…Coreference resolution refers to the task of detecting mentions of various entities and events and identifying groups of mentions referring to the same real-world entity or event. It is a fundamental NLP task that has several downstream applications such as question answering Bhattacharjee et al, 2020), textual entailment (Mitkov et al, 2012), building and maintaining KBs (Angeli et al, 2015;Angell et al, 2021), and multi-document summarization (Falke et al, 2017;Huang and Kurohashi, 2021). Often these downstream applications consume a set of documents, and thus require detection of coreference relations between event and entity mentions spread across documents such as multiple news articles.…”
Section: Introductionmentioning
confidence: 99%