Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1607
|View full text |Cite
|
Sign up to set email alerts
|

Zero-shot Reading Comprehension by Cross-lingual Transfer Learning with Multi-lingual Language Representation Model

Abstract: Because it is not feasible to collect training data for every language, there is a growing interest in cross-lingual transfer learning. In this paper, we systematically explore zeroshot cross-lingual transfer learning on reading comprehension tasks with a language representation model pre-trained on multi-lingual corpus. The experimental results show that with pre-trained language representation zeroshot learning is feasible, and translating the source data into the target language is not necessary and even de… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
40
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 40 publications
(40 citation statements)
references
References 17 publications
0
40
0
Order By: Relevance
“…Zero shot cross lingual transfer learning (Conneau et al, 2018;Pires et al, 2019) in the NLP context refers to transferring a model which is trained to solve a specific task in a source language to solve that specific task in a different language. Initially we get our baseline results using the zero-shot setting, where we use pretrained transformer models fine-tuned on English question answering task and check their performance on the Bengali evaluation dataset following similar research on Chinese (Hsu et al, 2019), French (d'Hoffschmidt et al, 2020 and Japanese (Siblini et al, 2019) language. In the next section we fine-tune these models further with our translated Bengali SQuAD dataset and compare the baselines with the fine-tuned models.…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…Zero shot cross lingual transfer learning (Conneau et al, 2018;Pires et al, 2019) in the NLP context refers to transferring a model which is trained to solve a specific task in a source language to solve that specific task in a different language. Initially we get our baseline results using the zero-shot setting, where we use pretrained transformer models fine-tuned on English question answering task and check their performance on the Bengali evaluation dataset following similar research on Chinese (Hsu et al, 2019), French (d'Hoffschmidt et al, 2020 and Japanese (Siblini et al, 2019) language. In the next section we fine-tune these models further with our translated Bengali SQuAD dataset and compare the baselines with the fine-tuned models.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…Aside from using translated dataset, there has also been attempts of curating large question answering datasets in multiple other languages including French (d'Hoffschmidt et al, 2020), Korean (Lim et al, 2019), Russian (Efimov et al, 2020), Chinese (Cui et al, 2018;Shao et al, 2018) and benchmark models like QANet (Yu et al, 2018), BiDAF (Seo et al, 2016), BERT (Devlin et al, 2018) have been trained on them. In contrast to gathering translated or human annotated dataset for model training, zero shot transfer learning where pretrained models were evaluated directly on a new language after task specific training on question answering has also been attempted on reading comprehension tasks (Artetxe et al, 2019;Hsu et al, 2019;Siblini et al, 2019). To the best of our knowledge none of similar work have ever been attempted on Bengali so far.…”
Section: Question Answering In Englishmentioning
confidence: 99%
See 1 more Smart Citation
“…For this reason, research has been focused lately on models that can work in a zero-shot setting, i.e., without being explicitly trained on data from the target language or domain. This training paradigm has been utilized with great effect for several popular NLP problems, such as cross-lingual document retrieval [25], sequence labeling [26], cross-lingual dependency parsing [27], and reading comprehension [28]. More specific to classification tasks, Ye et al [29] developed a reinforcement learning framework for cross-task text classification, which was tested also on the problem of sentiment classification in a monolingual setting.…”
Section: Related Workmentioning
confidence: 99%
“…mBERT is a multilingual version of BERT, which is trained on Wikipedia monolingual corpora in 104 languages. This model proves to be surprisingly effective in a wide range of cross-lingual tasks [32] [33], e.g., reading comprehension, document classification, etc.…”
Section: Baselinementioning
confidence: 99%