Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.87
|View full text |Cite
|
Sign up to set email alerts
|

Enhancing Answer Boundary Detection for Multilingual Machine Reading Comprehension

Abstract: Multilingual pre-trained models could leverage the training data from a rich source language (such as English) to improve the performance on low resource languages. However, the transfer effectiveness on the multilingual Machine Reading Comprehension (MRC) task is substantially poorer than that for sentence classification tasks, mainly due to the requirement of MRC to detect the word level answer boundary. In this paper, we propose two auxiliary tasks to introduce additional phrase boundary supervision in the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
38
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
2
1

Relationship

3
4

Authors

Journals

citations
Cited by 23 publications
(38 citation statements)
references
References 20 publications
0
38
0
Order By: Relevance
“…Machine Reading Comprehension. Machine reading comprehension (MRC) (Rajpurkar et al, 2016) has received increasing attention recently, which requires a model to extract an answer span to a question from reference documents (Yu et al, 2018;Devlin et al, 2019;Zheng et al, 2020;Yuan et al, 2020). Owing to the rise of pre-training models (Devlin et al, 2018), a machine is able to achieve highly competitive results on classic datasets (e.g.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Machine Reading Comprehension. Machine reading comprehension (MRC) (Rajpurkar et al, 2016) has received increasing attention recently, which requires a model to extract an answer span to a question from reference documents (Yu et al, 2018;Devlin et al, 2019;Zheng et al, 2020;Yuan et al, 2020). Owing to the rise of pre-training models (Devlin et al, 2018), a machine is able to achieve highly competitive results on classic datasets (e.g.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, the more challenging distantly supervised MRC task, TriviaQA (Joshi et al, 2017) was proposed, in which the provided evidences are noisy and collected based on the distant supervision. (Yuan et al, 2020) proposed a multilingual MRC task to facilitate the study on low resource languages. (Lee et al, 2019b) focused on annotating the unlabeled data with heuristic method and refine the labels by an extra Refinery model for multilingual MRC task.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…These language models aim to learn language agnostic contextual representations by leveraging large-scale monolingual and parallel corpuses, which show great potential on cross-lingual tasks, such as sentence classification tasks (Hsu et al, 2019;Pires et al, 2019;Conneau et al, 2018). However, there is still a big gap between the performance of CLMRC in rich-resource languages and that in low-resource languages, since CLMRC requires the capability of fine-grained representation at the phase-level (Yuan et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…To further boost the performance of multilingual PLM on CLMRC task, Yuan et al (2020) propose two auxiliary tasks mixMRC and LAKM on top of multilingual PLM. Those auxiliary tasks improve the answer boundary detection quality in low-resource languages.…”
Section: Introductionmentioning
confidence: 99%