Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1398
|View full text |Cite
|
Sign up to set email alerts
|

Meta-Learning for Low-Resource Neural Machine Translation

Abstract: In this paper, we propose to extend the recently introduced model-agnostic meta-learning algorithm (MAML, Finn et al., 2017) for lowresource neural machine translation (NMT). We frame low-resource translation as a metalearning problem, and we learn to adapt to low-resource languages based on multilingual high-resource language tasks. We use the universal lexical representation (Gu et al., 2018b) to overcome the input-output mismatch across different languages. We evaluate the proposed meta-learning strategy u… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
283
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 312 publications
(284 citation statements)
references
References 42 publications
(71 reference statements)
1
283
0
Order By: Relevance
“…There have been emerging research studies that utilize the above meta-learning algorithms to NLP tasks, including language modelling (Vinyals et al, 2016), text classification , machine translation (Gu et al, 2018), and relation learning (Xiong et al, 2018;Gao et al, 2019). In this paper, we propose to formulate the OOV word representation learning as a few-shot regression problem.…”
Section: Related Workmentioning
confidence: 99%
“…There have been emerging research studies that utilize the above meta-learning algorithms to NLP tasks, including language modelling (Vinyals et al, 2016), text classification , machine translation (Gu et al, 2018), and relation learning (Xiong et al, 2018;Gao et al, 2019). In this paper, we propose to formulate the OOV word representation learning as a few-shot regression problem.…”
Section: Related Workmentioning
confidence: 99%
“…Multi-task learning may favor highresource tasks over low-resource ones while metalearning aims at learning a good initialization that can be adapted to any task with minimal training samples. The figure is adapted from Gu et al (2018). jectives and exhibits strong performance on several benchmarks, attracting huge attention from researchers. Another line of research tries to apply multi-task learning to representation learning (Liu et al, 2015;Luong et al, 2015).…”
Section: Introductionmentioning
confidence: 99%
“…The optimization algorithm itself can be designed in a way that favors fast adaption. Model-agnostic meta-learning (MAML, [Finn et al, 2017;Yoon et al, 2018;Gu et al, 2018]) achieved state-of-theart performance by directly optimizing the gradient towards a good parameter initialization for easy fine-tuning on lowresource scenarios. It introduces no additional architectures nor parameters.…”
Section: Meta-learningmentioning
confidence: 99%
“…We obtained in total 600 annotations on each individual metric for each target domain. We calculated the Fleiss' kappa [Fleiss, 1971] to measure inter-rater consistency. The overall Fleiss' kappa values for informativeness and naturalness are 0.475 and 0.562, indicating "Moderate Agreement", and 0.637 for pairwise preferences, indicating "Substantial Agreement".…”
Section: Manual Evaluationmentioning
confidence: 99%