Proceedings of the 1st Workshop on Multilingual Representation Learning 2021
DOI: 10.18653/v1/2021.mrl-1.18
|View full text |Cite
|
Sign up to set email alerts
|

Multilingual Code-Switching for Zero-Shot Cross-Lingual Intent Prediction and Slot Filling

Abstract: Predicting user intent and detecting the corresponding slots from text are two key problems in Natural Language Understanding (NLU). Since annotated datasets are only available for a handful of languages, our work focuses particularly on a zero-shot scenario where the target language is unseen during training. In the context of zero-shot learning, this task is typically approached using representations from pre-trained multilingual language models such as mBERT or by fine-tuning on data automatically translate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 61 publications
0
5
0
Order By: Relevance
“…When porting a dialogue system to new languages, zero-shot transfer is an effective method to bypass costly data collection and annotation for every target language. However, based on prior work in general-purpose cross-lingual NLP, we detect three crucial gaps which require more attention in future work and that may play an instrumental role in the final task performance: 1) the choice of source language(s) (Zoph et al, 2016;Dabre et al, 2017;Lin et al, 2019), as recently hinted at for multilingual ToD NLU by Krishnan et al (2021); 2) harnessing multiple source languages rather than a single one (Zoph et al, 2016;Pan et al, 2020;Wu et al, 2020b;Ansell et al, 2021); 3) few-shot transfer with a small number of target-language examples, as opposed to fully zero-shot transfer (Lauscher et al, 2020). In other words, the go-to option of always transferring from English in a zero-shot fashion might be sub-optimal for a large number of target languages.…”
Section: Coping With Low-resource Scenarios In Nlumentioning
confidence: 92%
See 2 more Smart Citations
“…When porting a dialogue system to new languages, zero-shot transfer is an effective method to bypass costly data collection and annotation for every target language. However, based on prior work in general-purpose cross-lingual NLP, we detect three crucial gaps which require more attention in future work and that may play an instrumental role in the final task performance: 1) the choice of source language(s) (Zoph et al, 2016;Dabre et al, 2017;Lin et al, 2019), as recently hinted at for multilingual ToD NLU by Krishnan et al (2021); 2) harnessing multiple source languages rather than a single one (Zoph et al, 2016;Pan et al, 2020;Wu et al, 2020b;Ansell et al, 2021); 3) few-shot transfer with a small number of target-language examples, as opposed to fully zero-shot transfer (Lauscher et al, 2020). In other words, the go-to option of always transferring from English in a zero-shot fashion might be sub-optimal for a large number of target languages.…”
Section: Coping With Low-resource Scenarios In Nlumentioning
confidence: 92%
“…Cross-Lingual Transfer Methods. Given the absence of sufficiently large training data in many languages, the default approach to multilingual NLU is (zero-shot or few-shot) transfer of models trained on English datasets by means of massively multilingual Transformer-based encoders (Zhang et al, 2019b;Xu et al, 2020;Siddhant et al, 2020b;Krishnan et al, 2021). While most of the work relies on MEs pretrained via masked language modelling, such as mBERT (Devlin et al, 2019) and XLM-R , Siddhant et al (2020b) show that MEs pretrained via machine translation leads to more effective zero-shot transfer for intent classification.…”
Section: Natural Language Understanding (Nlu)mentioning
confidence: 99%
See 1 more Smart Citation
“…Generally, these approaches additionally require word alignment to project annotations between languages. Prior zero-shot cross-lingual work in SLU (Li et al, 2021;Zhu et al, 2020;Krishnan et al, 2021) similarly identifies a penalty for cross-lingual transfer and finds that pre-trained models and machine translation can only partially mitigate this error.…”
Section: Related Workmentioning
confidence: 99%
“…The progress in multilingual TOD is critically hampered by the paucity of training data for many of the world's languages. While cross-lingual transfer learning (Zhang et al, 2019;Xu et al, 2020;Krishnan et al, 2021) offers a partial remedy, its success is tenuous beyond typologically similar languages and generally hard to assess due to the lack of evaluation benchmarks (Razumovskaia et al, 2021). What is more, transfer learning often cannot leverage multi-source transfer and few-shot learning due to lack of language diversity in the TOD datasets (Zhu et al, 2020;Quan et al, 2020;Farajian et al, 2020).…”
Section: Introductionmentioning
confidence: 99%