Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.716
|View full text |Cite
|
Sign up to set email alerts
|

From English to Code-Switching: Transfer Learning with Strong Morphological Clues

Abstract: Linguistic Code-switching (CS) is still an understudied phenomenon in natural language processing. The NLP community has mostly focused on monolingual and multilingual scenarios, but little attention has been given to CS in particular. This is partly because of the lack of resources and annotated data, despite its increasing occurrence in social media platforms. In this paper, we aim at adapting monolingual models to code-switched text in various tasks. Specifically, we transfer English knowledge from a pre-tr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 29 publications
(16 citation statements)
references
References 19 publications
(31 reference statements)
0
16
0
Order By: Relevance
“…Code-mixed text processing. Previous research on code-mixed text processing focused on constructing formal grammars (Joshi, 1982) and tokenlevel language identification (Bali et al, 2014;Solorio et al, 2014;Barman et al, 2014), before progressing to named entity recognition and part-of-speech tagging (Ball and Garrette, 2018;AlGhamdi and Diab, 2019;Aguilar and Solorio, 2020). Recent work explores code-mixing in higher-level tasks such as question answering and task-oriented dialogue Ahn et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Code-mixed text processing. Previous research on code-mixed text processing focused on constructing formal grammars (Joshi, 1982) and tokenlevel language identification (Bali et al, 2014;Solorio et al, 2014;Barman et al, 2014), before progressing to named entity recognition and part-of-speech tagging (Ball and Garrette, 2018;AlGhamdi and Diab, 2019;Aguilar and Solorio, 2020). Recent work explores code-mixing in higher-level tasks such as question answering and task-oriented dialogue Ahn et al, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…Linguistic code-switching is a phenomenon where multilingual speakers alternate between languages. Recently, monolingual models have been adapted to code-switched text in entity recognition (Aguilar and Solorio, 2019), part-ofspeech tagging (Soto and Hirschberg, 2018;Ball and Garrette, 2018), sentiment analysis (Joshi et al, 2016) and language identification (Mave et al, 2018;Yirmibeşoglu and Eryigit, 2018;Mager et al, 2019). Recently, KhudaBukhsh et al, 2020 have proposed an approach to sample code-mixed documents using minimal supervision.…”
Section: Related Workmentioning
confidence: 99%
“…Bi-LSTM CRF (I): This is the most commonly used model for the LI and POS tasks. It uses Bi-LSTM and CRF consecutively for POS tagging the CM text (Aguilar and Solorio, 2019;Bhattu et al, 2020a).…”
Section: Fcrf (J)mentioning
confidence: 99%