Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1343
|View full text |Cite
|
Sign up to set email alerts
|

Dual Language Models for Code Switched Speech Recognition

Abstract: In this work, we present a simple and elegant approach to language modeling for bilingual code-switched text. Since codeswitching is a blend of two or more different languages, a standard bilingual language model can be improved upon by using structures of the monolingual language models. We propose a novel technique called dual language models, which involves building two complementary monolingual language models and combining them using a probabilistic model for switching between the two. We evaluate the eff… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
14
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(14 citation statements)
references
References 15 publications
0
14
0
Order By: Relevance
“…Baheti et al (2017) find that fine-tuning with CS data after pretraining on monolingual data works best. Finally, another line of works suggests using a dual language model, where two monolingual LMs are combined by a probabilistic model (Garg et al, 2017(Garg et al, , 2018.…”
Section: Related Workmentioning
confidence: 99%
“…Baheti et al (2017) find that fine-tuning with CS data after pretraining on monolingual data works best. Finally, another line of works suggests using a dual language model, where two monolingual LMs are combined by a probabilistic model (Garg et al, 2017(Garg et al, , 2018.…”
Section: Related Workmentioning
confidence: 99%
“…Recent work [6] attempts to train a CS language model using fine-tuning. Similar work [7] integrates two monolingual language models (LMs) by introducing a special "switch" token in both languages when training the LM, and further incorporating this within automatic speech recognition (ASR). Other works synthesize additional CS text using the modeled distribution from the data [8,9].…”
Section: Introductionmentioning
confidence: 99%
“…Equivalence Constraint and Functional Head Constraint are used to build a better CS language model [6,7,8], and CS models with syntactic and semantic features are built to exploit more information [9,10]. Because of a large amount of monolingual data, monolingual language models for host and guest languages are learned separately, and then combined with a probabilistic model for switching between the two [11].…”
Section: Introductionmentioning
confidence: 99%