ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8682850
|View full text |Cite
|
Sign up to set email alerts
|

Investigating End-to-end Speech Recognition for Mandarin-english Code-switching

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
50
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 65 publications
(50 citation statements)
references
References 20 publications
0
50
0
Order By: Relevance
“…A multilingual ASR approach does not need an additional LID module to identify speech segments since language information is incorporated directly into the system [1]. One technique is to use a linguistic knowledge-based method to establish a multilingual phone set mapping or clustering of similar phonetic features that share the training data [7]. Common examples are the International Phonetic Alphabet (IPA), Speech Assessment Methods Phonetic Alphabet (SAMPA) and Wordbet [15].…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…A multilingual ASR approach does not need an additional LID module to identify speech segments since language information is incorporated directly into the system [1]. One technique is to use a linguistic knowledge-based method to establish a multilingual phone set mapping or clustering of similar phonetic features that share the training data [7]. Common examples are the International Phonetic Alphabet (IPA), Speech Assessment Methods Phonetic Alphabet (SAMPA) and Wordbet [15].…”
Section: Related Workmentioning
confidence: 99%
“…The IPA-based phoneme set, and data-driven phoneme set contained 38 phonemes, excluding the silent phonemes. In this case, to train our multilingual acoustic model that effectively handled Sepedi-English code-switched speech, we adopted the technique used by Biswas et al [6], Shan et al [7] and Bhuvanagiri and Kopparapu [8]. Lastly, problematic words of Sepedi or English origin were manually reviewed for correct pronunciation prior to training the HMMs.…”
Section: Multilingual Dictionary and Phoneme Setmentioning
confidence: 99%
See 1 more Smart Citation
“…In the very first work [23], Seki et al explored an E2E ASR system for code-switching task on an artificially created dataset obtained by concatenating the monolingual utterances. In contrast, Shan et al [24] employed a real Mandarin-English code-switching dataset for developing the attention-based E2E ASR system. For improving the ASR performance, the multi-task learning (MTL) framework involving the language identification (LID) [25] was employed.…”
Section: Introductionmentioning
confidence: 99%
“…This includes techniques specifically targeting the acoustic model [1,2] and the language model [4,5,6] to handle code-mixing in speech. Apart from these cascaded ASR systems, there is also recent work on using end-to-end systems trained on multilingual data to recognize CM speech [7,8,9,10,11,12]. Leveraging monolingual sentences for code-mixed language models has been extensively studied in prior work [13,14,15,16,17].…”
Section: Introductionmentioning
confidence: 99%