2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP) 2021
DOI: 10.1109/iscslp49672.2021.9362086
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging Text Data Using Hybrid Transformer-LSTM Based End-to-End ASR in Transfer Learning

Abstract: In this work, we study leveraging extra text data to improve lowresource end-to-end ASR under cross-lingual transfer learning setting. To this end, we extend our prior work [1], and propose a hybrid Transformer-LSTM based architecture. This architecture not only takes advantage of the highly effective encoding capacity of the Transformer network but also benefits from extra text data due to the LSTM-based independent language model network. We conduct experiments on our in-house Malay corpus which contains lim… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 18 publications
0
5
0
Order By: Relevance
“…Using the Hybrid based LSTM transformer, the WER is reduced with 25.4% by transfer learning. Additionally, 13% WER is reduced by LSTM decoder (Zeng et al, 2021). Transformer model encoding and decoding can be carried with self-attention and multi-head attention layer (Lee et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Using the Hybrid based LSTM transformer, the WER is reduced with 25.4% by transfer learning. Additionally, 13% WER is reduced by LSTM decoder (Zeng et al, 2021). Transformer model encoding and decoding can be carried with self-attention and multi-head attention layer (Lee et al, 2021).…”
Section: Related Workmentioning
confidence: 99%
“…Here, we use for the science transformer mode a simple attention mechanism in an initial decoder but use a recurrent net LSTM for the encoder as shown in Figure 13. Such mixtures have been investigated and compared [74,75]. We compare the two architectures shown in Figure 13a,b.…”
Section: Using Transformers For Spatial Bagsmentioning
confidence: 99%
“…Deep learning and generative artificial intelligence have recently made unprecedented advances in computational biology 11,24,25 . Generative deep learning has been successful for the sampling of protein conformations 25 , such as long short-term memory (LSTM) 26 , autoencoders (AEs) 27,28 , variational autoencoders (VAEs) 2933 , generative adversarial networks (GANs) 31,34 , score-based models 35 , energy-based models 36 , Transformer 31,37,38 , and active learning 33 . Training of deep generative models relies on conformations extracted from molecular dynamics simulations.…”
Section: Introductionmentioning
confidence: 99%