2021
DOI: 10.1109/taslp.2020.3037543
|View full text |Cite
|
Sign up to set email alerts
|

Improving Automatic Speech Recognition and Speech Translation via Word Embedding Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(3 citation statements)
references
References 55 publications
0
3
0
Order By: Relevance
“…Information retrieval, question answering, machine translation, and other downstream applications use NER as a pre-processing step. In an end-to-end multitasking context, word embedding methods like Word2Vec and fastText are used to improve speech translation (Chuang et al 2021 ). Cross domains adversarial learning models comprised of CNN, BiLSTM, and Word2Vec embedding are utilized to categorize the information from EHR available in the Chinese language and achieve F1-score of 74.39%.…”
Section: Review On Text Analytics Word Embedding Application and Deep...mentioning
confidence: 99%
See 1 more Smart Citation
“…Information retrieval, question answering, machine translation, and other downstream applications use NER as a pre-processing step. In an end-to-end multitasking context, word embedding methods like Word2Vec and fastText are used to improve speech translation (Chuang et al 2021 ). Cross domains adversarial learning models comprised of CNN, BiLSTM, and Word2Vec embedding are utilized to categorize the information from EHR available in the Chinese language and achieve F1-score of 74.39%.…”
Section: Review On Text Analytics Word Embedding Application and Deep...mentioning
confidence: 99%
“… Catelli et al ( 2020 ) Italian language EHR classification English i2b2 2014 de-identification corpus, the Italian SIRM COVID-19 de-identification corpus Bi-LSTM BERT, MultiBPEmb, and Flair Multilingual Fast embeddings MultiBPEmb + Flair multi-fast achieves a micro F1-score of 94.48% 6. Chuang et al ( 2021 ) Automatic speech recognition and translation system LibriSpeech corpus, Augmented LibriSpeech, Fisher Spanish corpora LSTM, BiLSTM, CNN Word2Vec, fastText Word2Vec model efficiently maps speech signals to semantic space 7. Zhang et al ( 2021 ) Chinese word representation SogouCA data, Wikipedia dump, Fudan dataset LSTM Word2Vec, GloVe, BERT, CWE LSTM + CWE achieves an F1-score of 95.53% for the NER task 8.…”
Section: Appendix Amentioning
confidence: 99%
“…At present, the embedded speech recognition system based on the microprocessor platform can train from the speech library, create a training model, and complete the recognition by combining the human voice and pattern, rather than a specific person for speech recognition. In this paper, DTW algorithm combined with polynomial kernel function is used to effectively process the time-varying features of embedded speech, and a new PDTW-SVM algorithm is obtained [2]. It has local interpolation capabilities and global generalization capabilities, as well as the time-varying characteristics of speech signals.…”
Section: Introductionmentioning
confidence: 99%