2021
DOI: 10.3233/jifs-202841
|View full text |Cite
|
Sign up to set email alerts
|

Non-diacritized Arabic speech recognition based on CNN-LSTM and attention-based models

Abstract: Arabic language has a set of sound letters called diacritics, these diacritics play an essential role in the meaning of words and their articulations. The change in some diacritics leads to a change in the context of the sentence. However, the existence of these letters in the corpus transcription affects the accuracy of speech recognition. In this paper, we investigate the effect of diactrics on the Arabic speech recognition based end-to-end deep learning. The applied end-to-end approach includes CNN-LSTM and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 18 publications
0
9
0
Order By: Relevance
“…Finally, multi-head attention is used for the encoder and decoder. The data preprocessing, features ex-traction, and language modeling are performed as presented in [17,38]. Features extraction is very important aspect for machine learning systems [39][40][41][42][43][44].…”
Section: System Descriptionmentioning
confidence: 99%
See 4 more Smart Citations
“…Finally, multi-head attention is used for the encoder and decoder. The data preprocessing, features ex-traction, and language modeling are performed as presented in [17,38]. Features extraction is very important aspect for machine learning systems [39][40][41][42][43][44].…”
Section: System Descriptionmentioning
confidence: 99%
“…The look-ahead word-based LM probabilities are determined in each recognition phase according to the decoding of the word prefixes. The prefix trees are used to transform a word-based LM into a character-based LM [17,31]. We used a completely parallelized version of the decoding technique for GPUs that was presented in Espresso for Parallelization LM.…”
Section: Language Modelingmentioning
confidence: 99%
See 3 more Smart Citations