2021
DOI: 10.3390/app11198842
|View full text |Cite
|
Sign up to set email alerts
|

AUDD: Audio Urdu Digits Dataset for Automatic Audio Urdu Digit Recognition

Abstract: The ongoing development of audio datasets for numerous languages has spurred research activities towards designing smart speech recognition systems. A typical speech recognition system can be applied in many emerging applications, such as smartphone dialing, airline reservations, and automatic wheelchairs, among others. Urdu is a national language of Pakistan and is also widely spoken in many other South Asian countries (e.g., India, Afghanistan). Therefore, we present a comprehensive dataset of spoken Urdu di… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 48 publications
(13 citation statements)
references
References 28 publications
0
13
0
Order By: Relevance
“…For effective learning and implementation, a CNN model needs to be trained with a large amount of the training data [18,33]. e deep learningbased approaches need to be trained on large training datasets for avoiding overfitting and to maximize learning.…”
Section: Data Augmentationmentioning
confidence: 99%
“…For effective learning and implementation, a CNN model needs to be trained with a large amount of the training data [18,33]. e deep learningbased approaches need to be trained on large training datasets for avoiding overfitting and to maximize learning.…”
Section: Data Augmentationmentioning
confidence: 99%
“…Moreover, in order to avoid memorization, a dropout of 0.15 is used. For increasing the training samples, two augmentation strategies are utilized including random rotations and random Gaussian noise [26,32,58]. The utilized parameters for training the CNN model are shown in Table 1.…”
Section: Our Deep Learning Model (22)mentioning
confidence: 99%
“…Nevertheless, they proved to have poor performance especially when the network is large like CNN since the learning rate needs to be manually tuned in SGD. This significantly increases the training time for large-scale datasets [ 26 , 27 ]. To overcome this obstacle and improve the efficiency of adaptive, new variants of adaptive gradient methods are proposed such as Nostalgic Adam [ 28 ], which place bigger weights on the past gradient compared to the recent gradient, or YOGI [ 29 ], which increases in effective learning rate to achieve better convergence.…”
Section: Introductionmentioning
confidence: 99%