2021
DOI: 10.1109/access.2021.3069714
|View full text |Cite
|
Sign up to set email alerts
|

Arabic Sign Language Recognition System Using 2D Hands and Body Skeleton Data

Abstract: This paper presents a novel Arabic Sign Language (ArSL) recognition system, using selected 2D hands and body key points from successive video frames. The system recognizes the recorded video signs, for both signer dependent and signer independent modes, using the concatenation of a 3D CNN skeleton network and a 2D point convolution network. To accomplish this, we built a new ArSL video-based sign database. We will present the detailed methodology of recording the new dataset, which comprises 80 static and dyna… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
29
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 43 publications
(30 citation statements)
references
References 51 publications
0
29
0
1
Order By: Relevance
“…Figure 2. The Indonesian sign movement recognition system development research series [19] Regarding the 3 rd phase, the feature extraction techniques that are examined in this work are skeleton feature extraction [20], hand shape feature extraction [21], and MobileNetV2 [22]. The skeleton and hand shape feature did not utilize deep learning models and experiments were run on another dataset, the inflectional words dataset.…”
Section: Proposed Methods For Sibi Sentence Recognition Systemmentioning
confidence: 99%
See 1 more Smart Citation
“…Figure 2. The Indonesian sign movement recognition system development research series [19] Regarding the 3 rd phase, the feature extraction techniques that are examined in this work are skeleton feature extraction [20], hand shape feature extraction [21], and MobileNetV2 [22]. The skeleton and hand shape feature did not utilize deep learning models and experiments were run on another dataset, the inflectional words dataset.…”
Section: Proposed Methods For Sibi Sentence Recognition Systemmentioning
confidence: 99%
“…The skeleton and hand shape feature extraction methods yielded a best accuracy of 86.6% and 99% respectively [20], [21]. The MobileNet feature extraction [22], involving deep learning models, was running on the sentence dataset, the same dataset used by this paper. This extraction technique can pinpoint the identity for each video frame (separating epenthesis-gestures, suffixes, prefixes and root words), enabling sentence recognition from SIBI gestures with a 99% accuracy rate.…”
Section: Proposed Methods For Sibi Sentence Recognition Systemmentioning
confidence: 99%
“…Before the popularity of deep neural network, researchers focused on extracting hand-crafted features from the frames and videos, including HOG [13]- [15], hand motion trajectories [34], [35] and body joint coordinates [12]. In recent decades, convolutional neural network (CNN), including both 2D and 3D CNN, has gradually become the most common choice for feature extraction [20]- [22], [36]. Various types of CNN structures have been experimented in SLR, including 3D ResNet [37] and 3D Inception [26].…”
Section: Related Workmentioning
confidence: 99%
“…Output size = 1+ (input size -lter size )/stride size …………….. (2) In all the situations, invariance in the translation is offered by the pooling layer, which represents the identity of a speci c object with respect to its visibility within the frame.…”
Section: Pooling Layermentioning
confidence: 99%