2024
DOI: 10.1109/tcsvt.2023.3296668
|View full text |Cite
|
Sign up to set email alerts
|

Spatial–Temporal Enhanced Network for Continuous Sign Language Recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(1 citation statement)
references
References 0 publications
0
1
0
Order By: Relevance
“…1DCNN is widely used in CSLR for temporal feature extraction in recent years because of its advantages of simple structure and small number of parameters. 11,17,18 Most of the CSLR models based on these algorithms in recent years have been extracted from local features of fixed temporal receptive fields. 18,19 However, in the original video sequence, the length of the video clip sequence corresponding to different glosses is different, and the proficiency of sign language and some other interference caused by different sign language performers in the process of presentation will cause the same gloss to have inconsistent time used, as shown in Fig.…”
Section: Introductionmentioning
confidence: 99%
“…1DCNN is widely used in CSLR for temporal feature extraction in recent years because of its advantages of simple structure and small number of parameters. 11,17,18 Most of the CSLR models based on these algorithms in recent years have been extracted from local features of fixed temporal receptive fields. 18,19 However, in the original video sequence, the length of the video clip sequence corresponding to different glosses is different, and the proficiency of sign language and some other interference caused by different sign language performers in the process of presentation will cause the same gloss to have inconsistent time used, as shown in Fig.…”
Section: Introductionmentioning
confidence: 99%