Proceedings of the 2019 2nd International Conference on Signal Processing and Machine Learning 2019
DOI: 10.1145/3372806.3372815
|View full text |Cite
|
Sign up to set email alerts
|

A Vision-based Human Action Recognition System for Moving Cameras Through Deep Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 6 publications
0
7
0
Order By: Relevance
“…In addition, they estimated the liquid intake volume by assuming that each intake sip was constantly 100 mL, leading to a poor estimate that was not validated [ 103 ]. Chang et al used deep learning model trained with video and depth stream from a Kinect camera to classify several types of human activities, including drinking [ 104 ]. An average accuracy of 96.4% was achieved when combining color, depth and optical flow in a CNN algorithm.…”
Section: Vision- and Environmental-based Methodsmentioning
confidence: 99%
“…In addition, they estimated the liquid intake volume by assuming that each intake sip was constantly 100 mL, leading to a poor estimate that was not validated [ 103 ]. Chang et al used deep learning model trained with video and depth stream from a Kinect camera to classify several types of human activities, including drinking [ 104 ]. An average accuracy of 96.4% was achieved when combining color, depth and optical flow in a CNN algorithm.…”
Section: Vision- and Environmental-based Methodsmentioning
confidence: 99%
“…In a related survey, a vision-based human action recognition system using a deeplearning technique is proposed by Chang et al [16], which can recognise human actions by retrieving information from colour videos, optical flow videos, and depth videos from the camera. This core research of HAR is not focused on the classroom; rather, it is based on activities in an indoor environment.…”
Section: Related Studymentioning
confidence: 99%
“…The fusion of modalities like RGB and depth information further refines recognition. Recent strides in attention mechanisms and metaheuristic algorithms have optimized network architectures, emphasizing relevant regions for improved performance [9,[61][62][63][64][65][66][67][68][69][70][71][72].…”
Section: Human Action Recognitionmentioning
confidence: 99%