A wide variety of uses, such as video interpretation and surveillance, human-robot interaction, healthcare, and sport analysis, among others, make this technology extremely useful, human activity recognition has received a lot of attention in recent decades. human activity recognition from video frames or still images is a challenging procedure because of factors including viewpoint, partial occlusion, lighting, background clutter, scale differences, and look. Numerous applications, including human-computer interfaces, robotics for the analysis of human behavior, and video surveillance systems all require the activity recognition system. This work introduces the human activity recognition system, which includes 3 stages: preprocessing, feature extraction, and classification. The input video (image frames) are subjected for preprocessing stage which is processed with median filtering and background subtraction. Several features, including the Improved Bag of Visual Words, the local texton XOR pattern, and the Spider Local Picture Feature (SLIF) based features, are extracted from the pre-processed image. The next step involves classifying data using a hybrid classifier that blends Bidirectional Gated Recurrent (Bi-GRU) and Long Short Term Memory (LSTM). To boost the effectiveness of the suggested system, the weights of the Long Short Term Memory (LSTM) and Bidirectional Gated Recurrent (Bi-GRU) are both ideally determined using the Improved Aquila Optimization with City Block Distance Evaluation (IACBD) method. Finally, the effectiveness of the suggested approach is evaluated in comparison to other traditional models using various performance metrics.
Classifying human actions from still images or video sequences is a demanding task owing to issues, like lighting, backdrop clutter, variations in scale, partial occlusion, viewpoint, and appearance. A lot of appliances, together with video systems, human-computer interfaces, and surveillance necessitate a compound action recognition system. Here, the proposed system develops a novel scheme for HAR. Initially, filtering as well as background subtraction is done during preprocessing. Then, the features including local binary pattern (LBP), bag of the virtual word (BOW), and the proposed local spatio-temporal features are extracted. Then, in the recognition phase, an ensemble classification model is introduced that includes Recurrent Neural networks (RNN 1 and RNN 2) and Multi-Layer Perceptron (MLP 1 and MLP 2). The features are classified using RNN 1 and RNN 2, and the outputs from RNN 1 and RNN 2 are further classified using MLP 1 and MLP 2, respectively. Finally, the outputs attained from MLP 1 and MLP 2 are averaged and the final classified output is obtained. At last, the superiority of the developed approach is proved on varied measures. K E Y W O R D Saccuracy, human action, multi-layer perceptron, proposed spatio-temporal features, RNN INTRODUCTION"HAR is defined as a classification task where the movement of a person is categorized based on data received from different sources such as sensors, cameras, etc.". It has different appliances in healthcare, particularly in constant action monitoring and fall recognition of aged persons. [1][2][3] In addition, it has vast prospective to support emerging appliances namely, augmented reality, IoT, and indoor localization in smart construction controlling systems to offer a secure indoor surrounding with higher EE.Diverse studies were carried out on HAR techniques in the state of the works using computer vision, ambient devices, sensors, and smartphones. These approaches are usually classified into two classes, that is, passive and active monitoring schemes. [4][5][6] In an AMS, exterior devices like wearable sensors and cameras were deployed. Sensor-based HAR is extensively deployed in analysis, in which a person has to continually carry sensors such as gyroscope, pedometer, and accelerometer that can turn out to be difficult for the person in numerous cases, particularly for elder ones. [7][8][9]
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2025 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.