2021
DOI: 10.1007/s00371-021-02266-4
|View full text |Cite
|
Sign up to set email alerts
|

A multi-stream CNN for deep violence detection in video sequences using handcrafted features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 34 publications
(23 citation statements)
references
References 55 publications
0
23
0
Order By: Relevance
“…For example, the high accuracies of 100%, 98.3%, 97.5%, 97.1%, and 96.1% were offered by the Multi-stream CNN [36], Efficient 3D CNN [37], Xception + BiLSTM + Attentions [38], AlexNet + LSTM [40] and MobileNetV2 + LSTM [34] models, respectively. In contrast, our model offers 97.62% classification accuracy for this dataset, which is 2.38% and 0.68% lower than the highest accuracies offered by Multi-stream CNN [36] and Efficient 3D CNN [37], respectively. This fairly diminished classification accuracy of our model is effectively offset by the notable improvement in total parameters reduced by 77.82% and 53.84% compared to the respective models, resulting in a favourable trade-off.…”
Section: Results and Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…For example, the high accuracies of 100%, 98.3%, 97.5%, 97.1%, and 96.1% were offered by the Multi-stream CNN [36], Efficient 3D CNN [37], Xception + BiLSTM + Attentions [38], AlexNet + LSTM [40] and MobileNetV2 + LSTM [34] models, respectively. In contrast, our model offers 97.62% classification accuracy for this dataset, which is 2.38% and 0.68% lower than the highest accuracies offered by Multi-stream CNN [36] and Efficient 3D CNN [37], respectively. This fairly diminished classification accuracy of our model is effectively offset by the notable improvement in total parameters reduced by 77.82% and 53.84% compared to the respective models, resulting in a favourable trade-off.…”
Section: Results and Analysismentioning
confidence: 99%
“…For higher classification accuracy, Ehasan et al [35] proposed an UNet + PatchGAN-based unsupervised action translation network utilizing spatio-temporal features to identify violent behaviours and overcome the problem related to the insufficiency of relevant data. Similarly, Mohtavipour et al [36] proposed a multi-stream CNN-based AVDC approach. Despite the promising classification performance of this model, its computational efficiency in terms of total parameters remains suboptimal.…”
Section: Spatio-temporal Feature Modelsmentioning
confidence: 99%
“…Mohtavipour et al [22] offer a unique deep violence detection framework using characteristics extracted by manual labour. A convolutional neural network (CNN) receives these properties as streams of data in three different dimensions: space, time, and space-time.…”
Section: Related Workmentioning
confidence: 99%
“…In this study, we apply the suggested model to the UCF-Crime dataset [23], which contains a large amount of footage from public surveillance cameras documenting anomalous, unlawful, and violent behaviour in settings as diverse as schools, businesses, and streets. This dataset was chosen because its events are representative of those that occur often and in a variety of settings [24][25][26].…”
Section: Datasetmentioning
confidence: 99%
“…A new deep violence detection approach based on handcrafted techniques’ distinctive characteristics was presented ( Mohtavipour, Saeidi & Arabsorkhi, 2021 ). These characteristics are linked to appearance, movement speed, and representative images, and they are supplied to a CNN as spatial, temporal, and spatiotemporal streams.…”
Section: Classification Of Violence Detection Techniquesmentioning
confidence: 99%