2017
DOI: 10.1007/s11042-017-5460-9
|View full text |Cite
|
Sign up to set email alerts
|

End-to-end video background subtraction with 3d convolutional neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
75
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 102 publications
(75 citation statements)
references
References 62 publications
0
75
0
Order By: Relevance
“…Sakkos et al [161] designed an end-to-end 3D-CNN to track temporal changes in video sequences avoiding the use of a background model for the training. 3D-CNN can handle multiple scenes without further fine-tuning on each scene individually.…”
Section: D-cnnsmentioning
confidence: 99%
“…Sakkos et al [161] designed an end-to-end 3D-CNN to track temporal changes in video sequences avoiding the use of a background model for the training. 3D-CNN can handle multiple scenes without further fine-tuning on each scene individually.…”
Section: D-cnnsmentioning
confidence: 99%
“…Other authors emulated background subtraction strategies based on DNNs framework [10][11][12]23]. DNNs exhibit excellent performances in segmentation tasks such as semantic segmentation tasks [22,24,25] and foreground segmentation tasks [26].…”
Section: Related Workmentioning
confidence: 99%
“…DNN-based background subtraction methods have different frameworks [22,24,25]. DNNs in [22,24,25] used a single image as an input, whereas the DNN-based background subtraction in [10][11][12]23] used more than two images consisting of observed images and a background image.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Another typical preprocessing is to subtract the background to extract the foreground objects. In BEnd-to-end video background subtraction with 3d convolutional neural networks^ [7], Sakkos et al propose an end-to-end temporalaware background subtraction approach with 3D convolutional neural networks. By performing 3D convolutions on the 10 most recent frames of the video, the changes on the spatial and temporal dimension are simultaneously tracked.…”
mentioning
confidence: 99%