Moving object segmentation in video plays an important role in many computer vision applications such as outdoor video surveillance, robotics, self-driving cars, etc. Existing work comprises complicated training procedures considering more number of history frames, fine-tuning on test sequences, and pre-trained models adopted from other applications. Therefore, this paper proposes an end-to-end robust inter-frame change exposure (RIChEx) framework for segmenting moving objects using three consecutive frames. The RIChEx consists inter-frame change gradient (IFCG) aggregation approach responsible for extracting the relevant foreground features to capture substantial changes and spatio-temporal structural dependencies between the consecutive frames. This module consists of multi-scale addition (MSA) blocks for extracting multi-scale features in early layers. Furthermore, to enhance the temporal learning capability of the module, previous recurrent responses and optical flow information are integrated to estimate probable foreground availability and salient object motions. We have evaluated the proposed framework on the benchmark DAVIS-16 dataset to validate the model's performance.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.