2015
DOI: 10.1109/tip.2015.2438550
|View full text |Cite
|
Sign up to set email alerts
|

Robust Video Object Cosegmentation

Abstract: With ever-increasing volumes of video data, automatic extraction of salient object regions became even more significant for visual analytic solutions. This surge has also opened up opportunities for taking advantage of collective cues encapsulated in multiple videos in a cooperative manner. However, it also brings up major challenges, such as handling of drastic appearance, motion pattern, and pose variations, of foreground objects as well as indiscriminate backgrounds. Here, we present a cosegmentation framew… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 131 publications
(3 citation statements)
references
References 34 publications
0
3
0
Order By: Relevance
“…Video co-segmentation. Some proposal-based methods, such as the models in [6,9,15], extract the common information among the proposals from video frames, but proposal-based methods easily lead to a pixel misclassification problem. Therefore, we need to extract the common information between frames in pixel-level.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Video co-segmentation. Some proposal-based methods, such as the models in [6,9,15], extract the common information among the proposals from video frames, but proposal-based methods easily lead to a pixel misclassification problem. Therefore, we need to extract the common information between frames in pixel-level.…”
Section: Related Workmentioning
confidence: 99%
“…There are some methods [4][5][6][7][8][9][10][11] proposed to separate the common objects from the background information among video frames for video co-segmentation; however, most of these methods are not good at feature representation. On the other hand, deep learning models [12] have the high learning capability in feature representation that is needed for feature extraction in video co-segmentation.…”
Section: Introductionmentioning
confidence: 99%
“…Cheng et al [11] proposed a global contrast based algorithm, which employed the spatial distance between the current region and other regions as a factor to weigh the color histogram contrast. For temporal saliency detection [12]- [15], motion information is considered as the primarily influential factor. Gao et al [16] extended the image saliency model in [17] by adding the motion channel to predict human eye fixations in dynamic scenes.…”
Section: Introductionmentioning
confidence: 99%