2018
DOI: 10.1109/tpami.2017.2662005
|View full text |Cite
|
Sign up to set email alerts
|

Saliency-Aware Video Object Segmentation

Abstract: Video saliency, aiming for estimation of a single dominant object in a sequence, offers strong object-level cues for unsupervised video object segmentation. In this paper, we present a geodesic distance based technique that provides reliable and temporally consistent saliency measurement of superpixels as a prior for pixel-wise labeling. Using undirected intra-frame and inter-frame graphs constructed from spatiotemporal edges or appearance and motion, and a skeleton abstraction step to further enhance saliency… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
163
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 467 publications
(163 citation statements)
references
References 65 publications
0
163
0
Order By: Relevance
“…Superpixel segmentation can be used as pre‐processing step for many computer vision tasks. In this paper, we apply superpixel segmentation in saliency detection [WSS15, WSYP18]. A recent method for saliency optimization from robust background detection [ZLWS14] uses superpixel segmentation as input to generate salient object detection results.…”
Section: Methodsmentioning
confidence: 99%
“…Superpixel segmentation can be used as pre‐processing step for many computer vision tasks. In this paper, we apply superpixel segmentation in saliency detection [WSS15, WSYP18]. A recent method for saliency optimization from robust background detection [ZLWS14] uses superpixel segmentation as input to generate salient object detection results.…”
Section: Methodsmentioning
confidence: 99%
“…3, we provide the features visualization of the proposed network. As visible, the features progressively become discriminative (close to the final saliency map) which can effectively distinguish the foreground and background, such as the features in CU(0,3) and CU (1,3). In addition, one can find that the detail features (e.g., edges and textures) in the encoder path become more and more abstract with the downsampling, while the cluttered and noisy backgrounds gradually vanish with the nested connections and up-sampling in the decoder path.…”
Section: A Frameworkmentioning
confidence: 99%
“…where F (0,3) and F (1,3) are the features extracted by the convolution units CU (0,3) and CU (1,3) , respectively.…”
Section: Encoder-decoder Module With Nested Connectionsmentioning
confidence: 99%
“…Then automatically bootstraps an appearance model based on the initial foreground estimate, and uses it to refine the spatial accuracy of the segmentation and to also segment the object in frames where it does not move. The works [44], [45], [46] extend the concept of salient objects detection [47] as prior knowledge to infer the objects. Semi-supervised video segmentation, which also refers to label propagation, is usually achieved via propagating human annotation specified on one or a few key-frames onto the entire video sequence [48], [49], [50].…”
Section: Moving Object Segmentationmentioning
confidence: 99%