2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015
DOI: 10.1109/cvpr.2015.7298987
|View full text |Cite
|
Sign up to set email alerts
|

Semantic object segmentation via detection in weakly labeled video

Abstract: Semantic object segmentation in video is an important step for large-scale multimedia analysis. In many cases, however, semantic objects are only tagged at video-level, making them difficult to be located and segmented. To address this problem, this paper proposes an approach to segment semantic objects in weakly labeled video via object detection. In our approach, a novel video segmentationby-detection framework is proposed, which first incorporates object and region detectors pre-trained on still images to g… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
62
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 67 publications
(62 citation statements)
references
References 28 publications
(68 reference statements)
0
62
0
Order By: Relevance
“…In the same setting of multiple foreground vs. single background, several methods have proposed to rely on additional supervision. For instance, [63] relied on the CPMC [8] region detector, which has been trained from pixel-level annotations, to segment foreground from background. In [58] and [15], object proposal methods trained from pixel-level and bounding box annotations, respectively, were employed.…”
Section: Roadmentioning
confidence: 99%
“…In the same setting of multiple foreground vs. single background, several methods have proposed to rely on additional supervision. For instance, [63] relied on the CPMC [8] region detector, which has been trained from pixel-level annotations, to segment foreground from background. In [58] and [15], object proposal methods trained from pixel-level and bounding box annotations, respectively, were employed.…”
Section: Roadmentioning
confidence: 99%
“…We validate our proposed method on the Youtube-Object-Dataset (Prest et al, 2012;Jain and Grauman, 2014;Zhang et al, 2015). In experiments our model greatly improved the pre-trained model by mitigating its drawback even when we do not use the weak labels.…”
Section: Introductionmentioning
confidence: 92%
“…However, these methods rely on the quality of generated segment proposals and may produce inaccurate results when taking low-quality segments as the input. Zhang et al [44] propose to utilize object detectors integrated with object proposals to refine segmentations in videos. Furthermore, Tsai et al [40] develop a co-segmentation framework by linking object tracklets from all the videos and improve the result.…”
Section: Related Workmentioning
confidence: 99%
“…Compared to the baseline FCN model [20] used in our algorithm, there is a performance gain of 9%. In addition, while existing methods rely on training the segment classifier [34], integrating object proposals with detectors [44], co-segmentation via modeling relationships between videos [40], or self-paced fine-tuning [42], the proposed method utilizes a self-learning scheme to achieve better segmentation results. With the ResNet-101 architecture, we compare our method with DeepLab [2] and FSEG [12].…”
Section: Youtube-objects Datasetmentioning
confidence: 99%