Procedings of the British Machine Vision Conference 2017 2017
DOI: 10.5244/c.31.67
|View full text |Cite
|
Sign up to set email alerts
|

Weakly Supervised Saliency Detection with A Category-Driven Map Generator

Abstract: Top-down saliency detection aims to highlight the regions of a specific object category, and typically relies on pixel-wise annotated training data. In this paper, we address the high cost of collecting such training data by presenting a weakly supervised approach to object saliency detection, where only image-level labels, indicating the presence or absence of a target object in an image, are available. The proposed framework is composed of two deep modules, an image-level classifier and a pixel-level map gen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(7 citation statements)
references
References 40 publications
0
7
0
Order By: Relevance
“…Having obtained the spatial saliency maps by selecting maximum per-pixel responses, we can then use these spatial heatmaps to learn or predict temporal saliency. More recently, Hsu et al [21] develop two coupled ConvNets, one for image-level classifier and the other for pixel-level generator. By designing a well-formulated loss function and top-down guidance from class labels, the generator is demonstrated to output saliency estimation of good quality.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Having obtained the spatial saliency maps by selecting maximum per-pixel responses, we can then use these spatial heatmaps to learn or predict temporal saliency. More recently, Hsu et al [21] develop two coupled ConvNets, one for image-level classifier and the other for pixel-level generator. By designing a well-formulated loss function and top-down guidance from class labels, the generator is demonstrated to output saliency estimation of good quality.…”
Section: Related Workmentioning
confidence: 99%
“…Temporal consistent loss. Inspired by [21,67,16] that model correlation between discrete images in an selfsupervised manner by per-pixel displacement warping, smoothness regularization, etc., we design 3 loss functions to train our model and refine O t by temporal constraints: temporal reconstruction loss L recons , smoothness loss L smooth , and motion masking loss L motion . The total loss function of each time step t can be formulated as:…”
Section: Temporal Modelmentioning
confidence: 99%
“…On the other hand, previous works [15], [40], [59]- [61] have already demonstrated that semantic information, especially Fig. 5: The detailed architecture of the proposed bi-stream network.…”
Section: B Which Training Set Should Be Selected?mentioning
confidence: 96%
“…Wang et al [34] introduced a foreground inference network to produce potential saliency maps using image-level labels. Hsu et al [10] presented a categorydriven map generator to learn saliency from image-level labels. Similarly, Li et al [16] adopted an iterative learning strategy to update an initial saliency map generated from unsupervised saliency methods by learning saliency from image-level supervision.…”
Section: Learning Saliency From Weak Annotationsmentioning
confidence: 99%