Proceedings of the 10th International Conference on Computer Vision Theory and Applications 2015
DOI: 10.5220/0005294900440053
|View full text |Cite
|
Sign up to set email alerts
|

Weakly Supervised Object Localization with Large Fisher Vectors

Abstract: We propose a novel method for learning object localization models in a weakly supervised manner, by employing images annotated with object class labels but not with object locations. Given an image, the learned model predicts both the presence of the object class in the image and the bounding box that determines the object location. The main ingredients of our method are a large Fisher vector representation and a sparse classification model enabling efficient evaluation of patch scores. The method is able to r… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
2
2
1

Relationship

3
2

Authors

Journals

citations
Cited by 6 publications
(7 citation statements)
references
References 28 publications
(52 reference statements)
0
7
0
Order By: Relevance
“…Here we require the classification score gradient in order to be able to identify the positive patches. We observe that the combination of the group-sparsity and spatial layout model achieves the best classification AP (around 81%), which is 4 pp better than the appearance-based counterpart (row 7) and more than 9 pp better than [21] (row 2). The group-sparse model identifies only 7 visual words, so there are only 49 possible pairs to consider in the spatial model.…”
Section: Methodsmentioning
confidence: 89%
See 4 more Smart Citations
“…Here we require the classification score gradient in order to be able to identify the positive patches. We observe that the combination of the group-sparsity and spatial layout model achieves the best classification AP (around 81%), which is 4 pp better than the appearance-based counterpart (row 7) and more than 9 pp better than [21] (row 2). The group-sparse model identifies only 7 visual words, so there are only 49 possible pairs to consider in the spatial model.…”
Section: Methodsmentioning
confidence: 89%
“…The results show that the group-sparse model outperforms the 2 regularized model for 7 percentage points (pp). In comparison with the 1 regularized model [21], the group-sparse model is 17 times more sparse and achieves comparable AP. This implies substantial performance advantage in terms of execution time.…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations