2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00315
|View full text |Cite
|
Sign up to set email alerts
|

Weakly Supervised Complementary Parts Models for Fine-Grained Image Classification From the Bottom Up

Abstract: Given a training dataset composed of images and corresponding category labels, deep convolutional neural networks show a strong ability in mining discriminative parts for image classification. However, deep convolutional neural networks trained with image level labels only tend to focus on the most discriminative parts while missing other object parts, which could provide complementary information. In this paper, we approach this problem from a different perspective. We build complementary parts models in a we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
200
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 280 publications
(200 citation statements)
references
References 45 publications
0
200
0
Order By: Relevance
“…Nowadays, researchers focus on how to get better categorization accuracy under the weakly-supervised setting. Some works [10,19,48] utilize a localization-classification network and some of them [28,49] learn a discriminative feature representation by end-to-end feature encoding. Different from these models which exploit only visual information, we leverage textual descriptions as well as mine the complementary information between texts and images, which make our proposed approach learn more discriminative attribute-related features and obtain a 3.7% higher accuracy than the best performing result.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Nowadays, researchers focus on how to get better categorization accuracy under the weakly-supervised setting. Some works [10,19,48] utilize a localization-classification network and some of them [28,49] learn a discriminative feature representation by end-to-end feature encoding. Different from these models which exploit only visual information, we leverage textual descriptions as well as mine the complementary information between texts and images, which make our proposed approach learn more discriminative attribute-related features and obtain a 3.7% higher accuracy than the best performing result.…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 99%
“…√ indicates YES and blank expresses NO. 82.8 Mask-CNN [44] √ 87.3 STN [19] 84.1 Bilinear-CNN [28] 84.1 Low-rank Bilinear [21] 84.2 RA-CNN [8] 85.3 HBP [49] 87.1 DFL-CNN [42] 87.4 NTS-Net [48] 87.5 DCL [6] 87.8 TASN [56] 87.9 WS-BAN [15] 88.8 Ge et al [10] 90.4 Triplet-A [7] √ Manual labour 80.7 Multi-grained [ [34] 89.5 SDR [1] 90.5 ResNet-50 [11] 92.4 RIIR [46] 94.0 NAC-CNN [38] 95.3 MGE-CNN [51] 95.9 PBC [16] 96.1 CVL [12] 96.2 TEB (ours) 96.6…”
Section: Comparison With State-of-the-art Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Performances of baseline methods are directly from Maximum-Entropy [32]. We also compared our proposed CSDL with state-of-the-art weakly supervised fine-grained methods, including BCNN [26], Kernel Pooling [30], MA-CNN [20], NTS-Net [24], DFL-CNN [25], iSQRT-COV [31], Part Model [56], TASN [39], DCL [40]. Performances of existing methods are all taken directly from their original papers.…”
Section: Baselinesmentioning
confidence: 99%
“…For image classification task, Deep Neural Network (DNN) has demonstrated the ability of learning representative features by many literatures [4]. It can get better classification accuracy when using more annotated samples.…”
Section: Introductionmentioning
confidence: 99%