2019
DOI: 10.3390/app9091939
|View full text |Cite
|
Sign up to set email alerts
|

Two-Level Attentions and Grouping Attention Convolutional Network for Fine-Grained Image Classification

Abstract: The focus of fine-grained image classification tasks is to ignore interference information and grasp local features. This challenge is what the visual attention mechanism excels at. Firstly, we have constructed a two-level attention convolutional network, which characterizes the object-level attention and the pixel-level attention. Then, we combine the two kinds of attention through a second-order response transform algorithm. Furthermore, we propose a clustering-based grouping attention model, which implies t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
8
2

Relationship

2
8

Authors

Journals

citations
Cited by 18 publications
(5 citation statements)
references
References 44 publications
0
5
0
Order By: Relevance
“…While acquiring global features, the design generates special statistics-locationrelation descriptors to better describe the local features of parts and to generate special fine-grained part representation. In the overall structure, the hierachical CNN denoting skip-connections CNN (S-CCNN) [79] directly uses the feature maps of the first three levels of the convolutional layers and the last layer to combine and introduces a special module of "skip connection" for adjustment and classification. Two-level attentions and grouping attention CNN (TGA-CNN) [80] created a two-level attention model and a group attention model, which fuses the twostream features of pixel-level and object-level attention and merged the features with high similarity after channel feature processing to express the semantics better.…”
Section: Multi-stream Attentionmentioning
confidence: 99%
“…While acquiring global features, the design generates special statistics-locationrelation descriptors to better describe the local features of parts and to generate special fine-grained part representation. In the overall structure, the hierachical CNN denoting skip-connections CNN (S-CCNN) [79] directly uses the feature maps of the first three levels of the convolutional layers and the last layer to combine and introduces a special module of "skip connection" for adjustment and classification. Two-level attentions and grouping attention CNN (TGA-CNN) [80] created a two-level attention model and a group attention model, which fuses the twostream features of pixel-level and object-level attention and merged the features with high similarity after channel feature processing to express the semantics better.…”
Section: Multi-stream Attentionmentioning
confidence: 99%
“…e attention mechanism is a deep learning technology that originated from the study of human vision and has been widely used in natural language processing [26,27], recommendation systems, and image classification [28,29]. It mimics the characteristics of the human visual system that selectively focuses on the salient parts, and improves the efficiency of the model by dynamically selecting important features.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…SCA-CNN [31] introduces spatial attention and channel attention into the convolutional structure and uses them in image annotation tasks. GA-CNN [32] obtains part-level attention by performing similarity clustering and grouping convolution on feature chan-nels. The channel attention and spatial attention mentioned above have been widely used in various image tasks and achieved good results.…”
Section: Related Work a Attention Convolutional Modelmentioning
confidence: 99%