2022
DOI: 10.1609/aaai.v36i2.20126
|View full text |Cite
|
Sign up to set email alerts
|

Fully Attentional Network for Semantic Segmentation

Abstract: Recent non-local self-attention methods have proven to be effective in capturing long-range dependencies for semantic segmentation. These methods usually form a similarity map of R^(CxC) (by compressing spatial dimensions) or R^(HWxHW) (by compressing channels) to describe the feature relations along either channel or spatial dimensions, where C is the number of channels, H and W are the spatial dimensions of the input feature map. However, such practices tend to condense feature dependencies along the other d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 38 publications
(16 citation statements)
references
References 24 publications
0
14
0
Order By: Relevance
“…In addition, choices of the pooling size combination are investigated. It is observed in Table 4 that the pooling size combination (4,8,12) outperforms (1,2,4) and (2,4,8), which makes it our settlement in MFA.…”
Section: Ablation Study For Hyperparametersmentioning
confidence: 90%
See 1 more Smart Citation
“…In addition, choices of the pooling size combination are investigated. It is observed in Table 4 that the pooling size combination (4,8,12) outperforms (1,2,4) and (2,4,8), which makes it our settlement in MFA.…”
Section: Ablation Study For Hyperparametersmentioning
confidence: 90%
“…EMANet 11 iterates a set of compact bases using expectation maximization algorithm to represent the whole image and runs attention calculation on this set of bases, so that the computational complexity can be reduced significantly. FLANet 12 encodes both channel and spatial attentions in one single attention map, which not only considers all the information but also reduces the computational consumption. Even though these modules have reduced the computing cost and have improved the segmentation performance, they ignore the fusion of multi-resolution features.…”
Section: Introductionmentioning
confidence: 99%
“…Inspired by the Fully Attentional Block proposed in the literature 15 and Progressive Sampling proposed in the literature, 29 the FAPS module is proposed, shown in Figures 2 and 3. The basic idea of this module is to save the feature response from the global background under the same spatial position of horizontal and vertical coordinates and then use the self‐attention mechanism to capture the fully attention similarity between the two channel maps and their spatial positions.…”
Section: Methodsmentioning
confidence: 99%
“…Similarly, in non‐local spatial attention, a situation arises in which the action between each channel dimension is missing. Based on this, the literature 15 proposed Fully Attention Block (FLA), which uses global contextual features to preserve spatial response features when computing the channel attention map, which enables full attention in a single attention and improves computational efficiency. First, the feature response of the global context is captured at each spatial location.…”
Section: Introductionmentioning
confidence: 99%
“…EVALUATION OF SEGMENTATION RESULTS(%).To test the generality of the proposed feature consistency constraints, experiments on different network structures are conducted. In this experiment, four network structures, namely FCN, DeepLab V3+, PSPNet, and FLANet[56] are employed…”
mentioning
confidence: 99%