2021
DOI: 10.1007/s10489-021-02247-z
|View full text |Cite
|
Sign up to set email alerts
|

PSigmoid: Improving squeeze-and-excitation block with parametric sigmoid

Abstract: Squeeze-and-Excitation (SE) Networks won the last ImageNet Large-Scale Visual Recognition Challenge (ILSVRC) classification competition and is very popular in today's vision community. The SE block is the core of Squeeze-and-Excitation Network (SENet), which adaptively recalibrates channel-wise features and suppresses less useful ones. Since SE blocks can be directly used in existing models and effectively improve performance, SE blocks are widely used in a variety of tasks. In this paper, we propose a novel P… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(5 citation statements)
references
References 38 publications
0
5
0
Order By: Relevance
“…The σ and δ are activation functions. W 1 ∈ R f × f r , and W 2 ∈ f r × f represent two fully connected layers, where the r is reduction ratio to control capacity and computational cost [43].…”
Section: Excitationmentioning
confidence: 99%
“…The σ and δ are activation functions. W 1 ∈ R f × f r , and W 2 ∈ f r × f represent two fully connected layers, where the r is reduction ratio to control capacity and computational cost [43].…”
Section: Excitationmentioning
confidence: 99%
“…An adaptive variant of logistic sigmoid named parametric sigmoid (psigmoid) 57 was proposed in [463,464]. 58 Similarly as in generalized hyperbolic tangent, it introduces two scaling parameters to a logistic sigmoid:…”
Section: Parametric Sigmoid (Psigmoid)mentioning
confidence: 99%
“…where a i is a trainable parameter for each neuron or channel i and b is a global trainable parameter [463].…”
Section: Parametric Sigmoid (Psigmoid)mentioning
confidence: 99%
See 1 more Smart Citation
“…In the denoising task, each noise point is given weight, the low weight noise points are removed automatically, and the high weight noise points are retained. During this process, the network running efficiency can be improved, the parameters and computational cost can be reduced, and the recognition accuracy is improved [31]. As shown in Figure 4, by processing the feature map of convolutional, a one-dimensional vector with the same number of channels is obtained as the evaluation score of each channel [32], and then, the score is used for the corresponding channel to get the result.…”
mentioning
confidence: 99%