2021
DOI: 10.1049/cvi2.12020
|View full text |Cite
|
Sign up to set email alerts
|

TanhExp: A smooth activation function with high convergence speed for lightweight neural networks

Abstract: Lightweight or mobile neural networks used for real‐time computer vision tasks contain fewer parameters than normal networks, which lead to a constrained performance. Herein, a novel activation function named as Tanh Exponential Activation Function (TanhExp) is proposed which can improve the performance for these networks on image classification task significantly. The definition of TanhExp is f(x) = x tanh(ex). The simplicity, efficiency, and robustness of TanhExp on various datasets and network models is dem… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 39 publications
(18 citation statements)
references
References 19 publications
0
18
0
Order By: Relevance
“…In a recent study by Liu et al [24], the authors proposed the Tanh exponential activation function (TanhExp), which improves the performance of lightweight or mobile neural networks used for real-time computer vision tasks, and contains fewer parameters than usual. It enhances the performance of these networks on image classification.…”
Section: Related Workmentioning
confidence: 99%
“…In a recent study by Liu et al [24], the authors proposed the Tanh exponential activation function (TanhExp), which improves the performance of lightweight or mobile neural networks used for real-time computer vision tasks, and contains fewer parameters than usual. It enhances the performance of these networks on image classification.…”
Section: Related Workmentioning
confidence: 99%
“…The author of [26] has reported unstable training behaviour for a specific function which can be obtained from TanhSoft-1, however, we tested and failed to find any such instability. Also, in [31] the authors have mentioned a special case which can be obtained from TanhSoft-2.…”
Section: Tanhsoft-1 Tanhsoft-2 and Tanhsoft-3 And Their Propertiesmentioning
confidence: 99%
“…Worth noting that the author has reported unstable training behaviour for the specific function f (x; 1, 0, γ, 1) in [20], however, we failed to find any instability during the training process. Also, in [21] the authors have mentioned the function f (x; 0, 1, 1, 0), which arise as an example from the TanhSoft family. In fact, we show that because of the introduction of hyper-parameters, better activation functions of the form f (x; 0, β, γ, 0) can be obtained.…”
Section: Tanhsoft Activation Function Familymentioning
confidence: 99%