2021
DOI: 10.1109/access.2021.3105355
|View full text |Cite
|
Sign up to set email alerts
|

TanhSoft—Dynamic Trainable Activation Functions for Faster Learning and Better Performance

Abstract: Deep learning, at its core, contains functions that are the composition of a linear transformation with a non-linear function known as activation function. In the past few years, there is an increasing interest in the construction of novel activation functions resulting in better learning. In this work, we propose three novel activation functions with learnable parameters, namely TanhSoft-1, TanhSoft-2, and TanhSoft-3, which are shown to outperform several well-known activation functions. For instance, replaci… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(4 citation statements)
references
References 30 publications
0
4
0
Order By: Relevance
“…The TanhSoft is a family of AAFs proposed in [461] that combine the softplus and tanh that contains three notable cases -TanhSoft-1, TanhSoft-2, and TanhSoft-3 [461,462].…”
Section: Tanhsoftmentioning
confidence: 99%
See 2 more Smart Citations
“…The TanhSoft is a family of AAFs proposed in [461] that combine the softplus and tanh that contains three notable cases -TanhSoft-1, TanhSoft-2, and TanhSoft-3 [461,462].…”
Section: Tanhsoftmentioning
confidence: 99%
“…where a i is a trainable parameter [461,462]; it can be obtained from the general TanhSoft by setting b i = 0 and d i = 1 [461]. The second AF from [462], TanhSoft-2, is defined as…”
Section: Tanhsoftmentioning
confidence: 99%
See 1 more Smart Citation
“…Venkatappareddy et al [38] proposed a polynomial-based activation function; Hegui et al [39] proposed a nonmonotonic activation function called Logish; both show high accuracy on the CIFAR-10 dataset. TanhSoft-1, TanhSoft-2, and TanhSoft-3 [40] are combinations of Tanh; the exponential terms also increase the computational complexity. ALKHOULY et al [41] proposed the IpLU and AbsLU without exponential terms to make it simple in computation.…”
Section: Introductionmentioning
confidence: 99%