2018
DOI: 10.1016/j.neunet.2017.12.012
|View full text |Cite
|
Sign up to set email alerts
|

Sigmoid-weighted linear units for neural network function approximation in reinforcement learning

Abstract: In recent years, neural networks have enjoyed a renaissance as function approximators in reinforcement learning. Two decades after Tesauro's TD-Gammon achieved near top-level human performance in backgammon, the deep reinforcement learning algorithm DQN achieved human-level performance in many Atari 2600 games. The purpose of this study is twofold. First, we propose two activation functions for neural network function approximation in reinforcement learning: the sigmoid-weighted linear unit (SiLU) and its deri… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
550
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 1,084 publications
(552 citation statements)
references
References 15 publications
2
550
0
Order By: Relevance
“…For MobileNetV3, we use a combination of these layers as building blocks in order to build the most effective models. Layers are also upgraded with modified swish nonlinearities [36,13,16]. Both squeeze and excitation as well as the swish nonlinearity use the sigmoid which can be inefficient to compute as well challenging to maintain accuracy in fixed point arithmetic so we replace this with the hard sigmoid [2,11] as discussed in section 5.2. .…”
Section: Efficient Mobile Building Blocksmentioning
confidence: 99%
See 1 more Smart Citation
“…For MobileNetV3, we use a combination of these layers as building blocks in order to build the most effective models. Layers are also upgraded with modified swish nonlinearities [36,13,16]. Both squeeze and excitation as well as the swish nonlinearity use the sigmoid which can be inefficient to compute as well challenging to maintain accuracy in fixed point arithmetic so we replace this with the hard sigmoid [2,11] as discussed in section 5.2. .…”
Section: Efficient Mobile Building Blocksmentioning
confidence: 99%
“…In [36,13,16] a nonlinearity called swish was introduced that when used as a drop-in replacement for ReLU, that significantly improves the accuracy of neural networks. The nonlinearity is defined as…”
Section: Nonlinearitiesmentioning
confidence: 99%
“…Regarding activation, four algorithms were tested: ReLU [40], LeakyReLU [40], PReLU [41], and trainable Swish [42]. Swish is a recent algorithm similar to the sigmoid-weighted linear unit proposed in [43], but with a trainable parameter. Regarding convolutional layer initialization, two algorithms were tested, the so-called Glorot-normal (a.k.a., Xavier-normal, [44]) and the He-normal [41].…”
Section: Hyperparameter Tuning and Model Testingmentioning
confidence: 99%
“…E. Swish [19]: Swish was introduced to deep learning particularly in image classification and machine translation tasks by Google Brain team in 2017. In fact, it was similar to Sigmoid-weighted Linear Unit (SiL) [35] function which was used in reinforcement learning. It has the smooth property similar to Softplus.…”
Section: Existing Activation Functions For Comparisonmentioning
confidence: 99%