1997
DOI: 10.1016/s0020-0255(96)00200-9
|View full text |Cite
|
Sign up to set email alerts
|

The generalized sigmoid activation function: Competitive supervised learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
60
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 149 publications
(60 citation statements)
references
References 1 publication
0
60
0
Order By: Relevance
“…The MLP units take a number of real-valued inputs and generate a single real-valued output, according to an activation function (transfer function) applied to the weighted sum of the outputs of the units in the preceding layer. The most commonly used activation function in this network is a sigmoid function 40. The learning algorithm can be expressed using generalized Delta rule and back propagation gradient descent 41.…”
Section: Methodsmentioning
confidence: 99%
“…The MLP units take a number of real-valued inputs and generate a single real-valued output, according to an activation function (transfer function) applied to the weighted sum of the outputs of the units in the preceding layer. The most commonly used activation function in this network is a sigmoid function 40. The learning algorithm can be expressed using generalized Delta rule and back propagation gradient descent 41.…”
Section: Methodsmentioning
confidence: 99%
“…Both contained nodes to help capture nonlinearity in the input data, and an output layer, which contained a node to represent a dependent variable (VTA occurrence) 19,20 . We used rectified linear unit (RELU) 21 activation functions for the hidden layers, and the sigmoid activation function 22 for the output. The two hidden layers consisted of 22 neurons each.…”
Section: Sdmentioning
confidence: 99%
“…Hence, this layer enables seamless interaction between different convolutional groups that extract patterns from different representations and facilitates joint feature learning from multiple information sources during back-propagation [44]. These complex non-linear features are then passed as inputs to a dense layer with softmax activation function [45], which draws a linear decision boundary on the derived feature space for separating the anticancer peptides from peptides without anticancer activity. Figure 1 represents the architecture of our proposed model for joint feature extraction from multiple information sources.…”
Section: Multi-headed Convolutional Neural Network Architecturementioning
confidence: 99%