2016 International Conference on Inventive Computation Technologies (ICICT) 2016
DOI: 10.1109/inventive.2016.7824854
|View full text |Cite
|
Sign up to set email alerts
|

Simulating a pipelined RISC processor

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 4 publications
0
5
0
Order By: Relevance
“…(e) Activation functions: The authors of [167] explore the effect of different activation functions on image classification results. They note that CNNs perform better than machine learning techniques because of their multi-layer hierarchical feature extraction that is controlled by variables such as the number of hidden layers, activation functions, learning rates, initial weights, and decay functions, however, they attributed the non-linearity of the network only to the activation function, which motivates their comparative investigation, regarding under-researched problems including (a) vanishing and exploding gradients during back-propagation, (b) zero-mean and range of outputs, (c) computational complexity of activation functions and (d) predictive performance of the model.…”
Section: Properties Of Resnet-18mentioning
confidence: 99%
See 1 more Smart Citation
“…(e) Activation functions: The authors of [167] explore the effect of different activation functions on image classification results. They note that CNNs perform better than machine learning techniques because of their multi-layer hierarchical feature extraction that is controlled by variables such as the number of hidden layers, activation functions, learning rates, initial weights, and decay functions, however, they attributed the non-linearity of the network only to the activation function, which motivates their comparative investigation, regarding under-researched problems including (a) vanishing and exploding gradients during back-propagation, (b) zero-mean and range of outputs, (c) computational complexity of activation functions and (d) predictive performance of the model.…”
Section: Properties Of Resnet-18mentioning
confidence: 99%
“…While deep convolutional architectures had been shown to provide state-of-the-art performance on standard image classification benchmarks such as the ImageNet data set [39][40][41], it was quickly discovered that, in a somewhat counter-intuitive fashion, deeper networks only led to increased performance up to a point, after which increased network depth resulted in increasingly worse performance. This was due to what is known as the vanishing gradients problem [167,168]. In essence, the deeper a network becomes, the smaller the derivative used to adjust model weights becomes during backpropagation.…”
Section: Selection Of Resnet-18 As the Cnn Structure For The Work In ...mentioning
confidence: 99%
“…(e) Activation functions: [170] explore the effect of different activation functions on image classification results. They note that CNNs perform better than machine learning techniques because of their multi-layer hierarchical feature extraction which is controlled by variables such as number of hidden layers, activation functions, learning rates, initial weights, and decay functions, however, they attribute non-linearity of the network only to the activation function which motivates their comparative investigation, regarding under-researched problems including (a) vanishing and exploding gradients during back-propagation, (b) zero-mean and range of outputs, (c) compute complexity of activation functions and (d) predictive performance of the model.…”
Section: Properties Of Resnet-18mentioning
confidence: 99%
“…While deep convolutional architectures had been shown to provide state of the art performance on standard image classification benchmarks such as the ImageNet dataset [39][40][41], it was quickly discovered that, in a somewhat counter-intuitive fashion, deeper networks only led to increased performance up to a point, after which increased network depth resulted in increasingly worse performance. This was due to what is known as the vanishing gradients problem [170,171]. In essence, the deeper a network becomes, the smaller the derivative used to adjust model weights becomes during backpropagation.…”
Section: Applications Of Resnetsmentioning
confidence: 99%
“…This was due to what is known as the vanishing gradients problem (Pandey & Srivastava, 2023;Ruder, 2016). In essence, the deeper a network becomes, the smaller the derivative used to adjust model weights becomes during backpropagation.…”
Section: Applications Of Resnetsmentioning
confidence: 99%