2023 IEEE International Conference on Control, Electronics and Computer Technology (ICCECT) 2023
DOI: 10.1109/iccect57938.2023.10140896
|View full text |Cite
|
Sign up to set email alerts
|

IBM Stock Forecast Using LSTM, GRU, Attention and Transformer Models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…Then the input gate controls the proportion of input that is allowed to enter the memory cell. In the final step, the output gate determines the value that gets produced [11].…”
Section: Long Short-term Memory (Lstm)mentioning
confidence: 99%
“…Then the input gate controls the proportion of input that is allowed to enter the memory cell. In the final step, the output gate determines the value that gets produced [11].…”
Section: Long Short-term Memory (Lstm)mentioning
confidence: 99%
“…Unlike RNN-based models, the Transformer leverages a multi-head self-Attention mechanism to grasp global positional relationships, thereby strengthening its ability to learn longterm dependencies. Fu et al conducted several experiments to evaluate the efficacy of the Transformer model in predicting IBM's stock price trend [11]. Ding et al introduced a new methodology utilizing the Transformer with several enhancements [1].…”
Section: Literature Reviewmentioning
confidence: 99%
“…Common methods include taking the dot product of their vectors, calculating the cosine similarity of their vectors, or employing additional neural networks for evaluation. In this study, we adopt the approach of first calculating the dot product of the two vectors, followed by applying the tanh activation function to obtain the similarity: (11) In the second stage, a calculation method Softmax normalization is introduced to transform the scores obtained in the first stage. On one hand, it allows normalization by organizing the original score calculations into a probability distribution where the sum of all element weights is equal to 1.…”
Section: Attentionmentioning
confidence: 99%