2022
DOI: 10.1016/j.knosys.2022.109584
|View full text |Cite
|
Sign up to set email alerts
|

Muformer: A long sequence time-series forecasting model based on modified multi-head attention

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 25 publications
(5 citation statements)
references
References 22 publications
0
4
0
Order By: Relevance
“…3 , The multi-head attention mechanism achieves this goal by mapping the input sequence to a query (Q), key (K), and value (V) vector, respectively. Subsequently, it employs an attention-scoring function to compute the attentional weights of each position relative to the other positions, effectively capturing dependencies across the input sequence 47 , 48 . Finally, the output of each attention head is spliced together and then passed through a linear layer to obtain the final output..
Figure 3 Structure of the multi-head attention mechanism model.
…”
Section: Methodsmentioning
confidence: 99%
“…3 , The multi-head attention mechanism achieves this goal by mapping the input sequence to a query (Q), key (K), and value (V) vector, respectively. Subsequently, it employs an attention-scoring function to compute the attentional weights of each position relative to the other positions, effectively capturing dependencies across the input sequence 47 , 48 . Finally, the output of each attention head is spliced together and then passed through a linear layer to obtain the final output..
Figure 3 Structure of the multi-head attention mechanism model.
…”
Section: Methodsmentioning
confidence: 99%
“…The Transformer model proposed by Vaswani et al [20] has achieved tremendous success in natural language processing tasks, and Li et al [21] applied it to time series forecasting, addressing the issue of memory bottlenecks. However, traditional Transformer models exhibit high time and space complexity when dealing with long sequences, which limits their practical application in time series forecasting tasks [22] [23]. Therefore, in this paper, we propose a Patched Time Series Transformer model with independent channels to address this problem.…”
Section: Related Workmentioning
confidence: 99%
“…Time series forecasting plays a vital role in various domains such as finance (Ma et al 2022), weather forecasting (Liu et al 2022a), and sensor data analysis (Zhao et al 2023). Extracting meaningful patterns, understanding the underlying dynamics of time series to forecast future trends are crucial for informed decision-making and effective problemsolving (Zhang, Guo, and Wang 2023). With the advent of deep learning, convolutional neural networks (CNNs) (Fukushima 1980) and Transformers (Vaswani et al 2017) have shown remarkable progress in capturing temporal dependencies and extracting features from time series.…”
Section: Introductionmentioning
confidence: 99%