2022
DOI: 10.48550/arxiv.2211.14730
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Time Series is Worth 64 Words: Long-term Forecasting with Transformers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
75
0
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 39 publications
(76 citation statements)
references
References 0 publications
0
75
0
1
Order By: Relevance
“…The superior performance of ResNet is not surprising as the results are in agreement with [16]. The success of Transformer is also expected, as Transformers have achieved state-of-the-art performance in natural language processing [4,20,28], computer vision [7,12,13], and other time series problems [21,31,39]. On the other hand, the performance of RNN-based models (LSTM and GRU) is worse than simple baselines like ED and DTW.…”
Section: Methodsmentioning
confidence: 63%
“…The superior performance of ResNet is not surprising as the results are in agreement with [16]. The success of Transformer is also expected, as Transformers have achieved state-of-the-art performance in natural language processing [4,20,28], computer vision [7,12,13], and other time series problems [21,31,39]. On the other hand, the performance of RNN-based models (LSTM and GRU) is worse than simple baselines like ED and DTW.…”
Section: Methodsmentioning
confidence: 63%
“…The longterm patterns cannot be solved simply by extending the length of the lookback window because the long-and short-term repetitive patterns of PM temperature change do not have a certain regularity. Furthermore, increasing the lookback window's duration requires using more memory and processing power [29]. This paper suggests a convolutional neural network skip (CNN-skip) layer to capture the long-and shortterm local repetitive patterns in the temperature change of PMs by interval sampling to solve this problem.…”
Section: The Main Workmentioning
confidence: 99%
“…As for the time series analysis, TST (Zerveas et al, 2021) directly adopts the canonical masked modeling paradigm, which is learning to predict the removed time points based on the remaining time points. Afterward, PatchTST (Nie et al, 2022) learns to predict the masked subseries-level patches to capture the local semantic information and reduce memory usage. However, as we stated before, directly masking time series will ruin the essential temporal variations, making the reconstruction too difficult to guide the representation learning.…”
Section: Related Workmentioning
confidence: 99%