2020
DOI: 10.48550/arxiv.2002.12478
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Time Series Data Augmentation for Deep Learning: A Survey

Qingsong Wen,
Liang Sun,
Fan Yang
et al.

Abstract: Deep learning performs remarkably well on many time series analysis tasks recently. The superior performance of deep neural networks relies heavily on a large number of training data to avoid overfitting. However, the labeled data of many realworld time series applications may be limited such as classification in medical time series and anomaly detection in AIOps. As an effective way to enhance the size and quality of the training data, data augmentation is crucial to the successful application of deep learnin… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
112
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 92 publications
(113 citation statements)
references
References 7 publications
0
112
0
1
Order By: Relevance
“…Take the green line as the evaluation criterion, the data point will be misclassified as an anomalous sample with a high probability. Data augmentation can reduce or even eliminate the impact of such noisy data by increasing the size and quality of the data [26,8]. In our model, we use the methods such as window slicing and STL.…”
Section: Data Augmentationmentioning
confidence: 99%
See 1 more Smart Citation
“…Take the green line as the evaluation criterion, the data point will be misclassified as an anomalous sample with a high probability. Data augmentation can reduce or even eliminate the impact of such noisy data by increasing the size and quality of the data [26,8]. In our model, we use the methods such as window slicing and STL.…”
Section: Data Augmentationmentioning
confidence: 99%
“…The current machine learning models on BGP anomaly detection are based on the dataset constructed from the records of BGP update traffic packets, which are partitioned into "anomalous" and "normal" samples. To improve the anomaly detection performance, these models mainly focus on the following two dimensions [8]: feature dimension (to choose appropriate features) and time dimension (to choose appropriate models). For traditional machine learning methods (like SVM), the data (i.e., the features) at different timestamps are considered to be independent samples [9,10,11] in which the time correlations are totally ignored.…”
Section: Introductionmentioning
confidence: 99%
“…Unlike other augmentation methods such as cropping or rotation, time-series data is mainly augmented by using specific methods like window warping [35], flipping, Fourier transform [36], and down-sampling. Among them, a simple method of adding Gaussian noise is often used [37]. This data augmentation technique improves performance in time-series prediction models such as DeepAR [38].…”
Section: Data Augmentationmentioning
confidence: 99%
“…Further dataset details in Appendix F. ECG Data Augmentations. To augment each ECG for SimCLR (example in Appendix F, Figure 6), we apply three transformations in turn (based on prior work in time series augmentation [30,66]):…”
Section: Problem Setupmentioning
confidence: 99%