“…For this data, the input is scalar x t k ∈ R, i.e., the input size m = 1, and the output d t k ∈ R, where d t k = x t k+1 . For the parameter selection, we perform a grid search on the number of hidden neurons and learning rate in the intervals q = [3,20] and η = [10 −3 , 10 −6 ], respectively. For the window size of the classical LSTM architecture with the preprocessing method, we search on the interval [( max /2), max ], where max equals to 10, 20, and 50 ms, respectively.…”