2020
DOI: 10.48550/arxiv.2002.12530
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Temporal Convolutional Attention-based Network For Sequence Modeling

Abstract: With the development of feed-forward models, the default model for sequence modeling has gradually evolved to replace recurrent networks. Many powerful feed-forward models based on convolutional networks and attention mechanisms were proposed and show more potential to handle sequence modeling tasks. We wonder that is there an architecture that can not only achieve an approximate substitution of recurrent networks but also absorb the advantages of feed-forward models. So we propose an exploratory architecture … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
20
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(20 citation statements)
references
References 6 publications
0
20
0
Order By: Relevance
“…Also, deep learning models [31,38] are rising as the new state-of-the-arts for many time series forecasting tasks. Many of them use the sequenceto-sequence neural structures [13,48] based on recurrent neural networks [20,26,43], Convolutional neural networks [2,18], transformers [29,53,57], and graph neural networks [54]. Deep learning models are also efficient in modeling multivariate time series [59].…”
Section: Time Series Forecastingmentioning
confidence: 99%
“…Also, deep learning models [31,38] are rising as the new state-of-the-arts for many time series forecasting tasks. Many of them use the sequenceto-sequence neural structures [13,48] based on recurrent neural networks [20,26,43], Convolutional neural networks [2,18], transformers [29,53,57], and graph neural networks [54]. Deep learning models are also efficient in modeling multivariate time series [59].…”
Section: Time Series Forecastingmentioning
confidence: 99%
“…Our explorations into the domain of sequence modeling to generate these forecasts, i.e., time-series predictions, have taken us from statistical engines to multi-layer perceptrons (MLPs) to recursive models [26], [27].…”
Section: Related Workmentioning
confidence: 99%
“…Most successful neural networks are based on the encoder-decoder architectures [36,10,12,33,3,11,20,21], namely Seq2seq. Basically, various Seq2seq models based on RNNs [16,32,37,18,41,22], CNNs [4,15], and Transformers (self-attentions) [19,38] are proposed to model the non-linearity for time series.…”
Section: Related Workmentioning
confidence: 99%
“…We show that as we can identify the ground truth components of the mixture, and the ground truth assignment of each microscopic observation, independent modeling of time series data from each component could be improved due to lower variance, and further benefitting the estimation of macroscopic time series that are of interest. Second, inspired by recent successes of Seq2seq models [36,10,12] based on deep neural networks, e.g., variants of recurrent neural networks (RNNs) [16,41,22], convolutional neural networks (CNNs) [4,15], and Transformers [19,38], we propose Mixture of Seq2seq (MixSeq), a mixture model for time series, where the components come from a family of Seq2seq models parameterized by different parameters. Third, we conduct synthetic experiments to demonstrate the superiority of our approach, and extensive experiments on real-world data to show the power of our approach compared with canonical approaches.…”
Section: Introductionmentioning
confidence: 99%