2022
DOI: 10.1016/j.neunet.2021.11.022
|View full text |Cite
|
Sign up to set email alerts
|

Transformers for modeling physical systems

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
41
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 81 publications
(41 citation statements)
references
References 33 publications
0
41
0
Order By: Relevance
“…17 Transformer for physical system After the groundbreaking success in natural language processing, 29,56 attention has also been demonstrated as a promising tool for various other machine learning tasks. [57][58][59] Following the success in these fields, several previous works 17,18,[60][61][62] have explored using attention to model and simulate physical system, which can generally be divided into two orthogonal lines. In the first group, attention layers are used to capture the structures and patterns lie in the PDEs' spatial domain.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…17 Transformer for physical system After the groundbreaking success in natural language processing, 29,56 attention has also been demonstrated as a promising tool for various other machine learning tasks. [57][58][59] Following the success in these fields, several previous works 17,18,[60][61][62] have explored using attention to model and simulate physical system, which can generally be divided into two orthogonal lines. In the first group, attention layers are used to capture the structures and patterns lie in the PDEs' spatial domain.…”
Section: Related Workmentioning
confidence: 99%
“…17,18,62 On the contrary, in the second group, the attention is used to model the temporal evolution of the system while the spatial encoding is done by other mechanisms like CNNs or GNNs. 60,61 Our model lies in the former group where attention-based layers are used for encoding the spatial information of the input and query points, while the time marching is performed in the latent space using recurrent MLPs.…”
Section: Related Workmentioning
confidence: 99%
“…The population dynamics is thus learnt with the resulting embedding via a temporal transformer for the purpose of inference. There are many complex ways to create this embedding: [16] and [17] extracted embeddings from multivariate physical systems with Koopman operators before feeding the resulting representation into a temporal transformer; [18] used a graph neural network before transformers to embed interconnected-structures to perform skeleton-based action recognition; [19] used convolutional architecture to extract image embeddings before feeding them into transformer.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Multiple architectures have been proposed for using RNN for accomplishing the task future state prediction of a dynamical system. Recently Geneva & Zabara [44] and Eivazi et al [45] used LSTM and transformers as time integrator to predict flow evolution state. Hosseinyalamdary [46] used simple RNN for IMU modeling in deep Kalman filter.…”
Section: Recurrent Neural Networkmentioning
confidence: 99%