2020
DOI: 10.48550/arxiv.2007.12792
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Generative Models that Solve PDEs: Distributed Computing for Training Large Data-Free Models

Abstract: Recent progress in scientific machine learning (SciML) has opened up the possibility of training novel neural network architectures that solve complex partial differential equations (PDEs). Several (nearly data free) approaches have been recently reported that successfully solve PDEs, with examples including deep feed forward networks, generative networks, and deep encoder-decoder networks. However, practical adoption of these approaches is limited by the difficulty in training these models, especially to make… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
3

Relationship

2
1

Authors

Journals

citations
Cited by 3 publications
(11 citation statements)
references
References 19 publications
0
11
0
Order By: Relevance
“…One of the key outcomes of our experiments was to demonstrate a practical approach to train MGDiffNet on domain sizes up to 512 3 . We applied our framework to train MGDiffNet for resolutions up to 256 3 on GPU-based HPC clusters using on-demand multi-GPU virtual machines on Microsoft Azure.…”
Section: Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…One of the key outcomes of our experiments was to demonstrate a practical approach to train MGDiffNet on domain sizes up to 512 3 . We applied our framework to train MGDiffNet for resolutions up to 256 3 on GPU-based HPC clusters using on-demand multi-GPU virtual machines on Microsoft Azure.…”
Section: Resultsmentioning
confidence: 99%
“…We applied our framework to train MGDiffNet for resolutions up to 256 3 on GPU-based HPC clusters using on-demand multi-GPU virtual machines on Microsoft Azure. To train DiffNet for resolutions > 256 3 we used PSC Bridges2 HPC cluster with baremetal access to CPU nodes. We first talk about our experiments to study the multigrid approach and then the scaling studies using distributed deep learning.…”
Section: Resultsmentioning
confidence: 99%
See 3 more Smart Citations