2024
DOI: 10.1029/2023gl106324
|View full text |Cite
|
Sign up to set email alerts
|

Explainable Offline‐Online Training of Neural Networks for Parameterizations: A 1D Gravity Wave‐QBO Testbed in the Small‐Data Regime

Hamid A. Pahlavan,
Pedram Hassanzadeh,
M. Joan Alexander

Abstract: There are different strategies for training neural networks (NNs) as subgrid‐scale parameterizations. Here, we use a 1D model of the quasi‐biennial oscillation (QBO) and gravity wave (GW) parameterizations as testbeds. A 12‐layer convolutional NN that predicts GW forcings for given wind profiles, when trained offline in a big‐data regime (100‐year), produces realistic QBOs once coupled to the 1D model. In contrast, offline training of this NN in a small‐data regime (18‐month) yields unrealistic QBOs. However, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5

Relationship

3
2

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 55 publications
0
7
0
Order By: Relevance
“…We acknowledge that good offline performance (at least in terms of common metrics such as R 2 ) is not a sufficient indicator for stable and accurate online (coupled to climate model) performance (Guan et al, 2022;Ross et al, 2022), although it is a necessary first step. More strict metrics, such as R 2 of the PDF tails, may better connect the offline and online performance (Pahlavan et al, 2024).…”
Section: Summary and Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…We acknowledge that good offline performance (at least in terms of common metrics such as R 2 ) is not a sufficient indicator for stable and accurate online (coupled to climate model) performance (Guan et al, 2022;Ross et al, 2022), although it is a necessary first step. More strict metrics, such as R 2 of the PDF tails, may better connect the offline and online performance (Pahlavan et al, 2024).…”
Section: Summary and Discussionmentioning
confidence: 99%
“…Recently, there has been a growing interest in developing data-driven SGS parameterizations for different complex processes in the Earth system using machine learning (ML) techniques, particularly deep neural networks (NNs). Promising results have been demonstrated in a wide range of idealized applications, including prototype systems (Chattopadhyay, Subel, & Hassanzadeh, 2020;Frezat et al, 2022;Gagne et al, 2020;Guan et al, 2022;Maulik et al, 2019;Pahlavan et al, 2024;Rasp, 2020), ocean turbulent processes (Bolton & Zanna, 2019;C. Zhang et al, 2023), moist convection in the atmosphere (Beucler et al, 2021;Brenowitz & Bretherton, 2019;Iglesias-Suarez et al, 2023;O'Gorman & Dwyer, 2018;Yuval & O'Gorman, 2020), radiation (Belochitski & Krasnopolsky, 2021;Krasnopolsky et al, 2005;Song & Roh, 2021), and microphysics (Gettelman et al, 2021;Seifert & Rasp, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…This has important implications for any data-driven SGS modeling approach, including those using deep neural networks or any other statistical learning method (Fatkullin & Vanden-Eijnden, 2004;Grooms et al, 2021;Guan et al, 2022;Sun et al, 2023;Zanna & Bolton, 2021). In fact, this non-uniqueness and uncertainty of the true SGS term is a major shortcoming of the supervised/offline learning approach to data-driven closure modeling in real-world applications, and is one of the main motivations to pursue online or at least offline-online learning approaches (Pahlavan et al, 2024;Schneider, Stuart, & Wu, 2021;Schneider et al, 2023).…”
Section: Journal Of Advances In Modeling Earth Systemsmentioning
confidence: 99%
“…While some of these studies found the learned data-driven SGS closures to lead to stable and accurate LES (Frezat et al, 2022;Guan et al, 2022Guan et al, , 2023Yuval & O'Gorman, 2020), a number of major challenges remain (Balaji, 2021;Schneider, Jeevanjee, & Socolow, 2021). Perhaps the most important one is interpretability, which is difficult for neural networks, despite some recent advances in explainable ML for climate-related applications (Clare et al, 2022;Mamalakis et al, 2022), including for SGS modeling (Pahlavan et al, 2024;Subel et al, 2023). The black-box nature of neural network-based closures aside, there are also challenges related to generalizability, computational cost, and even implementation (Balaji, 2021;Chattopadhyay et al, 2020;Guan et al, 2022;Kurz & Beck, 2020;Maulik et al, 2019;Subel et al, 2021;Xie et al, 2019;Zhou et al, 2019), limiting the broad application of such closures in operational climate and weather models, at least for now.…”
Section: Introductionmentioning
confidence: 99%
“…Then, for out-ofsample states (that are from the same distribution as those of the training), the ANN predicts the systematic model tendency correction needed to nudge the state of NWP or climate model, thus improving the trajectory and potentially the simulated statistics. While ANNs are powerfully expressive, they have a number of major shortcomings: (a) they are difficult to interpret, (b) they do not generalize to out-of-distribution, (c) they are datahungry, and (d) their predictions fed into numerical models can cause instabilities and unphysical drifts (Bretherton et al, 2022;Clark et al, 2022;Farchi et al, 2023;Guan et al, 2022;Pahlavan et al, 2024;Slater et al, 2023;Subel et al, 2023). Challenges with interpretability hinder understanding the root cause(s) of the model errors and fixing them.…”
Section: Introductionmentioning
confidence: 99%