Spatially Extended Tests of a Neural Network Parametrization Trained by Coarse-Graining
Abstract
General circulation models (GCMs) typically have a grid size of 25-200 km. Parametrizations are used to represent diabatic processes such as radiative transfer and cloud microphysics and account for subgrid-scale motions and variability. Unlike traditional approaches, neural networks (NNs) can readily exploit recent observational data sets and global cloud-system resolving model (CRM) simulations to learn subgrid variability. This article describes an NN parametrization trained by coarse-graining a near-global CRM simulation with a 4-km horizontal grid spacing. The NN predicts the residual heating and moistening averaged over (160 km)2 grid boxes as a function of the coarse-resolution fields within the same atmospheric column. This NN is coupled to the dynamical core of a GCM with the same 160-km resolution. A recent study described how to train such an NN to be stable when coupled to specified time-evolving advective forcings in a single-column model, but feedbacks between NN and GCM components cause spatially extended simulations to crash within a few days. Analyzing the linearized response of such an NN reveals that it learns to exploit a strong synchrony between precipitation and the atmospheric state above 10 km. Removing these variables from the NN's inputs stabilizes the coupled simulations, which predict the future state more accurately than a coarse-resolution simulation without any parametrizations of subgrid-scale variability, although the mean state slowly drifts.
- Publication:
-
Journal of Advances in Modeling Earth Systems
- Pub Date:
- August 2019
- DOI:
- 10.1029/2019MS001711
- arXiv:
- arXiv:1904.03327
- Bibcode:
- 2019JAMES..11.2728B
- Keywords:
-
- machine learning;
- global cloud-system resolving model;
- parameterization;
- Physics - Atmospheric and Oceanic Physics
- E-Print:
- doi:10.1029/2019MS001711