Designing networks to accurately learn 2D turbulence closures
Abstract
Scientifically meaningful deployment of machine-learned subgrid closures in large-eddy simulations (LES) requires learned closures to be more accurate or faster to compute than existing closure models. Here we present a systematic study of the accuracy of neural LES closures for forced 2D turbulence as a function of the network architecture and hyperparameters. We examine statistically steady flows where we can control the location of the filtering scale with respect to the stationary spectrum, and include a range of architectures that allow us to distinguish the effects of nonlocality and finite-differencing errors in the closure accuracy. We consider fully-connected, convolutional, and u-net network architectures trained on filtered snapshots from highly resolved direct numerical simulations (DNS). We vary the breadth and depth of the networks as well as the selected input variables and cost functions used during training. We examine how these choices impact the accuracy of the learned closures in predicting true subgrid stresses from DNS, and how they affect the statistics of new coarse forward models / LES using the learned closures.
- Publication:
-
APS Division of Fluid Dynamics Meeting Abstracts
- Pub Date:
- 2020
- Bibcode:
- 2020APS..DFDS01030B