The University of Southampton
University of Southampton Institutional Repository

Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation

Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation
Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation

Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator's learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy.

High-performance computing, Inference-coupled large-eddy simulations, Super-resolution generative adversarial networks, Synchronous data-parallel distributed training, Turbulence closure modeling
0045-7930
Nista, Ludovico
a93303b2-3e96-484b-a8d1-cf11e1588814
Schumann, Christoph D.K.
4b993966-a0c3-438b-a2f0-0151e87d42cf
Petkov, Peicho
a50febee-289c-4a0a-9e47-2ebce543bbf4
Pavlov, Valentin
d0ac8484-b12b-44dd-a020-9f9352bddbba
Grenga, Temistocle
be0eba30-74b5-4134-87e7-3a2d6dd3836f
MacArt, Jonathan F.
1384a548-486e-4fae-9d5c-4177b0ed7825
Attili, Antonio
cd357d33-94e2-4a14-9aa0-9126a18687d0
Markov, Stoyan
697fb0c0-24a9-4060-beb3-d461483f0928
Pitsch, Heinz
3dc0eb6e-deca-4742-98a1-f0cdd62ff8b8
Nista, Ludovico
a93303b2-3e96-484b-a8d1-cf11e1588814
Schumann, Christoph D.K.
4b993966-a0c3-438b-a2f0-0151e87d42cf
Petkov, Peicho
a50febee-289c-4a0a-9e47-2ebce543bbf4
Pavlov, Valentin
d0ac8484-b12b-44dd-a020-9f9352bddbba
Grenga, Temistocle
be0eba30-74b5-4134-87e7-3a2d6dd3836f
MacArt, Jonathan F.
1384a548-486e-4fae-9d5c-4177b0ed7825
Attili, Antonio
cd357d33-94e2-4a14-9aa0-9126a18687d0
Markov, Stoyan
697fb0c0-24a9-4060-beb3-d461483f0928
Pitsch, Heinz
3dc0eb6e-deca-4742-98a1-f0cdd62ff8b8

Nista, Ludovico, Schumann, Christoph D.K., Petkov, Peicho, Pavlov, Valentin, Grenga, Temistocle, MacArt, Jonathan F., Attili, Antonio, Markov, Stoyan and Pitsch, Heinz (2025) Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation. Computers & Fluids, 288, [106498]. (doi:10.1016/j.compfluid.2024.106498).

Record type: Article

Abstract

Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator's learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy.

Text
1-s2.0-S0045793024003293-main - Version of Record
Download (2MB)

More information

Accepted/In Press date: 25 November 2024
e-pub ahead of print date: 2 December 2024
Published date: 15 February 2025
Additional Information: Publisher Copyright: © 2024 The Authors
Keywords: High-performance computing, Inference-coupled large-eddy simulations, Super-resolution generative adversarial networks, Synchronous data-parallel distributed training, Turbulence closure modeling

Identifiers

Local EPrints ID: 496828
URI: http://eprints.soton.ac.uk/id/eprint/496828
ISSN: 0045-7930
PURE UUID: 2160fc1a-153a-47be-9861-e43ebd6b4cdf
ORCID for Temistocle Grenga: ORCID iD orcid.org/0000-0002-9465-9505

Catalogue record

Date deposited: 08 Jan 2025 07:19
Last modified: 22 Aug 2025 02:38

Export record

Altmetrics

Contributors

Author: Ludovico Nista
Author: Christoph D.K. Schumann
Author: Peicho Petkov
Author: Valentin Pavlov
Author: Temistocle Grenga ORCID iD
Author: Jonathan F. MacArt
Author: Antonio Attili
Author: Stoyan Markov
Author: Heinz Pitsch

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×