Loading…

Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation

Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature...

Full description

Saved in:
Bibliographic Details
Published in:Computers & fluids 2025-02, Vol.288, p.106498, Article 106498
Main Authors: Nista, Ludovico, Schumann, Christoph D.K., Petkov, Peicho, Pavlov, Valentin, Grenga, Temistocle, MacArt, Jonathan F., Attili, Antonio, Markov, Stoyan, Pitsch, Heinz
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites cdi_FETCH-LOGICAL-c240t-569f610b23f39e5cca5c2025e8c557fc0744f84c9fbaaf6679ff4d56e6bae2d53
container_end_page
container_issue
container_start_page 106498
container_title Computers & fluids
container_volume 288
creator Nista, Ludovico
Schumann, Christoph D.K.
Petkov, Peicho
Pavlov, Valentin
Grenga, Temistocle
MacArt, Jonathan F.
Attili, Antonio
Markov, Stoyan
Pitsch, Heinz
description Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator’s learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy. •Accelerated SR-GAN training for turbulence closure using distributed data-parallelism.•Accurate a priori turbulence predictions require training on large subboxes.•Key factors in SR-GAN training: batch size, learning rate, subboxes, and its discriminator.•Integration of LES solvers with SR-GAN inference through SuperLES library is introduced.•Provided guidelines for GAN-based SR training and SR-LES closure.
doi_str_mv 10.1016/j.compfluid.2024.106498
format article
fullrecord <record><control><sourceid>elsevier_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1016_j_compfluid_2024_106498</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0045793024003293</els_id><sourcerecordid>S0045793024003293</sourcerecordid><originalsourceid>FETCH-LOGICAL-c240t-569f610b23f39e5cca5c2025e8c557fc0744f84c9fbaaf6679ff4d56e6bae2d53</originalsourceid><addsrcrecordid>eNqFkE1OwzAQhb0AiVI4A75AipPYTrOsKv6kSrCAteXY48rFiaNxUtQLcG7SFrFlNZrRe2_0PkLucrbIWS7vdwsT296F0dtFwQo-XSWvlxdkxhgXWVWX7Ipcp7Rj014WfEa-3zTqECBQ3_YBWugGPfjYUd1Z2gO6iK3uDNDoaBqnQ4aQYhhPmi10gJN8D1TbPWDS6HWgHQxfET_pMGIzBji622ghJDql0aBxCxlYe6DJt2M4vbshl06HBLe_c04-Hh_e18_Z5vXpZb3aZKbgbMiErJ3MWVOUrqxBGKOFmXoKWBohKmdYxblbclO7RmsnZVU7x62QIBsNhRXlnFTnXIMxJQSnevStxoPKmToiVDv1h1AdEaozwsm5OjunHrD3gCoZf-xmPYIZlI3-34wf7feGfg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation</title><source>ScienceDirect Freedom Collection</source><creator>Nista, Ludovico ; Schumann, Christoph D.K. ; Petkov, Peicho ; Pavlov, Valentin ; Grenga, Temistocle ; MacArt, Jonathan F. ; Attili, Antonio ; Markov, Stoyan ; Pitsch, Heinz</creator><creatorcontrib>Nista, Ludovico ; Schumann, Christoph D.K. ; Petkov, Peicho ; Pavlov, Valentin ; Grenga, Temistocle ; MacArt, Jonathan F. ; Attili, Antonio ; Markov, Stoyan ; Pitsch, Heinz</creatorcontrib><description>Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator’s learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy. •Accelerated SR-GAN training for turbulence closure using distributed data-parallelism.•Accurate a priori turbulence predictions require training on large subboxes.•Key factors in SR-GAN training: batch size, learning rate, subboxes, and its discriminator.•Integration of LES solvers with SR-GAN inference through SuperLES library is introduced.•Provided guidelines for GAN-based SR training and SR-LES closure.</description><identifier>ISSN: 0045-7930</identifier><identifier>DOI: 10.1016/j.compfluid.2024.106498</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>High-performance computing ; Inference-coupled large-eddy simulations ; Super-resolution generative adversarial networks ; Synchronous data-parallel distributed training ; Turbulence closure modeling</subject><ispartof>Computers &amp; fluids, 2025-02, Vol.288, p.106498, Article 106498</ispartof><rights>2024 The Authors</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c240t-569f610b23f39e5cca5c2025e8c557fc0744f84c9fbaaf6679ff4d56e6bae2d53</cites><orcidid>0009-0002-5786-9834 ; 0000-0003-2771-8832 ; 0000-0001-5656-0961 ; 0000-0001-6796-4072 ; 0000-0002-9465-9505 ; 0000-0002-0420-9480 ; 0000-0002-3254-3232</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Nista, Ludovico</creatorcontrib><creatorcontrib>Schumann, Christoph D.K.</creatorcontrib><creatorcontrib>Petkov, Peicho</creatorcontrib><creatorcontrib>Pavlov, Valentin</creatorcontrib><creatorcontrib>Grenga, Temistocle</creatorcontrib><creatorcontrib>MacArt, Jonathan F.</creatorcontrib><creatorcontrib>Attili, Antonio</creatorcontrib><creatorcontrib>Markov, Stoyan</creatorcontrib><creatorcontrib>Pitsch, Heinz</creatorcontrib><title>Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation</title><title>Computers &amp; fluids</title><description>Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator’s learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy. •Accelerated SR-GAN training for turbulence closure using distributed data-parallelism.•Accurate a priori turbulence predictions require training on large subboxes.•Key factors in SR-GAN training: batch size, learning rate, subboxes, and its discriminator.•Integration of LES solvers with SR-GAN inference through SuperLES library is introduced.•Provided guidelines for GAN-based SR training and SR-LES closure.</description><subject>High-performance computing</subject><subject>Inference-coupled large-eddy simulations</subject><subject>Super-resolution generative adversarial networks</subject><subject>Synchronous data-parallel distributed training</subject><subject>Turbulence closure modeling</subject><issn>0045-7930</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2025</creationdate><recordtype>article</recordtype><recordid>eNqFkE1OwzAQhb0AiVI4A75AipPYTrOsKv6kSrCAteXY48rFiaNxUtQLcG7SFrFlNZrRe2_0PkLucrbIWS7vdwsT296F0dtFwQo-XSWvlxdkxhgXWVWX7Ipcp7Rj014WfEa-3zTqECBQ3_YBWugGPfjYUd1Z2gO6iK3uDNDoaBqnQ4aQYhhPmi10gJN8D1TbPWDS6HWgHQxfET_pMGIzBji622ghJDql0aBxCxlYe6DJt2M4vbshl06HBLe_c04-Hh_e18_Z5vXpZb3aZKbgbMiErJ3MWVOUrqxBGKOFmXoKWBohKmdYxblbclO7RmsnZVU7x62QIBsNhRXlnFTnXIMxJQSnevStxoPKmToiVDv1h1AdEaozwsm5OjunHrD3gCoZf-xmPYIZlI3-34wf7feGfg</recordid><startdate>20250215</startdate><enddate>20250215</enddate><creator>Nista, Ludovico</creator><creator>Schumann, Christoph D.K.</creator><creator>Petkov, Peicho</creator><creator>Pavlov, Valentin</creator><creator>Grenga, Temistocle</creator><creator>MacArt, Jonathan F.</creator><creator>Attili, Antonio</creator><creator>Markov, Stoyan</creator><creator>Pitsch, Heinz</creator><general>Elsevier Ltd</general><scope>6I.</scope><scope>AAFTH</scope><scope>AAYXX</scope><scope>CITATION</scope><orcidid>https://orcid.org/0009-0002-5786-9834</orcidid><orcidid>https://orcid.org/0000-0003-2771-8832</orcidid><orcidid>https://orcid.org/0000-0001-5656-0961</orcidid><orcidid>https://orcid.org/0000-0001-6796-4072</orcidid><orcidid>https://orcid.org/0000-0002-9465-9505</orcidid><orcidid>https://orcid.org/0000-0002-0420-9480</orcidid><orcidid>https://orcid.org/0000-0002-3254-3232</orcidid></search><sort><creationdate>20250215</creationdate><title>Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation</title><author>Nista, Ludovico ; Schumann, Christoph D.K. ; Petkov, Peicho ; Pavlov, Valentin ; Grenga, Temistocle ; MacArt, Jonathan F. ; Attili, Antonio ; Markov, Stoyan ; Pitsch, Heinz</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c240t-569f610b23f39e5cca5c2025e8c557fc0744f84c9fbaaf6679ff4d56e6bae2d53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2025</creationdate><topic>High-performance computing</topic><topic>Inference-coupled large-eddy simulations</topic><topic>Super-resolution generative adversarial networks</topic><topic>Synchronous data-parallel distributed training</topic><topic>Turbulence closure modeling</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Nista, Ludovico</creatorcontrib><creatorcontrib>Schumann, Christoph D.K.</creatorcontrib><creatorcontrib>Petkov, Peicho</creatorcontrib><creatorcontrib>Pavlov, Valentin</creatorcontrib><creatorcontrib>Grenga, Temistocle</creatorcontrib><creatorcontrib>MacArt, Jonathan F.</creatorcontrib><creatorcontrib>Attili, Antonio</creatorcontrib><creatorcontrib>Markov, Stoyan</creatorcontrib><creatorcontrib>Pitsch, Heinz</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>CrossRef</collection><jtitle>Computers &amp; fluids</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nista, Ludovico</au><au>Schumann, Christoph D.K.</au><au>Petkov, Peicho</au><au>Pavlov, Valentin</au><au>Grenga, Temistocle</au><au>MacArt, Jonathan F.</au><au>Attili, Antonio</au><au>Markov, Stoyan</au><au>Pitsch, Heinz</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation</atitle><jtitle>Computers &amp; fluids</jtitle><date>2025-02-15</date><risdate>2025</risdate><volume>288</volume><spage>106498</spage><pages>106498-</pages><artnum>106498</artnum><issn>0045-7930</issn><abstract>Super-resolution (SR) generative adversarial networks (GANs) are promising for turbulence closure in large-eddy simulation (LES) due to their ability to accurately reconstruct high-resolution data from low-resolution fields. Current model training and inference strategies are not sufficiently mature for large-scale, distributed calculations due to the computational demands and often unstable training of SR-GANs, which limits the exploration of improved model structures, training strategies, and loss-function definitions. Integrating SR-GANs into LES solvers for inference-coupled simulations is also necessary to assess their a posteriori accuracy, stability, and cost. We investigate parallelization strategies for SR-GAN training and inference-coupled LES, focusing on computational performance and reconstruction accuracy. We examine distributed data-parallel training strategies for hybrid CPU–GPU node architectures and the associated influence of low-/high-resolution subbox size, global batch size, and discriminator accuracy. Accurate predictions require training subboxes that are sufficiently large relative to the Kolmogorov length scale. Care should be placed on the coupled effect of training batch size, learning rate, number of training subboxes, and discriminator’s learning capabilities. We introduce a data-parallel SR-GAN training and inference library for heterogeneous architectures that enables exchange between the LES solver and SR-GAN inference at runtime. We investigate the predictive accuracy and computational performance of this arrangement with particular focus on the overlap (halo) size required for accurate SR reconstruction. Similarly, a posteriori parallel scaling for efficient inference-coupled LES is constrained by the SR subdomain size, GPU utilization, and reconstruction accuracy. Based on these findings, we establish guidelines and best practices to optimize resource utilization and parallel acceleration of SR-GAN turbulence model training and inference-coupled LES calculations while maintaining predictive accuracy. •Accelerated SR-GAN training for turbulence closure using distributed data-parallelism.•Accurate a priori turbulence predictions require training on large subboxes.•Key factors in SR-GAN training: batch size, learning rate, subboxes, and its discriminator.•Integration of LES solvers with SR-GAN inference through SuperLES library is introduced.•Provided guidelines for GAN-based SR training and SR-LES closure.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.compfluid.2024.106498</doi><orcidid>https://orcid.org/0009-0002-5786-9834</orcidid><orcidid>https://orcid.org/0000-0003-2771-8832</orcidid><orcidid>https://orcid.org/0000-0001-5656-0961</orcidid><orcidid>https://orcid.org/0000-0001-6796-4072</orcidid><orcidid>https://orcid.org/0000-0002-9465-9505</orcidid><orcidid>https://orcid.org/0000-0002-0420-9480</orcidid><orcidid>https://orcid.org/0000-0002-3254-3232</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0045-7930
ispartof Computers & fluids, 2025-02, Vol.288, p.106498, Article 106498
issn 0045-7930
language eng
recordid cdi_crossref_primary_10_1016_j_compfluid_2024_106498
source ScienceDirect Freedom Collection
subjects High-performance computing
Inference-coupled large-eddy simulations
Super-resolution generative adversarial networks
Synchronous data-parallel distributed training
Turbulence closure modeling
title Parallel implementation and performance of super-resolution generative adversarial network turbulence models for large-eddy simulation
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-01T14%3A41%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-elsevier_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Parallel%20implementation%20and%20performance%20of%20super-resolution%20generative%20adversarial%20network%20turbulence%20models%20for%20large-eddy%20simulation&rft.jtitle=Computers%20&%20fluids&rft.au=Nista,%20Ludovico&rft.date=2025-02-15&rft.volume=288&rft.spage=106498&rft.pages=106498-&rft.artnum=106498&rft.issn=0045-7930&rft_id=info:doi/10.1016/j.compfluid.2024.106498&rft_dat=%3Celsevier_cross%3ES0045793024003293%3C/elsevier_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c240t-569f610b23f39e5cca5c2025e8c557fc0744f84c9fbaaf6679ff4d56e6bae2d53%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true