Loading…
Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology
A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimi...
Saved in:
Published in: | arXiv.org 2022-03 |
---|---|
Main Authors: | , |
Format: | Article |
Language: | English |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
cited_by | |
---|---|
cites | |
container_end_page | |
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Nasir, Yusuf Durlofsky, Louis J |
description | A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work. |
doi_str_mv | 10.48550/arxiv.2203.13375 |
format | article |
fullrecord | <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2644205979</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2644205979</sourcerecordid><originalsourceid>FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13</originalsourceid><addsrcrecordid>eNotjctqwzAUBUWh0JDmA7oTdO1UryvZy5I-IdBN9kGWr10HWUoluWn-voF2dWAY5hByx9la1QDswaaf8XstBJNrLqWBK7IQUvKqVkLckFXOB8aY0EYAyAWxT4hHmnAMfUwOJwyFerQpjGGgF0TjsYyT9fSE3lMXQ0nR0zHQPLd5Tr11SPM5F5wyPY3lk87BYSr2YgwYfRzOt-S6tz7j6n-XZPfyvNu8VduP1_fN47ayDUDlODLRNVLXwKGVjWudZbXqoDVONy32vVZOWbDaSu5AKtNBx5gGzqQxjssluf_LHlP8mjGX_SHOKVwe90IrJRg0ppG_We1X0Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2644205979</pqid></control><display><type>article</type><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><source>Publicly Available Content Database</source><creator>Nasir, Yusuf ; Durlofsky, Louis J</creator><creatorcontrib>Nasir, Yusuf ; Durlofsky, Louis J</creatorcontrib><description>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2203.13375</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Data assimilation ; Decision making ; Deep learning ; Flow mapping ; Geology ; Machine learning ; Markov processes ; Optimization ; Robustness ; Three dimensional models ; Two dimensional models ; Water injection</subject><ispartof>arXiv.org, 2022-03</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2644205979?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25752,27924,37011,44589</link.rule.ids></links><search><creatorcontrib>Nasir, Yusuf</creatorcontrib><creatorcontrib>Durlofsky, Louis J</creatorcontrib><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><title>arXiv.org</title><description>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</description><subject>Algorithms</subject><subject>Data assimilation</subject><subject>Decision making</subject><subject>Deep learning</subject><subject>Flow mapping</subject><subject>Geology</subject><subject>Machine learning</subject><subject>Markov processes</subject><subject>Optimization</subject><subject>Robustness</subject><subject>Three dimensional models</subject><subject>Two dimensional models</subject><subject>Water injection</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjctqwzAUBUWh0JDmA7oTdO1UryvZy5I-IdBN9kGWr10HWUoluWn-voF2dWAY5hByx9la1QDswaaf8XstBJNrLqWBK7IQUvKqVkLckFXOB8aY0EYAyAWxT4hHmnAMfUwOJwyFerQpjGGgF0TjsYyT9fSE3lMXQ0nR0zHQPLd5Tr11SPM5F5wyPY3lk87BYSr2YgwYfRzOt-S6tz7j6n-XZPfyvNu8VduP1_fN47ayDUDlODLRNVLXwKGVjWudZbXqoDVONy32vVZOWbDaSu5AKtNBx5gGzqQxjssluf_LHlP8mjGX_SHOKVwe90IrJRg0ppG_We1X0Q</recordid><startdate>20220324</startdate><enddate>20220324</enddate><creator>Nasir, Yusuf</creator><creator>Durlofsky, Louis J</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20220324</creationdate><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><author>Nasir, Yusuf ; Durlofsky, Louis J</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Data assimilation</topic><topic>Decision making</topic><topic>Deep learning</topic><topic>Flow mapping</topic><topic>Geology</topic><topic>Machine learning</topic><topic>Markov processes</topic><topic>Optimization</topic><topic>Robustness</topic><topic>Three dimensional models</topic><topic>Two dimensional models</topic><topic>Water injection</topic><toplevel>online_resources</toplevel><creatorcontrib>Nasir, Yusuf</creatorcontrib><creatorcontrib>Durlofsky, Louis J</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nasir, Yusuf</au><au>Durlofsky, Louis J</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</atitle><jtitle>arXiv.org</jtitle><date>2022-03-24</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2203.13375</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2022-03 |
issn | 2331-8422 |
language | eng |
recordid | cdi_proquest_journals_2644205979 |
source | Publicly Available Content Database |
subjects | Algorithms Data assimilation Decision making Deep learning Flow mapping Geology Machine learning Markov processes Optimization Robustness Three dimensional models Two dimensional models Water injection |
title | Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology |
url | http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-11T01%3A15%3A18IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20reinforcement%20learning%20for%20optimal%20well%20control%20in%20subsurface%20systems%20with%20uncertain%20geology&rft.jtitle=arXiv.org&rft.au=Nasir,%20Yusuf&rft.date=2022-03-24&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2203.13375&rft_dat=%3Cproquest%3E2644205979%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2644205979&rft_id=info:pmid/&rfr_iscdi=true |