Loading…

Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology

A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimi...

Full description

Saved in:
Bibliographic Details
Published in:arXiv.org 2022-03
Main Authors: Nasir, Yusuf, Durlofsky, Louis J
Format: Article
Language:English
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by
cites
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Nasir, Yusuf
Durlofsky, Louis J
description A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.
doi_str_mv 10.48550/arxiv.2203.13375
format article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2644205979</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2644205979</sourcerecordid><originalsourceid>FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13</originalsourceid><addsrcrecordid>eNotjctqwzAUBUWh0JDmA7oTdO1UryvZy5I-IdBN9kGWr10HWUoluWn-voF2dWAY5hByx9la1QDswaaf8XstBJNrLqWBK7IQUvKqVkLckFXOB8aY0EYAyAWxT4hHmnAMfUwOJwyFerQpjGGgF0TjsYyT9fSE3lMXQ0nR0zHQPLd5Tr11SPM5F5wyPY3lk87BYSr2YgwYfRzOt-S6tz7j6n-XZPfyvNu8VduP1_fN47ayDUDlODLRNVLXwKGVjWudZbXqoDVONy32vVZOWbDaSu5AKtNBx5gGzqQxjssluf_LHlP8mjGX_SHOKVwe90IrJRg0ppG_We1X0Q</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2644205979</pqid></control><display><type>article</type><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><source>Publicly Available Content Database</source><creator>Nasir, Yusuf ; Durlofsky, Louis J</creator><creatorcontrib>Nasir, Yusuf ; Durlofsky, Louis J</creatorcontrib><description>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2203.13375</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Data assimilation ; Decision making ; Deep learning ; Flow mapping ; Geology ; Machine learning ; Markov processes ; Optimization ; Robustness ; Three dimensional models ; Two dimensional models ; Water injection</subject><ispartof>arXiv.org, 2022-03</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.proquest.com/docview/2644205979?pq-origsite=primo$$EHTML$$P50$$Gproquest$$Hfree_for_read</linktohtml><link.rule.ids>780,784,25752,27924,37011,44589</link.rule.ids></links><search><creatorcontrib>Nasir, Yusuf</creatorcontrib><creatorcontrib>Durlofsky, Louis J</creatorcontrib><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><title>arXiv.org</title><description>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</description><subject>Algorithms</subject><subject>Data assimilation</subject><subject>Decision making</subject><subject>Deep learning</subject><subject>Flow mapping</subject><subject>Geology</subject><subject>Machine learning</subject><subject>Markov processes</subject><subject>Optimization</subject><subject>Robustness</subject><subject>Three dimensional models</subject><subject>Two dimensional models</subject><subject>Water injection</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>PIMPY</sourceid><recordid>eNotjctqwzAUBUWh0JDmA7oTdO1UryvZy5I-IdBN9kGWr10HWUoluWn-voF2dWAY5hByx9la1QDswaaf8XstBJNrLqWBK7IQUvKqVkLckFXOB8aY0EYAyAWxT4hHmnAMfUwOJwyFerQpjGGgF0TjsYyT9fSE3lMXQ0nR0zHQPLd5Tr11SPM5F5wyPY3lk87BYSr2YgwYfRzOt-S6tz7j6n-XZPfyvNu8VduP1_fN47ayDUDlODLRNVLXwKGVjWudZbXqoDVONy32vVZOWbDaSu5AKtNBx5gGzqQxjssluf_LHlP8mjGX_SHOKVwe90IrJRg0ppG_We1X0Q</recordid><startdate>20220324</startdate><enddate>20220324</enddate><creator>Nasir, Yusuf</creator><creator>Durlofsky, Louis J</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20220324</creationdate><title>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</title><author>Nasir, Yusuf ; Durlofsky, Louis J</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Algorithms</topic><topic>Data assimilation</topic><topic>Decision making</topic><topic>Deep learning</topic><topic>Flow mapping</topic><topic>Geology</topic><topic>Machine learning</topic><topic>Markov processes</topic><topic>Optimization</topic><topic>Robustness</topic><topic>Three dimensional models</topic><topic>Two dimensional models</topic><topic>Water injection</topic><toplevel>online_resources</toplevel><creatorcontrib>Nasir, Yusuf</creatorcontrib><creatorcontrib>Durlofsky, Louis J</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nasir, Yusuf</au><au>Durlofsky, Louis J</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology</atitle><jtitle>arXiv.org</jtitle><date>2022-03-24</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>A general control policy framework based on deep reinforcement learning (DRL) is introduced for closed-loop decision making in subsurface flow settings. Traditional closed-loop modeling workflows in this context involve the repeated application of data assimilation/history matching and robust optimization steps. Data assimilation can be particularly challenging in cases where both the geological style (scenario) and individual model realizations are uncertain. The closed-loop reservoir management (CLRM) problem is formulated here as a partially observable Markov decision process, with the associated optimization problem solved using a proximal policy optimization algorithm. This provides a control policy that instantaneously maps flow data observed at wells (as are available in practice) to optimal well pressure settings. The policy is represented by a temporal convolution and gated transformer blocks. Training is performed in a preprocessing step with an ensemble of prior geological models, which can be drawn from multiple geological scenarios. Example cases involving the production of oil via water injection, with both 2D and 3D geological models, are presented. The DRL-based methodology is shown to result in an NPV increase of 15% (for the 2D cases) and 33% (3D cases) relative to robust optimization over prior models, and to an average improvement of 4% in NPV relative to traditional CLRM. The solutions from the control policy are found to be comparable to those from deterministic optimization, in which the geological model is assumed to be known, even when multiple geological scenarios are considered. The control policy approach results in a 76% decrease in computational cost relative to traditional CLRM with the algorithms and parameter settings considered in this work.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2203.13375</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-03
issn 2331-8422
language eng
recordid cdi_proquest_journals_2644205979
source Publicly Available Content Database
subjects Algorithms
Data assimilation
Decision making
Deep learning
Flow mapping
Geology
Machine learning
Markov processes
Optimization
Robustness
Three dimensional models
Two dimensional models
Water injection
title Deep reinforcement learning for optimal well control in subsurface systems with uncertain geology
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-11T01%3A15%3A18IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20reinforcement%20learning%20for%20optimal%20well%20control%20in%20subsurface%20systems%20with%20uncertain%20geology&rft.jtitle=arXiv.org&rft.au=Nasir,%20Yusuf&rft.date=2022-03-24&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2203.13375&rft_dat=%3Cproquest%3E2644205979%3C/proquest%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-a955-c1e02d9368515b39cbca084d5b7c69beff64c4a5a6a31c5347d5d006510377c13%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=2644205979&rft_id=info:pmid/&rfr_iscdi=true