Loading…

Approximate dynamic programming for stochastic linear control problems on compact state spaces

•Proof of existence of optimal policies for stochastic linear control problems.•Proof of existence of convex bounded solutions to Average Cost Optimality Equation.•New ADP algorithm providing good policies and lower bounds to the optimal costs.•Tests of algorithm on multiple sourcing where optimalit...

Full description

Saved in:
Bibliographic Details
Published in:European journal of operational research 2015-02, Vol.241 (1), p.85-98
Main Authors: Woerner, Stefan, Laumanns, Marco, Zenklusen, Rico, Fertis, Apostolos
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
cited_by cdi_FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3
cites cdi_FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3
container_end_page 98
container_issue 1
container_start_page 85
container_title European journal of operational research
container_volume 241
creator Woerner, Stefan
Laumanns, Marco
Zenklusen, Rico
Fertis, Apostolos
description •Proof of existence of optimal policies for stochastic linear control problems.•Proof of existence of convex bounded solutions to Average Cost Optimality Equation.•New ADP algorithm providing good policies and lower bounds to the optimal costs.•Tests of algorithm on multiple sourcing where optimality gap can be bounded by 5%.•On all tested instances algorithm performs better than state-of-the-art heuristic. This paper addresses Markov Decision Processes over compact state and action spaces. We investigate the special case of linear dynamics and piecewise-linear and convex immediate costs for the average cost criterion. This model is very general and covers many interesting examples, for instance in inventory management. Due to the curse of dimensionality, the problem is intractable and optimal policies usually cannot be computed, not even for instances of moderate size. We show the existence of optimal policies and of convex and bounded relative value functions that solve the average cost optimality equation under reasonable and easy-to-check assumptions. Based on these insights, we propose an approximate relative value iteration algorithm based on piecewise-linear convex relative value function approximations. Besides computing good policies, the algorithm also provides lower bounds to the optimal average cost, which allow us to bound the optimality gap of any given policy for a given instance. The algorithm is applied to the well-studied Multiple Sourcing Problem as known from inventory management. Multiple sourcing is known to be a hard problem and usually tackled by parametric heuristics. We analyze several MSP instances with two and more suppliers and compare our results to state-of-the-art heuristics. For the considered scenarios, our policies are always at least as good as the best known heuristic, and strictly better in most cases. Moreover, by using the computed lower bounds we show for all instances that the optimality gap has never exceeded 5%, and that it has been much smaller for most of them.
doi_str_mv 10.1016/j.ejor.2014.08.003
format article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_1615270931</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0377221714006195</els_id><sourcerecordid>3468950131</sourcerecordid><originalsourceid>FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3</originalsourceid><addsrcrecordid>eNp9UE1LxDAQDaLg-vEHPBU8t06SNm3By7L4BQte9GpI0una0jY1yYr7701Zz55mmHnvzZtHyA2FjAIVd32GvXUZA5pnUGUA_ISsaFWyVFQCTskKeFmmjNHynFx43wMALWixIh_reXb2pxtVwKQ5TGrsTBInO6fGsZt2SWtd4oM1n8qHuBq6CZVLjJ2Cs8OC1AOOPrFTnI2zMiGiFy0fe_RX5KxVg8frv3pJ3h8f3jbP6fb16WWz3qYmr-uQNkzTUosCORO5znNkBee61aAqzQxjBbYCDRecagpVDahaBhUaobhuuEJ-SW6PutHQ1x59kL3duymelFTQgpVQcxpR7IgyznrvsJWzi5-7g6QglxxlL5cc5ZKjhErGHCPp_kjC6P-7Qye96XAy2HQOTZCN7f6j_wIo4n3d</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1615270931</pqid></control><display><type>article</type><title>Approximate dynamic programming for stochastic linear control problems on compact state spaces</title><source>ScienceDirect Journals</source><creator>Woerner, Stefan ; Laumanns, Marco ; Zenklusen, Rico ; Fertis, Apostolos</creator><creatorcontrib>Woerner, Stefan ; Laumanns, Marco ; Zenklusen, Rico ; Fertis, Apostolos</creatorcontrib><description>•Proof of existence of optimal policies for stochastic linear control problems.•Proof of existence of convex bounded solutions to Average Cost Optimality Equation.•New ADP algorithm providing good policies and lower bounds to the optimal costs.•Tests of algorithm on multiple sourcing where optimality gap can be bounded by 5%.•On all tested instances algorithm performs better than state-of-the-art heuristic. This paper addresses Markov Decision Processes over compact state and action spaces. We investigate the special case of linear dynamics and piecewise-linear and convex immediate costs for the average cost criterion. This model is very general and covers many interesting examples, for instance in inventory management. Due to the curse of dimensionality, the problem is intractable and optimal policies usually cannot be computed, not even for instances of moderate size. We show the existence of optimal policies and of convex and bounded relative value functions that solve the average cost optimality equation under reasonable and easy-to-check assumptions. Based on these insights, we propose an approximate relative value iteration algorithm based on piecewise-linear convex relative value function approximations. Besides computing good policies, the algorithm also provides lower bounds to the optimal average cost, which allow us to bound the optimality gap of any given policy for a given instance. The algorithm is applied to the well-studied Multiple Sourcing Problem as known from inventory management. Multiple sourcing is known to be a hard problem and usually tackled by parametric heuristics. We analyze several MSP instances with two and more suppliers and compare our results to state-of-the-art heuristics. For the considered scenarios, our policies are always at least as good as the best known heuristic, and strictly better in most cases. Moreover, by using the computed lower bounds we show for all instances that the optimality gap has never exceeded 5%, and that it has been much smaller for most of them.</description><identifier>ISSN: 0377-2217</identifier><identifier>EISSN: 1872-6860</identifier><identifier>DOI: 10.1016/j.ejor.2014.08.003</identifier><identifier>CODEN: EJORDT</identifier><language>eng</language><publisher>Amsterdam: Elsevier B.V</publisher><subject>Algorithms ; Decision making models ; Dual sourcing ; Dynamic programming ; Heuristic ; Inventory ; Inventory management ; Markov analysis ; Markov processes ; Multiple sourcing ; Optimization algorithms ; Studies</subject><ispartof>European journal of operational research, 2015-02, Vol.241 (1), p.85-98</ispartof><rights>2014 Elsevier B.V.</rights><rights>Copyright Elsevier Sequoia S.A. Feb 16, 2015</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3</citedby><cites>FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,27924,27925</link.rule.ids></links><search><creatorcontrib>Woerner, Stefan</creatorcontrib><creatorcontrib>Laumanns, Marco</creatorcontrib><creatorcontrib>Zenklusen, Rico</creatorcontrib><creatorcontrib>Fertis, Apostolos</creatorcontrib><title>Approximate dynamic programming for stochastic linear control problems on compact state spaces</title><title>European journal of operational research</title><description>•Proof of existence of optimal policies for stochastic linear control problems.•Proof of existence of convex bounded solutions to Average Cost Optimality Equation.•New ADP algorithm providing good policies and lower bounds to the optimal costs.•Tests of algorithm on multiple sourcing where optimality gap can be bounded by 5%.•On all tested instances algorithm performs better than state-of-the-art heuristic. This paper addresses Markov Decision Processes over compact state and action spaces. We investigate the special case of linear dynamics and piecewise-linear and convex immediate costs for the average cost criterion. This model is very general and covers many interesting examples, for instance in inventory management. Due to the curse of dimensionality, the problem is intractable and optimal policies usually cannot be computed, not even for instances of moderate size. We show the existence of optimal policies and of convex and bounded relative value functions that solve the average cost optimality equation under reasonable and easy-to-check assumptions. Based on these insights, we propose an approximate relative value iteration algorithm based on piecewise-linear convex relative value function approximations. Besides computing good policies, the algorithm also provides lower bounds to the optimal average cost, which allow us to bound the optimality gap of any given policy for a given instance. The algorithm is applied to the well-studied Multiple Sourcing Problem as known from inventory management. Multiple sourcing is known to be a hard problem and usually tackled by parametric heuristics. We analyze several MSP instances with two and more suppliers and compare our results to state-of-the-art heuristics. For the considered scenarios, our policies are always at least as good as the best known heuristic, and strictly better in most cases. Moreover, by using the computed lower bounds we show for all instances that the optimality gap has never exceeded 5%, and that it has been much smaller for most of them.</description><subject>Algorithms</subject><subject>Decision making models</subject><subject>Dual sourcing</subject><subject>Dynamic programming</subject><subject>Heuristic</subject><subject>Inventory</subject><subject>Inventory management</subject><subject>Markov analysis</subject><subject>Markov processes</subject><subject>Multiple sourcing</subject><subject>Optimization algorithms</subject><subject>Studies</subject><issn>0377-2217</issn><issn>1872-6860</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2015</creationdate><recordtype>article</recordtype><recordid>eNp9UE1LxDAQDaLg-vEHPBU8t06SNm3By7L4BQte9GpI0una0jY1yYr7701Zz55mmHnvzZtHyA2FjAIVd32GvXUZA5pnUGUA_ISsaFWyVFQCTskKeFmmjNHynFx43wMALWixIh_reXb2pxtVwKQ5TGrsTBInO6fGsZt2SWtd4oM1n8qHuBq6CZVLjJ2Cs8OC1AOOPrFTnI2zMiGiFy0fe_RX5KxVg8frv3pJ3h8f3jbP6fb16WWz3qYmr-uQNkzTUosCORO5znNkBee61aAqzQxjBbYCDRecagpVDahaBhUaobhuuEJ-SW6PutHQ1x59kL3duymelFTQgpVQcxpR7IgyznrvsJWzi5-7g6QglxxlL5cc5ZKjhErGHCPp_kjC6P-7Qye96XAy2HQOTZCN7f6j_wIo4n3d</recordid><startdate>20150216</startdate><enddate>20150216</enddate><creator>Woerner, Stefan</creator><creator>Laumanns, Marco</creator><creator>Zenklusen, Rico</creator><creator>Fertis, Apostolos</creator><general>Elsevier B.V</general><general>Elsevier Sequoia S.A</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7TB</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20150216</creationdate><title>Approximate dynamic programming for stochastic linear control problems on compact state spaces</title><author>Woerner, Stefan ; Laumanns, Marco ; Zenklusen, Rico ; Fertis, Apostolos</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2015</creationdate><topic>Algorithms</topic><topic>Decision making models</topic><topic>Dual sourcing</topic><topic>Dynamic programming</topic><topic>Heuristic</topic><topic>Inventory</topic><topic>Inventory management</topic><topic>Markov analysis</topic><topic>Markov processes</topic><topic>Multiple sourcing</topic><topic>Optimization algorithms</topic><topic>Studies</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Woerner, Stefan</creatorcontrib><creatorcontrib>Laumanns, Marco</creatorcontrib><creatorcontrib>Zenklusen, Rico</creatorcontrib><creatorcontrib>Fertis, Apostolos</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>European journal of operational research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Woerner, Stefan</au><au>Laumanns, Marco</au><au>Zenklusen, Rico</au><au>Fertis, Apostolos</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Approximate dynamic programming for stochastic linear control problems on compact state spaces</atitle><jtitle>European journal of operational research</jtitle><date>2015-02-16</date><risdate>2015</risdate><volume>241</volume><issue>1</issue><spage>85</spage><epage>98</epage><pages>85-98</pages><issn>0377-2217</issn><eissn>1872-6860</eissn><coden>EJORDT</coden><abstract>•Proof of existence of optimal policies for stochastic linear control problems.•Proof of existence of convex bounded solutions to Average Cost Optimality Equation.•New ADP algorithm providing good policies and lower bounds to the optimal costs.•Tests of algorithm on multiple sourcing where optimality gap can be bounded by 5%.•On all tested instances algorithm performs better than state-of-the-art heuristic. This paper addresses Markov Decision Processes over compact state and action spaces. We investigate the special case of linear dynamics and piecewise-linear and convex immediate costs for the average cost criterion. This model is very general and covers many interesting examples, for instance in inventory management. Due to the curse of dimensionality, the problem is intractable and optimal policies usually cannot be computed, not even for instances of moderate size. We show the existence of optimal policies and of convex and bounded relative value functions that solve the average cost optimality equation under reasonable and easy-to-check assumptions. Based on these insights, we propose an approximate relative value iteration algorithm based on piecewise-linear convex relative value function approximations. Besides computing good policies, the algorithm also provides lower bounds to the optimal average cost, which allow us to bound the optimality gap of any given policy for a given instance. The algorithm is applied to the well-studied Multiple Sourcing Problem as known from inventory management. Multiple sourcing is known to be a hard problem and usually tackled by parametric heuristics. We analyze several MSP instances with two and more suppliers and compare our results to state-of-the-art heuristics. For the considered scenarios, our policies are always at least as good as the best known heuristic, and strictly better in most cases. Moreover, by using the computed lower bounds we show for all instances that the optimality gap has never exceeded 5%, and that it has been much smaller for most of them.</abstract><cop>Amsterdam</cop><pub>Elsevier B.V</pub><doi>10.1016/j.ejor.2014.08.003</doi><tpages>14</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0377-2217
ispartof European journal of operational research, 2015-02, Vol.241 (1), p.85-98
issn 0377-2217
1872-6860
language eng
recordid cdi_proquest_journals_1615270931
source ScienceDirect Journals
subjects Algorithms
Decision making models
Dual sourcing
Dynamic programming
Heuristic
Inventory
Inventory management
Markov analysis
Markov processes
Multiple sourcing
Optimization algorithms
Studies
title Approximate dynamic programming for stochastic linear control problems on compact state spaces
url http://sfxeu10.hosted.exlibrisgroup.com/loughborough?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-01T03%3A10%3A10IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Approximate%20dynamic%20programming%20for%20stochastic%20linear%20control%20problems%20on%20compact%20state%20spaces&rft.jtitle=European%20journal%20of%20operational%20research&rft.au=Woerner,%20Stefan&rft.date=2015-02-16&rft.volume=241&rft.issue=1&rft.spage=85&rft.epage=98&rft.pages=85-98&rft.issn=0377-2217&rft.eissn=1872-6860&rft.coden=EJORDT&rft_id=info:doi/10.1016/j.ejor.2014.08.003&rft_dat=%3Cproquest_cross%3E3468950131%3C/proquest_cross%3E%3Cgrp_id%3Ecdi_FETCH-LOGICAL-c499t-d2b17b65e3264b44e2533bfb0a8b2c225ef6ec3631b10890eaf208ec6a3bd3ae3%3C/grp_id%3E%3Coa%3E%3C/oa%3E%3Curl%3E%3C/url%3E&rft_id=info:oai/&rft_pqid=1615270931&rft_id=info:pmid/&rfr_iscdi=true