Accelerated modified policy iteration algorithms for Markov decision processes
Author
Abstract
Suggested Citation
DOI: 10.1007/s00186-013-0432-y
Download full text from publisher
As the access to this document is restricted, you may want to search for a different version of it.
References listed on IDEAS
- Herzberg, Meir & Yechiali, Uri, 1996. "A K-step look-ahead analysis of value iteration algorithms for Markov decision processes," European Journal of Operational Research, Elsevier, vol. 88(3), pages 622-636, February.
- Meir Herzberg & Uri Yechiali, 1994. "Accelerating Procedures of the Value Iteration Algorithm for Discounted Markov Decision Processes, Based on a One-Step Lookahead Analysis," Operations Research, INFORMS, vol. 42(5), pages 940-946, October.
- Oleksandr Shlakhter & Chi-Guhn Lee & Dmitry Khmelev & Nasser Jaber, 2010. "Acceleration Operators in the Value Iteration Algorithms for Markov Decision Processes," Operations Research, INFORMS, vol. 58(1), pages 193-202, February.
- M. A. Pollatschek & B. Avi-Itzhak, 1969. "Algorithms for Stochastic Games with Geometrical Interpretation," Management Science, INFORMS, vol. 15(7), pages 399-415, March.
- Martin L. Puterman & Moon Chirl Shin, 1978. "Modified Policy Iteration Algorithms for Discounted Markov Decision Problems," Management Science, INFORMS, vol. 24(11), pages 1127-1137, July.
- D. P. de Farias & B. Van Roy, 2003. "The Linear Programming Approach to Approximate Dynamic Programming," Operations Research, INFORMS, vol. 51(6), pages 850-865, December.
Most related items
These are the items that most often cite the same works as this one and are cited by the same works as this one.- Oleksandr Shlakhter & Chi-Guhn Lee & Dmitry Khmelev & Nasser Jaber, 2010. "Acceleration Operators in the Value Iteration Algorithms for Markov Decision Processes," Operations Research, INFORMS, vol. 58(1), pages 193-202, February.
- Herzberg, Meir & Yechiali, Uri, 1996. "A K-step look-ahead analysis of value iteration algorithms for Markov decision processes," European Journal of Operational Research, Elsevier, vol. 88(3), pages 622-636, February.
- Thomas W. M. Vossen & Dan Zhang, 2015. "Reductions of Approximate Linear Programs for Network Revenue Management," Operations Research, INFORMS, vol. 63(6), pages 1352-1371, December.
- Novoa, Clara & Storer, Robert, 2009. "An approximate dynamic programming approach for the vehicle routing problem with stochastic demands," European Journal of Operational Research, Elsevier, vol. 196(2), pages 509-515, July.
- Mercedes Esteban-Bravo & Jose M. Vidal-Sanz & Gökhan Yildirim, 2014.
"Valuing Customer Portfolios with Endogenous Mass and Direct Marketing Interventions Using a Stochastic Dynamic Programming Decomposition,"
Marketing Science, INFORMS, vol. 33(5), pages 621-640, September.
- Vidal-Sanz, Jose M. & Yildirim, Gökhan, 2012. "Valuing customer portfolios with endogenous mass-and-direct-marketing interventions using a stochastic dynamic programming decomposition," DEE - Working Papers. Business Economics. WB wb121304, Universidad Carlos III de Madrid. Departamento de EconomÃa de la Empresa.
- Matthew S. Maxwell & Mateo Restrepo & Shane G. Henderson & Huseyin Topaloglu, 2010. "Approximate Dynamic Programming for Ambulance Redeployment," INFORMS Journal on Computing, INFORMS, vol. 22(2), pages 266-281, May.
- Sauré, Antoine & Patrick, Jonathan & Tyldesley, Scott & Puterman, Martin L., 2012. "Dynamic multi-appointment patient scheduling for radiation therapy," European Journal of Operational Research, Elsevier, vol. 223(2), pages 573-584.
- Cai, Yongyang & Judd, Kenneth L. & Lontzek, Thomas S. & Michelangeli, Valentina & Su, Che-Lin, 2017.
"A Nonlinear Programming Method For Dynamic Programming,"
Macroeconomic Dynamics, Cambridge University Press, vol. 21(2), pages 336-361, March.
- Yongyang Cai & Kenneth L. Judd & Thomas S. Lontzek & Valentina Michelangeli & Che-Lin Su, 2013. "Nonlinear Programming Method for Dynamic Programming," NBER Working Papers 19034, National Bureau of Economic Research, Inc.
- Yu, Yugang & Luo, Yifei & Shi, Ye, 2022. "Adoption of blockchain technology in a two-stage supply chain: Spillover effect on workforce," Transportation Research Part E: Logistics and Transportation Review, Elsevier, vol. 161(C).
- Vijay V. Desai & Vivek F. Farias & Ciamac C. Moallemi, 2012. "Pathwise Optimization for Optimal Stopping Problems," Management Science, INFORMS, vol. 58(12), pages 2292-2308, December.
- Jonathan Patrick & Martin L. Puterman & Maurice Queyranne, 2008. "Dynamic Multipriority Patient Scheduling for a Diagnostic Resource," Operations Research, INFORMS, vol. 56(6), pages 1507-1525, December.
- Meissner, Joern & Strauss, Arne, 2012.
"Network revenue management with inventory-sensitive bid prices and customer choice,"
European Journal of Operational Research, Elsevier, vol. 216(2), pages 459-468.
- Joern Meissner & Arne Strauss, 2008. "Network Revenue Management with Inventory-Sensitive Bid Prices and Customer Choice," Working Papers MRG/0008, Department of Management Science, Lancaster University, revised Apr 2010.
- Michael H. Veatch, 2013. "Approximate Linear Programming for Average Cost MDPs," Mathematics of Operations Research, INFORMS, vol. 38(3), pages 535-544, August.
- Daniel Adelman & Adam J. Mersereau, 2008. "Relaxations of Weakly Coupled Stochastic Dynamic Programs," Operations Research, INFORMS, vol. 56(3), pages 712-727, June.
- Oscar Dowson & Lea Kapelevich, 2021. "SDDP.jl : A Julia Package for Stochastic Dual Dynamic Programming," INFORMS Journal on Computing, INFORMS, vol. 33(1), pages 27-33, January.
- Saghafian, Soroush, 2018. "Ambiguous partially observable Markov decision processes: Structural results and applications," Journal of Economic Theory, Elsevier, vol. 178(C), pages 1-35.
- John N. Tsitsiklis, 2010. "Commentary ---Perspectives on Stochastic Optimization Over Time," INFORMS Journal on Computing, INFORMS, vol. 22(1), pages 18-19, February.
- J. G. Dai & Pengyi Shi, 2019. "Inpatient Overflow: An Approximate Dynamic Programming Approach," Manufacturing & Service Operations Management, INFORMS, vol. 21(4), pages 894-911, October.
- David L. Kaufman & Andrew J. Schaefer, 2013. "Robust Modified Policy Iteration," INFORMS Journal on Computing, INFORMS, vol. 25(3), pages 396-410, August.
- Garud N. Iyengar, 2005. "Robust Dynamic Programming," Mathematics of Operations Research, INFORMS, vol. 30(2), pages 257-280, May.
More about this item
Keywords
Markov decision processes; Policy iteration; Modified policy iteration; Accelerated convergence;All these keywords.
Statistics
Access and download statisticsCorrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:mathme:v:78:y:2013:i:1:p:61-76. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .
Please note that corrections may take a couple of weeks to filter through the various RePEc services.