Access the full text.
Sign up today, get DeepDyve free for 14 days.
Xianping Guo, O. Hernández-Lerma, T. Prieto-Rumeau, Xi-Ren Cao, Junyu Zhang, Q. Hu, M. Lewis, R. Vélez (2006)
A survey of recent results on continuous-time Markov decision processesTOP, 14
Quanxin Zhu, Xianping Guo (2005)
Another Set of Conditions for Strong n (n = −1, 0) Discount Optimality in Markov Decision ProcessesStochastic Analysis and Applications, 23
X.R. Cao, H.F. Chen (1997)
Potentials, perturbation realization and sensitivity analysis of Markov processesIEEE Trans. Automat. Control, 42
T. Prieto-Rumeau, O. Hernández-Lerma (2008)
Ergodic control of continuous-time Markov chains with pathwise constraintsProceedings of the 48h IEEE Conference on Decision and Control (CDC) held jointly with 2009 28th Chinese Control Conference
K. Ross, R. Varadarajan (1989)
Markov Decision Processes with Sample Path Constraints: The Communicating CaseOper. Res., 37
P. Mandl (1973)
A connection between controlled Markov chains and martingalesKybernetika, 9
Xianping Guo, Ke Liu (2001)
A note on optimality conditions for continuous-time Markov decision processes with average cost criterionIEEE Trans. Autom. Control., 46
A. Arapostathis, V. Borkar, E. Fernández-Gaucherand, M. Ghosh, S. Marcus (1993)
Discrete-time controlled Markov processes with average cost criterion: a surveySiam Journal on Control and Optimization, 31
O. Hernández-Lerma, Óscar Vega-Amaya, Guadalupe Carrasco (1999)
Sample-Path Optimality and Variance-Minimization of Average Cost Markov Control ProcessesSIAM J. Control. Optim., 38
Quanxin Zhu (2007)
Average optimality inequality for continuous-time Markov decision processes in Polish spacesMathematical Methods of Operations Research, 66
Xi-Ren Cao, Han-Fu Chen (1997)
Perturbation realization, potentials, and sensitivity analysis of Markov processesIEEE Trans. Autom. Control., 42
Quanxin Zhu (2010)
Variance minimization for continuous-time Markov decision processes: two approachesApplied Mathematics-A Journal of Chinese Universities, 25
Quanxin Zhu, Xianping Guo (2006)
Another set of conditions for Markov decision processes with average sample-path costs ✩Journal of Mathematical Analysis and Applications, 322
Xianping Guo, O. Hernández-Lerma (2003)
Drift and monotonicity conditions for continuous-time controlled Markov chains with an average criterionIEEE Trans. Autom. Control., 48
Quanxin Zhu, T. Prieto-Rumeau (2008)
Bias and Overtaking Optimality for Continuous-Time Jump Markov Decision Processes in Polish SpacesJournal of Applied Probability, 45
Quanxin Zhu (2008)
Average optimality for continuous-time Markov decision processes with a policy iteration approach☆Journal of Mathematical Analysis and Applications, 339
R. Bhattacharya (1982)
On the functional central limit theorem and the law of the iterated logarithm for Markov processesZeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete, 60
Quanxin Zhu, Xinsong Yang, Chuangxia Huang (2009)
Policy Iteration for Continuous-Time Average Reward Markov Decision Processes in Polish SpacesAbstract and Applied Analysis, 2009
Xi-Ren Cao (1998)
The Relations Among Potentials, Perturbation Analysis, and Markov Decision ProcessesDiscrete Event Dynamic Systems, 8
Xianping Guo, Xi-Ren Cao (2005)
Optimal Control of Ergodic Continuous-Time Markov Chains with Average Sample-Path RewardsSIAM J. Control. Optim., 44
Xianping Guo, U. Rieder (2006)
Average optimality for continuous-time Markov decision processes in polish spacesAnnals of Applied Probability, 16
M. Puterman (1994)
Markov Decision Processes: Discrete Stochastic Dynamic Programming
T. Prieto-Rumeau, O. Hernández-Lerma (2006)
Bias Optimality for Continuous-Time Controlled Markov ChainsSIAM J. Control. Optim., 45
M.L. Puterman (1994)
Markov Decision Process
Xianping Guo (2007)
Continuous-Time Markov Decision Processes with Discounted Rewards: The Case of Polish SpacesMath. Oper. Res., 32
O. Hernández-Lerma, J. Lasserre (1999)
Further topics on discrete-time Markov control processes
Quanxin Zhu, Xianping Guo (2007)
Markov Decision Processes with Variance Minimization: A New Condition and ApproachStochastic Analysis and Applications, 25
P. Hall, E. Lukács, Z. Birnbaum, C. Heyde (1980)
Martingale Limit Theory and Its Application
In this paper we study the average sample-path cost (ASPC) problem for continuous-time Markov decision processes in Polish spaces. To the best of our knowledge, this paper is a first attempt to study the ASPC criterion on continuous-time MDPs with Polish state and action spaces. The corresponding transition rates are allowed to be unbounded, and the cost rates may have neither upper nor lower bounds. Under some mild hypotheses, we prove the existence of ɛ (ɛ ≥ 0)-ASPC optimal stationary policies based on two different approaches:one is the “optimality equation” approach and the other is the “two optimality inequalities” approach.
Acta Mathematicae Applicatae Sinica – Springer Journals
Published: Sep 9, 2011
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.