© The Institution of Engineering and Technology
This study introduces the notion of safety for the controlled Markov chains in the continuous-time horizon. The concept is a non-trivial extension of safety control for stochastic systems modelled as discrete-time Markov decision processes, where the safety means that the probability distributions of the system states will not visit the given forbidden set at any time. In this paper study a unit-interval-valued vector that serves as an upper bound on the state probability distribution vector characterises the forbidden set. A probability distribution is then called safe if it does not exceed the upper bound. Under mild conditions the author derives two results: (i) the necessary and sufficient conditions that guarantee the all-time safety of the probability distributions if the starting distribution is safe, and (ii) the characterisation of the supreme set of safe initial probability vectors that remain safe as time passes. In particular, study the paper identifies an upper bound on time and shows that if a distribution is always safe before that time, the distribution is safe at all times. Numerical examples are provided to illustrate the two results.
References
-
-
1)
-
Feinberg, E.A.: `Constrained finite continuous-time Markov decision processes with average rewards', Proc. IEEE Conf. on Decisions and Control, 2002, p. 3805–3810.
-
2)
-
X. Guo ,
O. Hernandez-Lerma
.
Continuous-time controlled Markov chains.
Ann. Appl. Probab.
,
1 ,
363 -
388
-
3)
-
Arapostathis, A., Kumar, R., Hsu, S.-P.: `State-feedback control of Markov chains with safety bounds', Proc. IEEE Conf. on Decision and Control, 2003, Maui, Hawaii, p. 6283–6288.
-
4)
-
A. Arapostathis ,
R. Kumar ,
S. Tangirala
.
Controlled Markov chains with safety upper bounds.
IEEE Trans. Autom. Control
,
7 ,
1230 -
1234
-
5)
-
E. Altman
.
(1999)
Constrained Markov Decision Processes.
-
6)
-
A. Arapostathis ,
R. Kumar ,
S.-P. Hsu
.
Control of Markov chains with safety bounds.
IEEE Trans. Autom. Sci. Eng.
,
4 ,
333 -
343
-
7)
-
E. Seneta
.
(1981)
Non-negative Matrices and Markov Chains.
-
8)
-
M.L. Puterman
.
(1994)
Markov Decision Process: discrete stochastic dynamic programming.
http://iet.metastore.ingenta.com/content/journals/10.1049/iet-cta.2010.0101
Related content
content/journals/10.1049/iet-cta.2010.0101
pub_keyword,iet_inspecKeyword,pub_concept
6
6