Markov Decision Processes with Their Applications

Markov Decision Processes with Their Applications

Author
Qiying Hu, Wuyi Yue
Publication Year
2008
Publisher
Springer
Language
English
Document Type
Book
Faculty / Subject Heading
Mathematics and Statistics

Markov decision processes (MDPs), also called stochastic dynamic programming, were first studied in the 1960s. MDPs can be used to model and solve dynamic decision-making problems that are multi-period and occur in stochastic circumstances. There are three basic branches in MDPs: discrete-time MDPs, continuous-time MDPs and semi-Markov decision processes. Starting from these three branches, many generalized MDPs models have been applied to various practical problems. These models include partially observable MDPs, adaptive MDPs, MDPs in stochastic environments, and MDPs with multiple objectives, constraints or imprecise parameters.


Keywords: Mathematics and Statistics / Markov decision process / Observable / Optimal control / Decision making problems / Decision processes / Discrete event systems / Stochastic dynamic programming