Abstract
In this paper, we introduce a basic framework for mutually dependent Markov decision processes (MDMDP) showing recursive mutual dependence. Our model is structured upon two types of finite-stage Markov decision processes. At each stage, the reward in one process is given by the optimal value of the alternative process problem, whose initial state is determined by the current state and decision in the original process. We formulate the MDMDP model and derive mutually dependent recursive equations by dynamic programming. Furthermore, MDMDP is illustrated in a numerical example. The model enables easier treatment of some classes of complex multi-stage decision processes.
Original language | English |
---|---|
Pages (from-to) | 992-998 |
Number of pages | 7 |
Journal | Journal of Advanced Computational Intelligence and Intelligent Informatics |
Volume | 18 |
Issue number | 6 |
DOIs | |
Publication status | Published - 2014 Nov 1 |
Keywords
- Additive reward
- Dynamic programming
- Markov decision process
- Nonserial system
ASJC Scopus subject areas
- Human-Computer Interaction
- Computer Vision and Pattern Recognition
- Artificial Intelligence