Abstract :
In this paper, we propose a new approach to the theory of finite multichain Markov decision processes (MDPs) with different performance optimization criteria. We first propose the concept of nth-order bias; then, using the average reward and bias difference formulas derived in this paper, we develop an optimization theory for finite MDPs that covers a complete spectrum from average optimality, bias optimality, to all high-order bias optimality, in a unified way. The approach is simple, direct, natural, and intuitive; it depends neither on Laurent series expansion nor on discounted MDPs. We also propose one-phase policy iteration algorithms for bias and high-order bias optimal policies, which are more efficient than the two-phase algorithms in the literature. Furthermore, we derive high-order bias optimality equations. This research is a part of our effort in developing sensitivity-based learning and optimization theory.
Keywords :
Markov processes; decision theory; iterative methods; optimisation; series (mathematics); Laurent series expansion; average reward-bias difference formula; finite multichain Markov decision process; nth-order bias optimality; one-phase policy iteration algorithm; optimization criteria; Computer vision; Convergence; Equations; Linear programming; Optimization; Sun; User-generated content; $n$ th-bias optimality; $n$ th potentials; Average optimality; Markov decision processes (MDPs); bias optimality; discrete-event systems; policy iteration;