Title :
Lifted-Rollout for Approximate Policy Iteration of Markov Decision Process
Author :
Dai, Wang-Zhou ; Yu, Yang ; Zhou, Zhi-Hua
Author_Institution :
Nat. Key Lab. for Novel Software Technol., Nanjing Univ., Nanjing, China
Abstract :
Sampling-based approximate policy iteration, which samples (or "rollout") the current policy and find improvement from the samples, is an efficient and practical approach for solving policies in Markov decision process. Such an approach, however, suffers from the inherent variance of sampling. In this paper, we propose the lifted-rollout approach. This approach models the decision process using a directed a cyclic graph and then lifts the possibly huge graph by compressing similar nodes. Finally the approximate policy is obtained by inference on the lifted graph. Experiments show that our approach avoids the sampling variance and achieves significantly better performance.
Keywords :
Markov processes; decision making; directed graphs; sampling methods; Markov decision process; directed acyclic graph; lifted graph; lifted rollout; sampling based approximate policy iteration; Approximation algorithms; Approximation methods; Argon; Equations; Grounding; Markov processes; Trajectory;
Conference_Titel :
Data Mining Workshops (ICDMW), 2011 IEEE 11th International Conference on
Conference_Location :
Vancouver, BC
Print_ISBN :
978-1-4673-0005-6
DOI :
10.1109/ICDMW.2011.112