DocumentCode :
1853117
Title :
Efficient learning algorithms for episodic tasks with acyclic state spaces
Author :
Reveliotis, Spyros ; Bountourelis, Theologos
Author_Institution :
Sch. of Ind. & Syst. Eng., Georgia Inst. of Technol., Atlanta, GA
fYear :
2006
fDate :
8-10 Oct. 2006
Firstpage :
411
Lastpage :
418
Abstract :
This paper considers the problem of computing an optimal policy for a Markov decision process (MDP), under lack of complete a priori knowledge of (i) the branching probability distributions determining the evolution of the process state upon the execution of the different actions, and (ii) the probability distributions characterizing the immediate rewards returned by the environment as a result of the execution of these actions at different states of the process. In addition, it is assumed that the underlying task evolves in a repetitive, episodic manner, with each episode starting from a well-defined initial state and evolving over an acyclic state space. A novel efficient algorithm for this problem is proposed, and its convergence properties and computational complexity are rigorously characterized in the formal framework of computational learning theory
Keywords :
Markov processes; computational complexity; learning (artificial intelligence); statistical distributions; Markov decision process; acyclic state spaces; branching probability distributions; computational complexity; computational learning theory; efficient learning algorithms; episodic tasks; Aerospace industry; Algorithm design and analysis; Automation; Computational complexity; Convergence; Distributed computing; Probability distribution; State-space methods; Systems engineering and theory; USA Councils;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Automation Science and Engineering, 2006. CASE '06. IEEE International Conference on
Conference_Location :
Shanghai
Print_ISBN :
1-4244-0310-3
Electronic_ISBN :
1-4244-0311-1
Type :
conf
DOI :
10.1109/COASE.2006.326917
Filename :
4120383
Link To Document :
بازگشت