DocumentCode :
1675680
Title :
Towards possibilistic reinforcement learning algorithms
Author :
Sabbadin, Reégis
Author_Institution :
Toulouse-Unite de Biometrie et Intelligence Artificielle, INRA, Castanet-Tolosan, France
Volume :
1
fYear :
2001
fDate :
6/23/1905 12:00:00 AM
Firstpage :
404
Lastpage :
407
Abstract :
We propose a framework and algorithms for reinforcement learning in sequential decision problems under uncertainty in which the rewards are qualitative, and/or are temporarily aggregated by a "minimum" instead of a sum as in the classical Markov decision processes framework. The framework is based on a "possibilistic" version of Markov decision processes and the learning algorithms are based on indirect methods in which the possibilistic model of the problem is learned while the problem itself is solved, using dynamic programming
Keywords :
Markov processes; decision theory; dynamic programming; learning (artificial intelligence); possibility theory; uncertainty handling; Markov decision processes; dynamic programming; indirect methods; possibilistic reinforcement learning algorithms; sequential decision problems; uncertainty; Computational modeling; Decision making; Large Hadron Collider; Learning; Possibility theory; Stochastic processes; Uncertainty; Utility theory;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Fuzzy Systems, 2001. The 10th IEEE International Conference on
Conference_Location :
Melbourne, Vic.
Print_ISBN :
0-7803-7293-X
Type :
conf
DOI :
10.1109/FUZZ.2001.1007334
Filename :
1007334
Link To Document :
بازگشت