Title :
Towards possibilistic reinforcement learning algorithms
Author :
Sabbadin, Reégis
Author_Institution :
Toulouse-Unite de Biometrie et Intelligence Artificielle, INRA, Castanet-Tolosan, France
fDate :
6/23/1905 12:00:00 AM
Abstract :
We propose a framework and algorithms for reinforcement learning in sequential decision problems under uncertainty in which the rewards are qualitative, and/or are temporarily aggregated by a "minimum" instead of a sum as in the classical Markov decision processes framework. The framework is based on a "possibilistic" version of Markov decision processes and the learning algorithms are based on indirect methods in which the possibilistic model of the problem is learned while the problem itself is solved, using dynamic programming
Keywords :
Markov processes; decision theory; dynamic programming; learning (artificial intelligence); possibility theory; uncertainty handling; Markov decision processes; dynamic programming; indirect methods; possibilistic reinforcement learning algorithms; sequential decision problems; uncertainty; Computational modeling; Decision making; Large Hadron Collider; Learning; Possibility theory; Stochastic processes; Uncertainty; Utility theory;
Conference_Titel :
Fuzzy Systems, 2001. The 10th IEEE International Conference on
Conference_Location :
Melbourne, Vic.
Print_ISBN :
0-7803-7293-X
DOI :
10.1109/FUZZ.2001.1007334