DocumentCode :
288537
Title :
Dynamic concept model learns optimal policies
Author :
Szepesvári, Cs
Author_Institution :
Dept. of Math., Jozsef Attila Univ., Szeged, Hungary
Volume :
3
fYear :
1994
fDate :
27 Jun-2 Jul 1994
Firstpage :
1738
Abstract :
Dynamic concept model (DCM) is a goal-oriented neural controller, that builds an internal representation of events and chains of events in the form of a directed graph and uses spreading activation for decision making. It is shown, that a special case of DCM is equivalent to reinforcement learning (RL) and is capable of learning the optimal policy in a probabilistic world. The memory and computational requirements of both DCM and RL are analyzed and a special algorithm is introduced, that ensures intentional behavior
Keywords :
Markov processes; decision theory; directed graphs; learning (artificial intelligence); neurocontrollers; computational requirements; decision making; directed graph; dynamic concept model; goal-oriented neural controller; intentional behavior; internal representation; optimal policies; probabilistic world; spreading activation; Algorithm design and analysis; Artificial neural networks; Brain modeling; Control systems; Cost function; Decision making; Explosions; Learning; Mathematics; Problem-solving;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
Type :
conf
DOI :
10.1109/ICNN.1994.374418
Filename :
374418
Link To Document :
بازگشت