DocumentCode
1661868
Title
Modified Q-learning method with fuzzy state division and adaptive rewards
Author
Maeda, Yoichiro
Author_Institution
Fac. of Inf. Sci. & Arts, Osaka Electro-Commun. Univ., Japan
Volume
2
fYear
2002
fDate
6/24/1905 12:00:00 AM
Firstpage
1556
Lastpage
1561
Abstract
Reinforcement learning method can be considered as an adaptive learning method for autonomous agents. It is important to balance between searching behavior of the unknown knowledge and using behavior of the obtained knowledge. However, the learning is not always efficient in every searching stage because of constant learning parameters in the ordinary Q-learning. For this problem, we have already proposed an adaptive Q-learning method with learning parameters tuned by fuzzy rules. Furthermore, it is hard to deal with the continuous states and behaviors in the ordinary reinforcement learning method. It is also difficult to learn the problem with multiple purposes. Therefore, in this research, we propose a modified Q-learning method where the reward values are tuned according to its state and can deal with multiple purposes in the continuous state space by using fuzzy reasoning. We also report some results for the simulation of object chase agents by using this method
Keywords
fuzzy logic; inference mechanisms; learning (artificial intelligence); multi-agent systems; adaptive learning method; adaptive rewards; autonomous agent; continuous state space; fuzzy reasoning; fuzzy rules; fuzzy state division; modified Q-learning method; object chase agents; reinforcement learning; reward values; searching behavior; Art; Autonomous agents; Equations; Fuzzy reasoning; Information science; Learning systems; Mechatronics; Orbital robotics; Robot control; State-space methods;
fLanguage
English
Publisher
ieee
Conference_Titel
Fuzzy Systems, 2002. FUZZ-IEEE'02. Proceedings of the 2002 IEEE International Conference on
Conference_Location
Honolulu, HI
Print_ISBN
0-7803-7280-8
Type
conf
DOI
10.1109/FUZZ.2002.1006738
Filename
1006738
Link To Document