Title :
Relative reward strength algorithms for learning automata
Author :
Simha, Rahul ; Kurose, James F.
Author_Institution :
Dept. of Comput. & Inf. Sci., Massachusetts Univ., Amherst, MA, USA
Abstract :
A novel class of action probability update algorithms for learning automata that use the relative reward strengths of responses from the environment is examined. Specifically, update algorithms for S-model automata in which `recent´ environmental responses for each of the actions retained are used. A convergence result is proven and the behavior of these automat is studied by simulation. A major result is that the performance of these algorithms is superior, in several respects, to that of the well-known SLR-1 update algorithm. Additional results are presented on the variability of performance, the cost of learning and, in the case of static environments, modifications that result in improved convergence
Keywords :
automata theory; probability; S-model automata; action probability update algorithms; convergence; learning automata; relative reward strengths; Algorithm design and analysis; Convergence; Costs; History; Information science; Learning automata; Probability distribution; Stochastic processes;
Journal_Title :
Systems, Man and Cybernetics, IEEE Transactions on