Title :
Using reinforcement learning to catch a baseball
Author :
Das, Sreerupa ; Das, Rajarshi
Author_Institution :
Dept. of Comput. Sci., Colorado Univ., Boulder, CO, USA
fDate :
27 Jun-2 Jul 1994
Abstract :
Moments after a baseball batter has hit a fly ball, an outfielder has to decide whether to run forward or backward to catch the ball. Judging a fly ball is a difficult task, especially when the fielder is in the plane of the ball´s trajectory. A previous study in experimental psychology suggests that to intercept the ball, the fielder has to run such that d2(tanφ)/dt2 is close to zero, where φ is the elevation angle of the ball from the fielder´s perspective. The authors investigate whether d2(tanφ)/dt2 information is sufficient to learn this task in two reinforcement learning models: AHC and Q learning. The authors´ results indicate that although d2(tanφ)/dt2 provides initial clue as to the ball´s landing point, it is not a good indicator in the latter stages of the ball´s trajectory. Thus the two models fail to learn to intercept fly balls. However, when information about the perpendicular velocity of the ball with respect to the fielder is also included as an input to the system, it provides the necessary discriminability in the latter stages of the ball´s trajectory, and the two models are able to successfully learn this reinforcement problem
Keywords :
learning (artificial intelligence); neural nets; AHC learning; Q learning; baseball catching; discriminability; fly ball; perpendicular velocity; reinforcement learning; Computer science; Learning; Motion analysis; Physics; Psychology;
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
DOI :
10.1109/ICNN.1994.374676