• DocumentCode
    1940482
  • Title

    Double action Q-learning for obstacle avoidance in a dynamically changing environment

  • Author

    Ngai, Daniel C K ; Yung, Nelson W C

  • Author_Institution
    Dept. of Electron. & Electr. Eng., Hong Kong Univ., China
  • fYear
    2005
  • fDate
    6-8 June 2005
  • Firstpage
    211
  • Lastpage
    216
  • Abstract
    In this paper, we propose a new method for solving the reinforcement learning problem in a dynamically changing environment, as in vehicle navigation, in which the Markov decision process used in traditional reinforcement learning is modified so that the response of the environment is taken into consideration for determining the agent\´s next state. This is achieved by changing the action-value function to handle three parameters at a time, namely, the current state, action taken by the agent, and action taken by the environment. As it considers the actions by the agent and environment, it is termed "double action". Based on the Q-learning method, the proposed method is implemented and the update rule is modified to handle all of the three parameters. Preliminary results show that the proposed method has the sum of rewards (negative) 89.5% less than that of the traditional method. Apart from that, our new method also has the total number of collisions and mean steps used in one episode 89.5% and 15.5% lower than that of the traditional method respectively.
  • Keywords
    Markov processes; automated highways; collision avoidance; decision theory; learning (artificial intelligence); problem solving; Markov decision process; action-value function; double action Q-learning; dynamically changing environment; obstacle avoidance; reinforcement learning problem; vehicle navigation; Computational efficiency; Cost function; Delay; Dynamic programming; Electronic mail; Genetic programming; Learning; Legged locomotion; Navigation; Vehicles;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Intelligent Vehicles Symposium, 2005. Proceedings. IEEE
  • Print_ISBN
    0-7803-8961-1
  • Type

    conf

  • DOI
    10.1109/IVS.2005.1505104
  • Filename
    1505104