• DocumentCode
    1803501
  • Title

    Continuous action for multi-agent q-learning

  • Author

    Hwang, Kao-Shing ; Chen, Yu-Jen ; Lin, Tzung-Feng ; Jiang, Wei-Cheng

  • Author_Institution
    Dept. of Electr. Eng., Nat. Chung Cheng Univ., Ming-Hsiung, Taiwan
  • fYear
    2011
  • fDate
    15-18 May 2011
  • Firstpage
    418
  • Lastpage
    423
  • Abstract
    Q-learning, a most widely used reinforcement learning method, normally needs well-defined quantized state and action spaces to obtain an optimal policy for accomplishing a given task. This means it difficult to be applied to real robot tasks because of poor performance of learned behavior due to the failure of quantization of continuous state and action spaces. In this paper, we proposed a fuzzy-based Cerebellar Model Articulation Controller method to calculate contribution values to estimate a continuous action value in order to make motion smooth and effective. And we implement it to a multi-agent system for real robot applications.
  • Keywords
    cerebellar model arithmetic computers; control engineering computing; fuzzy control; learning (artificial intelligence); motion control; multi-agent systems; robots; action spaces; continuous state quantization; fuzzy based cerebellar model articulation controller; multiagent Q-learning; real robot tasks; reinforcement learning; smooth motion; Learning; Logic gates; Multiagent systems; Quantization; Robot kinematics; Robot sensing systems; Cerebellar Model Articulation Controller; Multi-agent; Reinforcement learning;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Control Conference (ASCC), 2011 8th Asian
  • Conference_Location
    Kaohsiung
  • Print_ISBN
    978-1-61284-487-9
  • Electronic_ISBN
    978-89-956056-4-6
  • Type

    conf

  • Filename
    5899108