• DocumentCode
    617815
  • Title

    A comparison of PSO and Reinforcement Learning for multi-robot obstacle avoidance

  • Author

    Di Mario, Ezequiel ; Talebpour, Zeynab ; Martinoli, Alcherio

  • Author_Institution
    Distrib. Intell. Syst. & Algorithms Lab, Ecole Polytech. Fed. de Lausanne, Lausanne, Switzerland
  • fYear
    2013
  • fDate
    20-23 June 2013
  • Firstpage
    149
  • Lastpage
    156
  • Abstract
    The design of high-performing robotic controllers constitutes an example of expensive optimization in uncertain environments due to the often large parameter space and noisy performance metrics. There are several evaluative techniques that can be employed for on-line controller design. Adequate benchmarks help in the choice of the right algorithm in terms of final performance and evaluation time. In this paper, we use multi-robot obstacle avoidance as a benchmark to compare two different evaluative learning techniques: Particle Swarm Optimization and Q-learning. For Q-learning, we implement two different approaches: one with discrete states and discrete actions, and another one with discrete actions but a continuous state space. We show that continuous PSO has the highest fitness overall, and Q-learning with continuous states performs significantly better than Q-learning with discrete states. We also show that in the single robot case, PSO and Q-learning with discrete states require a similar amount of total learning time to converge, while the time required with Q-learning with continuous states is significantly larger. In the multi-robot case, both Q-learning approaches require a similar amount of time as in the single robot case, but the time required by PSO can be significantly reduced due to the distributed nature of the algorithm.
  • Keywords
    collision avoidance; continuous time systems; control system synthesis; discrete time systems; learning (artificial intelligence); multi-robot systems; particle swarm optimisation; PSO; Q-learning approach; continuous state space; discrete action; discrete state; multirobot; noisy performance metrics; obstacle avoidance; online controller design; particle swarm optimization; reinforcement learning; robotic controller design; Collision avoidance; Mobile robots; Optimization; Robot sensing systems; Wheels;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Evolutionary Computation (CEC), 2013 IEEE Congress on
  • Conference_Location
    Cancun
  • Print_ISBN
    978-1-4799-0453-2
  • Electronic_ISBN
    978-1-4799-0452-5
  • Type

    conf

  • DOI
    10.1109/CEC.2013.6557565
  • Filename
    6557565