Title :
Scalarized multi-objective reinforcement learning: Novel design techniques
Author :
Van Moffaert, K. ; Drugan, Madalina M. ; Nowe, Ann
Author_Institution :
Dept. of Comput. Sci., Vrije Univ. Brussel, Brussels, Belgium
Abstract :
In multi-objective problems, it is key to find compromising solutions that balance different objectives. The linear scalarization function is often utilized to translate the multi-objective nature of a problem into a standard, single-objective problem. Generally, it is noted that such as linear combination can only find solutions in convex areas of the Pareto front, therefore making the method inapplicable in situations where the shape of the front is not known beforehand, as is often the case. We propose a non-linear scalarization function, called the Chebyshev scalarization function, as a basis for action selection strategies in multi-objective reinforcement learning. The Chebyshev scalarization method overcomes the flaws of the linear scalarization function as it can (i) discover Pareto optimal solutions regardless of the shape of the front, i.e. convex as well as non-convex , (ii) obtain a better spread amongst the set of Pareto optimal solutions and (iii) is not particularly dependent on the actual weights used.
Keywords :
Chebyshev approximation; Pareto optimisation; learning (artificial intelligence); Chebyshev scalarization function; Pareto front; Pareto optimal solutions; linear scalarization function; multiobjective nature translation; nonlinear scalarization function; scalarized multiobjective reinforcement learning; Benchmark testing; Chebyshev approximation; Learning (artificial intelligence); Measurement; Pareto optimization; Shape;
Conference_Titel :
Adaptive Dynamic Programming And Reinforcement Learning (ADPRL), 2013 IEEE Symposium on
Conference_Location :
Singapore
DOI :
10.1109/ADPRL.2013.6615007