DocumentCode
1797299
Title
The scalarized multi-objective multi-armed bandit problem: An empirical study of its exploration vs. exploitation tradeoff
Author
Yahyaa, Saba Q. ; Drugan, Madalina M. ; Manderick, Bernard
Author_Institution
Artificial Intell. Lab., Vrije Univ. Brussel, Brussels, Belgium
fYear
2014
fDate
6-11 July 2014
Firstpage
2290
Lastpage
2297
Abstract
The multi-armed bandit (MAB) problem is the simplest sequential decision process with stochastic rewards where an agent chooses repeatedly from different arms to identify as soon as possible the optimal arm, i.e. the one of the highest mean reward. Both the knowledge gradient (KG) policy and the upper confidence bound (UCB) policy work well in practice for the MAB-problem because of a good balance between exploitation and exploration while choosing arms. In case of the multi-objective MAB (or MOMAB)-problem, arms generate a vector of rewards, one per arm, instead of a single scalar reward. In this paper, we extend the KG-policy to address multi-objective problems using scalarization functions that transform reward vectors into single scalar reward. We consider different scalarization functions and we call the corresponding class of algorithms scalarized KG. We compare the resulting algorithms with the corresponding variants of the multi-objective UCBl-policy (MO-UCB1) on a number of MOMAB-problems where the reward vectors are drawn from a multivariate normal distribution. We compare experimentally the exploration versus exploitation trade-off and we conclude that scalarized-KG outperforms MO-UCB1 on these test problems.
Keywords
decision theory; normal distribution; stochastic processes; KG-policy; MAB-problem; MO-UCB1; MOMAB-problems; exploration-versus-exploitation trade-off; knowledge gradient; multiobjective MAB; multiobjective UCB1-policy; multivariate normal distribution; reward vectors; scalarization functions; scalarized KG; scalarized multiobjective multiarmed bandit problem; sequential decision process; single scalar reward; stochastic rewards; upper confidence bound; Chebyshev approximation; Gaussian distribution; Indexes; Measurement; Pareto optimization; Standards; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), 2014 International Joint Conference on
Conference_Location
Beijing
Print_ISBN
978-1-4799-6627-1
Type
conf
DOI
10.1109/IJCNN.2014.6889390
Filename
6889390
Link To Document