Title :
On the myopic policy for a class of restless bandit problems with applications in dynamic multichannel access
Author :
Liu, Keqin ; Zhao, Qing
Author_Institution :
Dept. of Electr. & Comput. Eng., Univ. of California, Davis, CA, USA
Abstract :
We consider a class of restless multi-armed bandit problems that arises in multi-channel opportunistic communications, where channels are modeled as independent and stochastically identical Gilbert-Elliot channels and channel state observations are subject to errors. We show that the myopic channel selection policy has a semi-universal structure that obviates the need to know the Markovian transition probabilities of the channel states. Based on this structure, we establish closed-form lower and upper bounds on the steady-state throughput achieved by the myopic policy. Furthermore, we characterize the approximation factor of the myopic policy to bound its worst-case performance loss with respect to the optimal performance.
Keywords :
Markov processes; approximation theory; channel allocation; probability; Gilbert-Elliot channel; Markovian transition probability; approximation factor; channel state observation; dynamic multichannel access; multiarmed bandit problem; multichannel opportunistic communication; myopic channel selection policy; restless bandit problem; Cognitive radio; Communication system control; Detectors; H infinity control; Performance analysis; Performance loss; Stochastic processes; Throughput; Upper bound; Dynamic multi-channel access; myopic policy; restless multi-armed bandit;
Conference_Titel :
Decision and Control, 2009 held jointly with the 2009 28th Chinese Control Conference. CDC/CCC 2009. Proceedings of the 48th IEEE Conference on
Conference_Location :
Shanghai
Print_ISBN :
978-1-4244-3871-6
Electronic_ISBN :
0191-2216
DOI :
10.1109/CDC.2009.5400366