DocumentCode
2272907
Title
On the control of jump linear Markov systems with Markov state estimation
Author
Gupta, Vijay ; Murray, Richard M. ; Hassibi, Babak
Author_Institution
Div. of Eng. & Appl. Sci., California Inst. of Technol., CA, USA
Volume
4
fYear
2003
fDate
4-6 June 2003
Firstpage
2893
Abstract
We analyze a jump linear Markov system being stabilized using a linear controller. We consider the case when the Markov state is associated with the probability distribution of a measured variable. We assume that the Markov state is not known, but rather is being estimated based on the observations of the variable. We present conditions for the stability of such a system and also solve the optimal LQR control problem for the case when the state estimate update uses only the last observation value. In particular we consider a suboptimal version of the casual Viterbi estimation algorithm and show that a separation property does not hold between the optimal control and the Markov state estimate. Some simple examples are also presented.
Keywords
Markov processes; linear quadratic control; maximum likelihood estimation; stability; state estimation; statistical distributions; stochastic systems; Markov state estimation; causal Viterbi estimation algorithm; jump linear Markov systems; linear controller; linear quadratic regulation; optimal LQR control problem; probability distribution; Communication channels; Communication system control; Control systems; Delay; Optimal control; Radar tracking; Sensor systems; Stability; State estimation; Wireless sensor networks;
fLanguage
English
Publisher
ieee
Conference_Titel
American Control Conference, 2003. Proceedings of the 2003
ISSN
0743-1619
Print_ISBN
0-7803-7896-2
Type
conf
DOI
10.1109/ACC.2003.1243762
Filename
1243762
Link To Document