Back to Search Start Over

The Optimal Control of Partially Observable Markov Processes over a Finite Horizon.

Authors :
Smallwood, Richard D.
Sondik, Edward J.
Source :
Operations Research; Sep/Oct73, Vol. 21 Issue 5, p1071-1088, 18p
Publication Year :
1973

Abstract

This paper formulates the optimal control problem for a class of mathematical models in which the system to be controlled is characterized by a finite-state discrete-time Markov process. The states of this internal process are not directly observable by the controller; rather, he has available a set of observable outputs that are only probabilistically related to the internal state of the system. The formulation is illustrated by a simple machine-maintenance example, and other specific application areas are also discussed. The paper demonstrates that, if there are only a finite number of control intervals remaining, then the optimal payoff function is a piecewise-linear, convex function of the current state probabilities of the internal Markov process. In addition, an algorithm for utilizing this property to calculate the optimal control policy and payoff function for any finite horizon is outlined. These results are illustrated by a numerical example for the machine-maintenance problem. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
0030364X
Volume :
21
Issue :
5
Database :
Complementary Index
Journal :
Operations Research
Publication Type :
Academic Journal
Accession number :
8735898
Full Text :
https://doi.org/10.1287/opre.21.5.1071