Journal of the Operations Research Society of Japan
Online ISSN : 2188-8299
Print ISSN : 0453-4514
ISSN-L : 0453-4514
OPTIMAL CONTROL FOR PARTIALLY OBSERVABLE MARKOV DECISION PROCESSES OVER AN INFINITE HORIZON
Katsushige SawakiAkira Ichikawa
Author information
JOURNAL FREE ACCESS

1978 Volume 21 Issue 1 Pages 1-16

Details
Abstract
In this paper we consider an optimal control problem for partially observable Markov decision processes with finite states, signals and actions over an infinite horizon. It is shown that there are <isins>-optimal piecewise-linear value functions and piecewise-constant policies which are simple. Simple means that there are only finitely many pieces, each of which is defined on a convex polyhedral set. An algorithm based on the method of successive approximation is developed to compute <isins>-optimal policy and <isins>-optimal cost. Furthermore, a special class of stationary policies, called finitely transient, will be considered. It will be shown that such policies have attractive properties which enable us to convert a partially observable Markov decision chain into a usual finite state Markov one.
Content from these authors
© 1978 The Operations Research Society of Japan
Previous article Next article
feedback
Top