Time-Sharing Policies for Controlled Markov Chains
نویسندگان
چکیده
We propose a class of non-stationary policies called \policy time sharing" (p.t.s.), which possess several desirable properties for problems where the criteria are of the average-cost type; an optimal policy exists within this class, the computation of optimal policies is straightforward, and the implementation of this policy is easy. While in the nite state case stationary policies are also known to share these properties, the new policies are much more exible, in the sense that they can be applied to solve adaptive problems, and that they suggest new ways to incorporate the particular structure of the problem at hand into the derivation of optimal policies. In addition, they provide insight into the pathwise-structure of Controlled Markov chains. To use p.t.s. policies one alternates between the use of several stationary deterministic policies, switching when reaching some predetermined state. In some (countable state) cases optimal solutions of the policy time sharing type are available and easy to compute, whereas optimal stationary policies are not available. Examples illustrating the last point and the usefulness of the new approach are discussed, involving constrained optimization problems with countable state space or compact action space.
منابع مشابه
Action Time Sharing Policies for Ergodic Control of Markov Chains
Ergodic control for discrete time controlled Markov chains with a locally compact state space and a compact action space is considered under suitable stability, irreducibility and Feller continuity conditions. A flexible family of controls, called action time sharing (ATS) policies, associated with a given continuous stationary Markov control, is introduced. It is shown that the long term avera...
متن کاملDrift and monotonicity conditions for continuous-time controlled Markov chains with an average criterion
In this paper, we give conditions for the existence of average optimal policies for continuous-time controlled Markov chains with a denumerable state–space and Borel action sets. The transition rates are allowed to be unbounded, and the reward/cost rates may have neither upper nor lower bounds. In the spirit of the “drift and monotonicity” conditions for continuous-time Markov processes, we pro...
متن کاملVariance minimization and the overtaking optimality approach to continuous-time controlled Markov chains
This paper deals with denumerable-state continuous-time controlled Markov chains with possibly unbounded transition and reward rates. It concerns optimality criteria that improve the usual expected average reward criterion. First, we show the existence of average reward optimal policies with minimal average variance. Then we compare the variance minimization criterion with overtaking optimality...
متن کاملAnalysis of Multi-server Systems via Dimensionality Reduction of Markov Chains
The performance analysis of multiserver systems is notoriously hard, especially when the systeminvolves resource sharing or prioritization. We provide two new analytical tools for the perfor-mance analysis of multiserver systems: moment matching algorithms and dimensionality reductionof Markov chains (DR).Moment matching algorithms allow us to approximate a general distribution ...
متن کاملOptimal Policies for Controlled Markov Chains with a Constraint
The time average reward for a discrete-time controlled Markov process subject to a time-average cost constraint is maximized over the class of al causal policies. Each epoch, a reward depending on the state and action, is earned, and a similarly constituted cost is assessed; the time average of the former is maximized, subject to a hard limit on the time average of the latter. It is assumed tha...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Operations Research
دوره 41 شماره
صفحات -
تاریخ انتشار 1993