Page 179 - A First Course In Stochastic Models
P. 179

172                 CONTINUOUS-TIME MARKOV CHAINS

                     6l         5l      4l        3l         2l         l
                  6         5        4         3         2         1         crash


                                                         m         m
                                                                           l
                                                             2l
                                       h
                                                       sleep 2   sleep 1
                                         h

                           Figure 4.5.1  The transition rate diagram for the telescope

                  Taking one year as time unit, consider the numerical example with the data

                                     λ = 0.1, µ = 100 and η = 5.

                The uniformization method is applied with the choice ν = 100 for the uniformized
                leaving rate ν (the value 0 is taken for the leaving rate from the state crash). The
                calculations yield the value 0.000504 for the probability 1−Q(T ) that a crash will
                occur in the next T = 10 years. Similarly, one can calculate that with probability
                0.3901 the sleep mode will not be reached in the next 10 years. In other words,
                the probability of no shuttle mission in the next 10 years equals 0.3901. However,
                if one wishes to calculate the probability distribution of the number of required
                shuttle missions in the next 10 years, one must use the Markov reward model with
                lump rewards (assume a lump reward of 1 each time the process jumps from either
                state 2 or state 1 to the sleep mode). This Markov reward model is much more
                difficult to solve and will be discussed in the next section.


                4.6  TRANSIENT DISTRIBUTION OF CUMULATIVE REWARDS

                A basic and practically important problem is the calculation of the transient prob-
                ability distribution of the cumulative reward in a continuous-time Markov chain.
                For example, a practical application is an oil-production platform which has to
                meet a contractually agreed production level over a specified time period. The pro-
                duction rate is not constant but depends on the stochastically varying state of the
                platform. In this example a continuous-time Markov chain with reward rates may
                be appropriate, where the reward rate r(j) represents the production rate in state j.
                  In Section 4.6.1 we first consider the special case of a Markov reward model
                with reward rates that are either 0 or 1. The cumulative reward in this model
                corresponds to the cumulative sojourn time of the process in a certain set of (good)
                states. A nice and simple extension of the uniformization method can be given to
                compute the transient distribution of the cumulative sojourn time in some given
                set of states. The general Markov reward model with both reward rates and lump
                rewards is dealt with in Section 4.6.2. A discretization algorithm will be discussed
   174   175   176   177   178   179   180   181   182   183   184