Page 176 - A First Course In Stochastic Models
P. 176

TRANSIENT STATE PROBABILITIES                169

                distribution, the truncation integer M can be chosen as
                                                    √
                                           M = νt + c νt
                for some constant c with 0 < c ≤ c 0 (ε), where c 0 (ε) depends only on the tolerance
                number ε. As a consequence the computational complexity of the uniformization
                              2
                method is O(νtN ) where N is the number of states of the Markov chain. Hence
                the uniformization method should be applied with the choice
                                            ν = max ν i .
                                                 i∈I
                The number νt is a crucial factor for the computational complexity of the uni-
                formization method, as it is for the Runge–Kutta method, and is called the index
                of stiffness. Also, the following remark may be helpful. For fixed initial state i,
                the recursion scheme (4.5.6) boils down to the multiplication of a vector with the
                matrix P. In many applications the matrix P is sparse. Then the computational effort
                can be considerably reduced by using a data structure for sparse matrix multiplica-
                tions. The uniformization results (4.5.5) and (4.5.6) need only a minor modification
                when the initial state X(0) has a given probability distribution {π 0 (i), i ∈ I}. The
                          (n)                        (n)             (n)
                probability p  should then be replaced by p  =  π 0 (i)p  and this prob-
                          ij                         j      i∈I      ij
                                                     (n)         (n−1)
                ability can recursively be computed from p  =  p     p kj starting with
                                                     j       k∈I  k
                 (0)
                p   = π 0 (j) for j ∈ I. For example, this modification may be used to compute
                 j
                the transient state probabilities in finite-capacity queueing systems with a non-
                stationary Poisson arrival process and exponential services, where the arrival rate
                function λ(t) is (approximately) a piecewise-constant function. One then computes
                the transient state probabilities for each interval separately on which λ(t) is constant
                and uses the probability distribution of the state at the beginning of the interval as
                the distribution of the initial state.

                Expected transient rewards
                Assume that a reward at rate r(j) is earned whenever the continuous-time Markov
                chain {X(t)} is in state j, while a lump reward of F jk is earned each time the
                process makes a transition from state j to state k ( = j). Let

                           R(t) = the total reward earned up to time t,  t ≥ 0.
                The following lemma shows that it is a simple matter to compute the expected
                value of the reward variable R(t). The computation of the probability distribution
                of R(t) is much more complex and will be addressed in Section 4.6.

                Lemma 4.5.4 Suppose that the continuous-time Markov chain {X(t)} satisfies
                Assumption 4.1.2. Then

                  E[R(t) | X(0) = i] =  r(j)E ij (t) +  E ij (t)  q jk F jk ,  t > 0,  (4.5.7)
                                     j∈I          j∈I      k =j
   171   172   173   174   175   176   177   178   179   180   181