Page 182 - A First Course In Stochastic Models
P. 182

TRANSIENT DISTRIBUTION OF CUMULATIVE REWARDS         175

                           n

                       =     P {Y 1 + · · · + Y k ≤ x} α(n, k)
                          k=0
                           n        n
                                           x j
                                       n            x n−j


                       =     α(n, k)            1 −       ,  0 ≤ x < t.
                                       j    t       t
                          k=0      j=k
                This gives the desired expression
                                ∞           n        n
                                                             x j

                                    −νt  (νt)  n         n           x n−j
                  P {O(t) ≤ x} =   e          α(n, k)             1 −        (4.6.1)
                                        n!               j   t        t
                               n=0         k=0      j=k
                for 0 ≤ x < t. The random variable O(t) assumes the value t if the uniformized
                process visits only operational states in (0, t). Thus
                                              ∞                  n
                                                           −νt  (νt)
                                P {O(t) = t} =  α(n, n + 1)e      .
                                                               n!
                                             n=0
                The above expression for P {O(t) ≤ x} is well suited for numerical computations,
                since the summations involve only positive terms. As before, the infinite sum can
                be truncated to M terms, where the error associated with the truncation is bounded
                               n
                by    ∞  e −νt (νt) /n! so that M can be determined beforehand for a given error
                    n=M
                tolerance.
                Computation of the α(n, k)

                The probabilities α(n, k) are determined by the discrete-time Markov chain {X n }
                that governs the state transitions in the uniformized process. The one-step transition
                probabilities of this discrete-time Markov chain are given by p = (ν i /ν)p ij for
                                                                    ij
                j  = i and p = 1−ν i /ν, where p ij = q ij /ν i . To calculate the α(n, k), let α(n, k, j)
                         ii
                be the joint probability that the discrete-time Markov chain {X t } visits k times an
                operational state over the first n state transitions and is in state j after the nth
                transition. Then


                   α(n, k) =   α(n, k, j),  k = 0, 1, . . . , n + 1 and n = 0, 1, . . . .
                            j∈I
                The probabilities α(n, k, j) can be recursively computed. In the recursion we have
                to distinguish between states j ∈ I 0 and states j ∈ I f . Obviously,


                              α(n, k, j) =  α(n − 1, k − 1, i)p ,  j ∈ I 0
                                                           ij
                                         i∈I
                and

                               α(n, k, j) =  α(n − 1, k, i)p ,  j ∈ I f .
                                                         ij
                                          i∈I
   177   178   179   180   181   182   183   184   185   186   187