Page 126 - A First Course In Stochastic Models
P. 126

118                   DISCRETE-TIME MARKOV CHAINS

                  For the generated sequence of successive states X 0 , X 1 , . . . , it holds that
                                   n          N
                                1
                            lim      f (X k ) =  f (j)π j  with probability 1
                            n→∞ n
                                  k=0        j=1
                for any given function f . Thus the Metropolis—Hastings algorithm can be used to
                find performance measures of the Markov chain {X n } such as the long-run average
                cost per time unit when a cost structure is imposed on the Markov chain.
                  The most widely used version of the Metropolis—Hastings algorithm is the
                Gibbs sampler. Suppose that (N 1 , . . . , N d ) is a d-dimensional stochastic vector
                whose probability distribution

                               p(x 1 , . . . , x d ) = P {N 1 = x 1 , . . . , N d = x d }
                is known up to a multiplicative constant. This situation will be encountered in
                Section 5.6 in the context of a closed queueing network. In this particular applica-
                tion the univariate conditional distribution

                             P {N k = x k |N j = x j for j = 1, . . . , d with j  = k}  (3.4.17)
                is explicitly known for each k = 1, . . . , d. In order to apply the Gibbs sampler,
                it is required that the univariate conditional distributions in (3.4.17) are known.
                The Gibbs sampler generates a sequence of successive states (x 1 , . . . , x d ) from a
                Markov chain whose equilibrium distribution is given by p(x 1 , . . . , x d ).

                Gibbs sampler
                Step 0. Choose an initial state x = (x 1 , . . . , x d ).
                Step 1. For the current state x choose a coordinate which is equally likely to be
                any of the coordinates 1, . . . , d. If coordinate k is chosen, then generate a random
                variable Y whose probability distribution is given by
                       P {Y = y} = P {X k = y|X j = x j for j = 1, . . . , d with j  = k}.

                If Y = y, let the candidate state y = (x 1 , . . . , x k−1 , y, x k+1 , . . . , x d ).
                Step 2. The next state x = (x 1 , . . . , x d ) is set equal to y. Repeat step 1 with this
                new state x.
                  The Gibbs sampler uses the Metropolis—Hastings algorithm with the choice
                                1
                         m x,y =  P {X k = y|X j = x j for j = 1, . . . , d with j  = k}
                                d
                for the Markov matrix M. It is not difficult to verify that for this choice the
                acceptance probability α x,y is given by


                                               p(y)p(x)
                                    α x,y = min        , 1 = 1.
                                               p(x)p(y)
                Hence the candidate state is always accepted as the next state of the Markov chain.
   121   122   123   124   125   126   127   128   129   130   131