Page 203 - Applied Probability
P. 203

9. Descent Graph Methods
                              188
                              9.11 The Lander-Green-Kruglyak Algorithm
                              Descent graphs also provide a basis for deterministically computing the
                              likelihood of a small pedigree [18, 19, 20, 21]. Success within this framework
                              depends on (a) the assumption of genetic equilibrium, (b) the product
                              rule for calculating penetrances across many loci, and (c) the application
                              of results from the theory of hidden Markov chains and Fourier analysis.
                              In explaining how these parts fit together, our first order of business is
                              to identify an appropriate Markov chain. Instead of viewing the chain as
                              evolving over time, we think of it as evolving from one locus to the next
                              along a sequence of ordered loci. At each locus, the state of the chain is the
                              pedigree’s descent graph at that locus. Likelihood calculation proceeds via
                              Baum’s forward algorithm as described here and in Chapter 11 [1, 4].
                                Baum’s algorithm recursively updates the joint probabilities
                                          α i (j)=Pr(Y 1 = y 1 ,...,Y i−1 = y i−1 ,Z i = j),
                              where in the current context Y i represents the random vector of pheno-
                              types at locus i, and Z i represents the random descent graph at that lo-
                              cus. In Section 9.5 we dealt with the problem of computing the likelihood
                              φ i (y i | j) that Y i = y i given that Z i = j. Each of the m meiotic events
                              in the pedigree involves a random choice of whether the contributing par-
                              ent transmits a grandmaternally or a grandpaternally derived allele. These
                              choices determine 2 m  a priori equally likely descent graphs at each locus.
                              Hence, α 1 (j)= 2 −m  at the first locus. Because of the inherent phase un-
                              certainties within a pedigree founder, it is also possible to force one child of
                              each founder to inherit the founder’s grandmaternal allele at an autosomal
                              locus [20]. This action decreases the size of the state space of the Markov
                              chain and speeds up likelihood evaluation. In the interests of brevity, we
                              omit further discussion of this subtle effect.
                                Baum’s forward algorithm updates α i (j)via


                                             α i+1 (k)=      α i (j)φ i (y i | j)t i (k | j),  (9.13)
                                                          j
                              where t i (k | j) is the conditional probability that descent graph k occurs
                              at locus i + 1 given descent graph j at locus i. At the last locus, say
                              locus n, we recover the likelihood of the pedigree by forming the sum

                                  φ n (y n | j)α n (j). At first sight it appears that the update (9.13) takes
                                 j
                              on the order of O(2 2m ) arithmetic operations. This discouraging estimate
                              neglects crucial symmetries, however.
                                To expose these symmetries, we represent the descent graphs j and k by
                              m-vectors of indicators
                                                      j  =(j 1 ,...,j m )
                                                      k  =(k 1 ,...,k m ),
   198   199   200   201   202   203   204   205   206   207   208