Page 169 - Macromolecular Crystallography
P. 169

158  MACROMOLECULAR CRYS TALLOGRAPHY

          (a)                        (b)                           (c)










                        Target distance








                           Interatomic                Rotamer angle
                            distance                                               Rotamer angle
        Figure 11.2 This figure shows the probability functions for two typical crystallographic examples. In (a) two atoms show a preference for a
        certain bond distance, which is inversely related to an energy function of minimization target. This function is unimodal and the corresponding
        optimization problem is easy to solve. In (b) four different rotamer positions are depicted. The probability distributions are derived from
        knowledge-based potential of this side chain. This function is multimodal and requires a sufficiently wide sampling to solve as an optimization
        problem. Even so, the ‘best’ result may not be correct. In (c) the electron density, shown in darker colour, is also taken into account that, when
        correctly weighted, points clearly to the right solution. It would be misleading to state that this is a standard situation but is does show the
        significance of taking all available data into account and weighting them appropriately.


        such studies of linear problems are called Integer  its gradient
        Programming. Further generalization is covered by
                                                       ∂f(x)
        Quadratic Programming that handles objective func-
                                                        ∂x
        tions with quadratic terms, but still with linear
                                                     (the first derivative vector of the function with
        constraints. If there is no requirement for linear-
                                                     respect to the variables – slope) can be used effec-
        ity and the variables are still integer, the problem
                                                     tively to determine where to sample next. A simple
        is considered to be one of Combinatorial Optimiza-
                                                     method that uses gradient information to find a local
        tion. Non-linear Programming is an even more general
                                                     minimum is gradient descent or the method of steepest
        topic and allows the objective function or the con-
                                                     descent. Ifthesecondderivativeofthefunctionexists,
        straints or both to contain non-linear parts. Dynamic
                                                     the Hessian (the second-order derivative matrix of
        Programming cuts the optimization problem into
                                                     all combinations of variables) can be employed to
        smaller subproblems from which the global solution
                                                     help the search and to classify stationary points. The
        is constructed. Many more classifications exist and
                                                     use of Hessian information significantly improves
        dedicated handbooks or software manuals should
                                                     the search and reduces the number of required steps
        be consulted for further information. The book on
                                                     to reach the minimum. Here each step in the iter-
        statistical learning by Hastie et al. (2001) provides a
                                                     ative procedure is proportional to the gradient at
        good basic introduction to many of these topics.
                                                     the current point. Given an appropriate step size
          Most optimization procedures work in an iterative
                                                     α (which can be problematic to determine) this
        manner. An initial guess is made at where the min-
                                                     approach works very well, also in high-dimensional
        imum may lie and various techniques are applied
                                                     space:
        to improve that estimate. This is generally achieved

        by random to very sophisticated sampling methods.            −1     ∂f(x)
                                                       x new = x old  − αH  f(x)
        If the objective function f(x) is differentiable then                ∂x
   164   165   166   167   168   169   170   171   172   173   174