Page 166 - Machine Learning for Subsurface Characterization
P. 166

140    Machine learning for subsurface characterization


            model formulates the predicted target ^ y as a weighted sum of basis function
            B(x):
                                         p
                                        X

                                     ^ y ¼  α q B q x qi                (5.9)
                                      i
                                         q¼1
            where x qi is the value of qth feature (log) x q at the ith depth point, B q (x qi )
            is a basis function, and α q is the coefficient of B q . The basis function can
            have many different forms, but most likely, it is a hinge function. Each
            knot generates two basis functions (hinge functions): max(0,x qi   C q ) and
            max(0,C q   x qi ), where C q is the knot. Hinge function partitions the feature
            space into subspaces by using different values of C q . In our study, we use 10
            knots to partition the feature space; consequently, there are 21 parameters
            that need to be computed in the training phase. The MARS model
                                               2
            synthesizes DTC and DTS logs with R of 0.85 and 0.83, respectively.
            LASSO and ElasticNet models showed that RLA0, RLA4, and RLA5 logs
            have high correlations with RLA1, RLA2, and RLA3 logs. Therefore, when
            training the MARS model, we do not use the RLA0, RLA4, and RLA5 logs.
            In conclusion, the MARS algorithm involves a forward phase and a
            backward phase. The forward phase places candidate knots at random
            positions within the range of each feature to define a pair of basis functions
            (BFs). At each step, the model adapts the knot and its corresponding pair of
            BFs to give the maximum reduction in sum-of-squares residual error. This
            process of adding BFs continues until a user-defined threshold. The
            backward phase involves deleting the redundant knots and corresponding
            BFs that made the least contribution to the model fitting.

            2.4.6 Artificial neural network (ANN) model
            ANN is a widely used machine learning model suitable for both linear and
            nonlinear regression tasks. A neural network comprises an input layer, an
            output layer, and several hidden layers. The capacity of the neural network
            model to fit data can be adjusted by changing the number of hidden layers and
            the number of neurons in each hidden layer. Each hidden layer and output
            layer are made of neurons. In a densely connected neural network architecture,
            each neuron in a layer is connected to neurons in the immediately previous
            layer. The neuron in a layer receives the output from the neurons in the
            immediately previous layer multiplied by weight of the connection between
            the two neurons. Each neuron sums the weighted outputs from the neurons in
            the immediately previous layer and feeds the summation through a activation
            function that generates values between  1and1or0and1. Activation
            function, also referred as transfer function, controls the nature of output of a
            neuron. For the neurons in hidden layers, it is recommended to use Rectified
            Linear Unit as the activation function because it overcomes the vanishing
            gradient problem of once popular Sigmoid and Tanh activation functions. For
            the neurons in output layer, softmax is used as the activation function for
   161   162   163   164   165   166   167   168   169   170   171