Page 230 - Machine Learning for Subsurface Characterization
P. 230

200 Machine learning for subsurface characterization


            then shifts the random sample by the mean vector in the latent layer and scales it
            by the variance vector in the latent layer. The first 3-dimensional layer of the
            decoder is followed by 16-dimensional hidden layer and the 64-dimensional
            output layer that upscale the latent vector to develop a reconstruction of the
            original input. For any sampling of the latent distributions, the decoder
            model learns to accurately reconstruct the input, such that the samples that
            are nearby in the latent space result in very similar reconstructions. Similar
            to the VAE, the loss function of the VAEc model has two components:
            reconstruction loss and the Kullback-Leibler (KL) divergence loss. The
            reconstruction loss emphasizes the difference between the original and
            synthesized NMR T2 distribution, and the KL divergence loss evaluates the
            difference between the distribution of the latent vector and a unit Gaussian
            distribution.
               In the second stage of training the VAEc-NN (Fig. 7.5), a four-layered NN
            model is connected to the frozen pretrained three-layered decoder. The
            decoder trained in the first stage of VAEc-NN training is frozen to preserve
            the generalization of the T2 distributions. Only the four-layered neural
            network undergoes weight updates in the second stage of training. The
            4-layered neural network comprises one 10-dimensional input layer, one
            30-dimensional hidden layer, one 10-dimensional hidden layer, and one
            final 3-dimensional hidden layer attached to the frozen decoder network. In
            doing so, the 4-layered neural network processes the 7 mineral composition
            and 3 fluid saturation logs to produce a 3-dimensional vector, which can be
            decoded by the trained decoder to generate the 64-dimensional T2
            distribution. After the training process is complete, for purposes of testing
            and deployment, the trained four-layered neural network followed by the
            frozen three-layered decoder synthesizes the T2 distributions of subsurface
            formations by processing the seven formation mineral content logs and
            three fluid saturation logs (similar to Fig 7.3).



            4.5 LSTM architecture, training, and testing
            Long short-term memory (LSTM) network is a type of recurrent neural network
            (RNN) that operates over sequences of input/output vectors. Unlike other neural
            networks that are stateless, LSTM and RNN use the internal state at one
            timestep as one of the inputs for calculating the internal state and output at
            the next timestep. The recurrence of internal state serves as a form of
            memory that allows contextual information to flow through the network so
            that relevant outputs or hidden states from previous timesteps can be used by
            the network at the current timestep. In our implementation the LSTM
            network processes the sequence of 10 inversion-derived logs to encode an
            intermediate vector that can be used to generate the sequence of 64 discrete
            values that constitute the 64-dimensional NMR T2 distribution (Fig. 7.6).
   225   226   227   228   229   230   231   232   233   234   235