Page 263 - Machine Learning for Subsurface Characterization
P. 263

Shallow and deep machine learning models Chapter  8 227



               TABLE 8.2 Computational times (in seconds) for training the
               shallow-learning models
                           OLS    LASSO     ElasticNet  SVR   kNNR     ANN
               Inverted log  0.002  0.33    0.009      0.33   0.0010   0.32
               Raw log     0.005  0.34      0.009      0.21   0.005    0.20





                Median R2 for NMR T2 synthesis using OLS, LASSO, and ElasticNet models
             are relatively similar. For NMR T2 synthesis, grid search was performed to find
             the best penalty parameters (hyperparameter) for LASSO and ElasticNet models.
             Best performance was achieved when the penalty parameters are very small
             values, which indicates that none of the features are redundant, trivial, or
             highly correlated. SVR model performs well on inverted well logs but not on
             raw logs. This indicates that the feature space created by the inverted logs lead
             to a better containment of the data points within the predefined margin. kNNR
             synthesizes the log by computing the weighted average of targets for the k-
             nearest training points around the testing points; nonetheless the algorithm can
             be computationally intensive for large-sized, high-dimensional dataset due to
             the numerous distance calculations that are performed to find the neighbors.
             With raw logs, kNNR exhibits the best log-synthesis performance with a R2
             of 0.68, and SVR performs the best on inverted logs with R2 of 0.74. The
             computational times for model training range from 0.001 to 0.3 s. The models
             are run on a Dell workstation with Intel 6-core Xeon CPU and 32-GB RAM.
             kNNR and OLS are the fastest on the raw log with the best accuracy. ANN,
             SVR, and LASSO models require large computational time, especially when
             processing the inverted log.
                Figs. 8.2 and 8.3 compare synthesized and measured NMR T2 distributions.
             The OLS, LASSO, and ElasticNet have intermediate performance. When
             processing inverted log, SVR and ANN exhibit good log-synthesis
             performance (Fig. 8.2). However, SVR’s log-synthesis performance drops to
             0.46 when using raw logs as input (Fig. 8.3). ANN seems to overfit with raw
             logs. ANN-generated NMR T2 is not smooth. OLS, LASSO, and ElasticNet
             models cannot handle nonlinear relationships required to synthesize the
             NMR T2 by processing the easy-to-acquire logs. On the other hand, ANN
             model with 200 neurons each in two hidden layers has relatively large
             estimation capacity but ignores the basic spatial dependency among the 64
             bins of the NMR T2 spectra. In the following section, we use four different
             deep learning models to synthesis NMR T2 distribution with an emphasis on
             learning the spatial and sequential features of typical NMR T2 distribution
             and then relating these multiscale features with the easy-to-acquire logs.
   258   259   260   261   262   263   264   265   266   267   268