Accepted Manuscript A parametric study of machine learning techniques in petroleum reservoir permeability prediction by integrating seismic attributes and wireline data Fatai Anifowose, Abdulazeez Abdulraheem, Abdullatif Al-Shuhail PII:
S0920-4105(19)30124-X
DOI:
https://doi.org/10.1016/j.petrol.2019.01.110
Reference:
PETROL 5754
To appear in:
Journal of Petroleum Science and Engineering
Received Date: 7 September 2018 Revised Date:
29 January 2019
Accepted Date: 30 January 2019
Please cite this article as: Anifowose, F., Abdulraheem, A., Al-Shuhail, A., A parametric study of machine learning techniques in petroleum reservoir permeability prediction by integrating seismic attributes and wireline data, Journal of Petroleum Science and Engineering (2019), doi: https:// doi.org/10.1016/j.petrol.2019.01.110. This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT 1 2 3 4
A PARAMETRIC STUDY OF MACHINE LEARNING TECHNIQUES IN PETROLEUM RESERVOIR PERMEABILITY PREDICTION BY INTEGRATING SEISMIC ATTRIBUTES AND WIRELINE DATA
5
Fatai Anifowosea, Abdulazeez Abdulraheemb, Abdullatif Al-Shuhailc The Research Institute, Center for Petroleum and Minerals, b
7
Department of Petroleum Engineering, c
8 9 10
Department of Earth Sciences,
RI PT
a
6
King Fahd University of Petroleum and Minerals, Dhahran 31261, Saudi Arabia. a
[email protected],
[email protected],
[email protected]
11
One of the recipes for the big data and artificial intelligence paradigms is multi-dimensional
13
data integration for improved decision making in petroleum reservoir characterization.
14
Various machine learning (ML) techniques have been applied. However, there is still ample
15
room for improvement. This paper carries out a rigorous parametric study to investigate the
16
comparative performance of common and sophisticated ML techniques in the estimation of
17
the permeability of a carbonate reservoir in the Middle East. The study integrates seismic
18
attributes and wireline data for improved permeability prediction. The effects of tuning
19
hyperparameters on the performance of the techniques are also studied. The techniques are
20
evaluated on two versions of the seismic-log integrated data: globally-averaged and depth-
21
matched. The results show that using the depth-matched dataset gives marginal improvement
22
on the permeability prediction as reflected in the higher correlation coefficient and lower
23
errors than the globally-averaged version. The outcome of this study will assist users of ML
24
techniques to make informed choices on the appropriate ML techniques to use in petroleum
25
reservoir characterization for more accurate predictions and better decision making especially
26
when faced with limited and sparse data.
27 28 29 30
Keywords: Computational intelligence; reservoir characterization; seismic attributes; wireline log; depth-matching; global averaging.
31
1.
32
Over the years, machine learning (ML) has attracted keen attention in various areas of
33
petroleum engineering and geosciences applications. Many successful implementations of
34
this science with real oil/gas industry cases have encouraged considerable application of these
35
techniques to predict reservoir permeability in uncored wells. Some of the areas in which ML
36
techniques have been successfully utilized in the petroleum industry include reservoir
37
characterization, seismic pattern recognition, reservoir properties prediction, lithofacies
AC C
EP
TE D
M AN U
SC
12
INTRODUCTION
1
ACCEPTED MANUSCRIPT identification, well production prediction, and various reservoir portfolio management (Duch
2
et al. 1997; Hosmer and Lemeshow 2000; Guojie 2004; Salah et al. 2005).
3
Permeability is an important reservoir property and a major ingredient of reservoir
4
characterization and simulation processes. It is part of the inputs used to determine
5
hydrocarbon production rate, estimation of recovery, optimal well placement, downhole
6
pressure, fluid contact evaluation, etc. A more accurate prediction of reservoir permeability
7
will lead to an integral improvement in the exploration and development of a field. More
8
accurate prediction of permeability in uncored reservoirs has remained a challenge in the
9
petroleum industry.
RI PT
1
The most reliable estimation of reservoir permeability is the special core analysis (SCAL)
11
performed on core samples in the laboratory. Both the coring process and the consequent
12
SCAL process are expensive and require huge manpower. Core samples are only available
13
over limited reservoir intervals. The reservoir simulation process requires that SCAL
14
measurements are available for the whole length of a reservoir. Hence, there has been the
15
need to develop relationships based on existing and abundant wireline and SCAL data that
16
can be used to predict reservoir permeability for uncored wells. The utility of ML techniques
17
in petroleum reservoir characterization is one of the alternative methods to reduce the need for
18
extensive coring. Traditional methods of permeability estimation are based on statistical
19
regressions and empirical equations that use correlations between wireline data and SCAL.
20
The linear regression techniques assume and establish a linear relationship between the input
21
and target variables. This may not be valid for a robust permeability prediction as the
22
variables are believed to be highly nonlinear. The empirical equations are derived based on
23
local experiments that may not be applicable in different locations.
24
The recent advancement in seismic attributes derivation from seismic volumes has given a
25
boost to the reservoir characterization process (Brahmantio et al. 2012; Maity and Aminzadeh
26
2012). This has allowed the integration of such diverse data in history matching for improved
27
reservoir simulation (Kazemi and Stephen 2010; Jin et al. 2011), with production data for
28
improved reservoir portfolio management (Nalonnil and Marion 2010; Liang et al. 2011;
29
Suman et al. 2011), and with stratigraphic data for improved forward modeling
30
(Lertlamnaphakul et al. 2012). Previous studies have investigated the prediction of reservoir
31
permeability from seismic attributes (Behzadi et al. 2011; Liang et al. 2011). This is the
32
rationale for this study. We aim to extend the existing effort of predicting reservoir
33
permeability by integrating seismic attributes and well logs to achieve improved results.
34
Leveraging recent developments in ML and seismic attributes extraction along with the
35
wireline data currently being used can have potential utility in improving the prediction of full
36
permeability profile in a given reservoir where seismic attributes and wireline data are
37
available without recourse to an extensive coring process. Artificial Neural Networks (ANN)
AC C
EP
TE D
M AN U
SC
10
2
ACCEPTED MANUSCRIPT and other traditional ML techniques have over some decades been used to predict reservoir
2
permeability and other petroleum reservoir properties. Such techniques, however, have
3
various disadvantages that impose structural and technical limitations that affect their
4
predictive performance. In the recent time, some more sophisticated ML techniques have been
5
successfully applied in other fields. It is essential for such techniques to be explored for
6
improved petroleum reservoir characterization through a rigorous comparative study that
7
assesses their applicability and performance in different data scenarios.
8
This study explores the comparative performance of traditional and more sophisticated ML
9
techniques in the prediction of petroleum reservoir permeability using integrated seismic
10
attributes and wireline data. The choice of these methods is based on their common use in the
11
petroleum industry literature and research community, and to demonstrate their comparative
12
performance as professionals transit from the traditional to the more sophisticated methods.
13
The outcome of this study will assist researchers in the petroleum industry to decide on which
14
set of ML techniques to use in cases of different operational and data scenarios such as with
15
limited and high-dimensional data.
16 17
2.
18
INTELLIGENCE
19
One of the most fundamental equations used to calculate permeability is the generalized
20
Kozeny-Carman equation (Duch et al. 1997) that approximates the fluid flow in porous media
21
as given in equation 1.
26 27 28 29 30 31
where k is in µm2 and Sgv is in µm-1.
……………..(1)
is a function that characterizes the
geology of porous media and variations in pore geometry (Blasingame 2008, Orodu et al. 2009).
is the shape factor, is tortuosity, and
is the specific area.
Equation 1 can be rewritten as follows:
0.0314
32 33 34 35
=
AC C
25
EP
23
TE D
PERMEABILITY PREDICTION AND COMPUTATIONAL
22
24
M AN U
SC
RI PT
1
=
where k is in mD.
3
……………..(2)
ACCEPTED MANUSCRIPT In addition to the independent terms in the equations above, reservoir permeability has been
3
predicted separately from log measurements and seismic attributes. Details of log
4
measurements and their relationship with various reservoir properties such as permeability are
5
found in a number of geophysical literatures such as Sengel (1981), Etnyre (1989), Darling
6
(2005), Schlumberger Oil Field Glossary (2018), and while those of seismic attributes have
7
been well documented in Taner et al. (1979), Chen and Sidney (1997), Walls et al. (1999),
8
and Brown (2001).
9
The utility of ML techniques has been widely reported in various fields including petroleum
10
engineering and the geosciences. This interdisciplinary collaboration has succeeded in
11
creating the necessary synergy among computer scientists, petroleum engineers, and
12
geoscientists in improving the petroleum reservoir characterization process. ML is the
13
computational aspect of artificial intelligence (AI) that deals with algorithms that
14
automatically learn from data without explicit programming. More details about the AI field
15
and the related subfields of computational intelligence and ML have been extensively
16
discussed in Fulcher (2008). Details of the various ML techniques that have been applied to
17
predict various reservoir properties can be found in Duch et al., (1997), Hosmer and
18
Lemeshow (2000), Guojie (2004), Salah et al. (2005), Lauría and Duchessi (2006), and El-
19
Sebakhy (2011). Such reservoir properties include PVT, drive mechanism, well spacing, well-
20
bore integrity, structure and seal, diagenesis, porosity and permeability while such ML
21
techniques include K-Nearest Neighbor, Multilayer Perceptrons, Logistic Regression, Radial
22
Basis Function, Naïve Bayes, Random Forests, Functional Networks, Bayesian Belief
23
Networks, Support Vector Machines, Probabilistic Neural Networks, Adaptive-Neuro Fuzzy
24
Systems, Artificial Neural Networks, and Decision Trees.
25
The main objective of this study is to investigate some of these traditional techniques as well
26
as the more sophisticated ones with respect to their performance on limited and integrated
27
data scenarios for the possibility of obtaining improved predictive performance of such
28
models. In addition to demonstrating the capability of the proposed methods in the petroleum
29
industry, they are essentially relevant to handle the sparse and limited datasets we handle in
30
the petroleum industry. Core measurements are limited to few samples in the entire length of
31
a reservoir due to their huge cost, enormous time consumption, and excessive manpower
32
requirement. The utility of seismic datasets is also limited due to their low resolution.
33
The following sections give brief overviews of the ML techniques that are implemented and
34
evaluated in this study.
AC C
EP
TE D
M AN U
SC
RI PT
1 2
35
4
ACCEPTED MANUSCRIPT 3.
MACHINE LEARNING TECHNIQUES IMPLEMENTED
2
3.1
ARTIFICIAL NEURAL NETWORKS
3
Artificial neural networks (ANN) is a machine learning technique developed in an attempt to
4
mimic the way the human brain works. It consists of an array of processing units called
5
neurons that communicate with other neurons by sending weighted signals. They have been
6
successfully applied to build various models for pattern recognition, signal processing,
7
robotics, control, and decision making (Ali 1994). The most widely used neural network
8
architecture for handling prediction and regression problems is the multilayer perceptron
9
(MLP) with back-propagation. The MLP architecture produces its output by learning the
10
patterns embedded in the data. More details of ANN, its different architectures, and
11
underlining mathematical bases can be found in Petrus et al. (1995).
12 13
3.2
14
Functional networks (FN) is a generalization of ANN. It is similar to ANN by consisting of
15
different layers of neurons connected by links but different in that each computing unit or
16
neuron performs a simple calculation typically a scalar, typically monotone, function f of a
17
weighted sum of inputs. The function associated with each neuron is fixed and the weights are
18
learned automatically from data using some common algorithms such as the least-squares
19
fitting (Castillo et al. 2001). FN ensures that the weights are suppressed when the functions
20
are undergoing learning. The functions can be multivariate or composed of single variables.
21
Solving the system of equations leads to a great simplification of the initial functions f
22
associated with the neurons (Castillo et al. 2001; El-Sebakhy 2011). The FN architecture can
23
be mathematically defined as:
SC
RI PT
1
TE D
M AN U
FUNCTIONAL NETWORKS
Assume that we have a neuron with s inputs: (x1,…, xs) and k outputs: (y1,…, yk), then
26
we assume that there exist k functions Fj; j = 1,…, k, such that yj = Fj(x1,…, xs); j =
27
1,…, k.
AC C
28
EP
24 25
29
More details about the FN architecture can be found in Castillo et al. (2001), El-Sebakhy
30
(2011), and Anifowose and Abdulraheem (2011).
31 32 33
3.3
34
Support vector machines (SVM) is a supervised learning method that belongs to the class of
35
generalized linear classifiers and a special case of Tikhonov Regularization. It was initially
36
developed for classification until Vapnik et al. (1995) developed a new ε-sensitive loss
37
function technique that seeks to minimize an upper bound of the generalization error ideal for
SUPPORT VECTOR MACHINES
5
ACCEPTED MANUSCRIPT 1
regression problems. It is founded on the statistical learning theory and follows the principle
2
of structural risk minimization. It maps input vectors to a higher dimensional space where a
3
maximal separating hyperplane is constructed (Peng and Wang 2009; Anifowose et al.
4
2014). SVM has been reported to exhibit excellent predictive performance. The kernel
5
functions that are commonly used in SVM are:
6
8 9
Linear:
RI PT
7
……………..(3)
RBF:
SC
……………..(4)
Polynomial:
12 13
M AN U
10 11
……………..(5)
where param is the kernel parameter in the SVM file.
14 3.4
TYPE-2 FUZZY LOGIC SYSTEM
16
Type-2 fuzzy logic system is a generalization of the Type-1. Details about Type-1 fuzzy logic
17
system can be found in Zadeh (1975) and Lee (1990). Fuzzy sets (FS) enable modeling in
18
uncertain and ambiguous data conditions. They are capable of solving ill-posed problems.
19
Each element in a FS is mapped to [0,1], a set of real numbers between 0 and 1 (inclusive) by
20
a membership function. A type-2 FS allows the incorporation of uncertainty about the
21
membership function into fuzzy set theory. It would be noted that in the absence of
22
uncertainties, a type-2 FS reduces to a type-1 FS, which is similar to probability reducing to
23
deterministic approaches when unpredictability vanishes. The membership function of a type-
24
2 FS is three-dimensional where the third dimension is the value of the membership function
25
at each point on its two-dimensional domain. This is referred to as the footprint of uncertainty
26
that provides new design degrees of freedom for handling uncertainties (Karnik and Mendel
27
1999; Mendel and John 2002). Further details of type-2 FS and its successful applications
28
can be found in Mendel (2006).
29 30
3.5
31
Decision tree (DT), also known as classification or regression tree, is a machine learning
32
technique that uses the principle of information gain theory to map observations about a
33
problem to conclusions about the problem's target value. In the DT structure, leaves represent
AC C
EP
TE D
15
DECISION TREES
6
ACCEPTED MANUSCRIPT the outcome classes and branches represent conjunctions of features that lead to the outcome
2
classes. DT learning is a commonly used data mining method that is capable of creating
3
models that predict the value of a target variable based on one or more input features (White
4
and Liu 1994). The DT algorithm learns by splitting the input features into subsets based on a
5
value test. This process, called recursive partitioning, is repeated on each derived subset in a
6
recursive manner. The recursion is stopped when further splitting no longer adds value to the
7
predictions or when the subset at a node has the same value of the target variable. More
8
details about the DT algorithm can be found in Sherrod (2008).
9 10
3.6
11
Extreme learning machine (ELM) is a unified framework of generalized single-layer feed-
12
forward networks (SLFFN). ELM runs much faster than conventional methods. It
13
automatically determines all the network parameters by randomly choosing the input weights
14
and analytically estimating the output weights of the SLFFN. This gives it the capability to
15
avoid trivial human intervention and makes it efficient in online and real-time applications. It
16
was proposed to overcome the drawbacks of gradient-based algorithms such as ANN and
17
SVM (Han et al. 2006). Mathematically, the output function of an ELM model with L hidden
18
nodes can be given as:
M AN U
SC
EXTREME LEARNING MACHINES
RI PT
1
19
"
TE D
!" # = $ %& ℎ& # &(
22 23 24 25 26 27
The output function of the ith hidden node is further given as:
EP
21
where %& is the output weight and ℎ& is the output function of the ith hidden node.
AC C
20
……………..(6)
ℎ& # = ) *& , ,& , #
……………..(7)
where *& and ,& are the parameters of the ith hidden node.
The output of the hidden layer is expressed as:
28 ℎ # = [) ℎ # , ℎ # , … , ℎ" # ]
29 30
Finally, with N samples the output matrix of the hidden layer is given as:
31
7
……………..(8)
ACCEPTED MANUSCRIPT ℎ 3 ℎ 2 0=2 2 2 1ℎ
# ) * , , , # ) * , , , # … ) *" , ," , # 7 3 # ) * , , , # ) * , , , # … ) *" , ," , # 6 2 . 6 2 . . . = . 6 2 . . . . 6 2 . . . #4 5 1) * , , , #4 ) * , , , #4 … ) *" , ," , #4
7 6 6 6 6 5
……………..(9)
1 3
……………..(10)
More details about this technique can be found in Huang et al. (2004, 2006)
6
M AN U
4 5
9 39 7 2.6 8=2 . 6 2 6 2.6 194 5
RI PT
and the training data target matrix is given as:
SC
2
4.
METHODOLOGY
8
4.1.
DESCRIPTION OF DATA
9
Seismic attributes and wireline data from 17 wells in a Middle East carbonate reservoir are
10
used to design and evaluate the models. The seismic data are available in four intervals: 10ms,
11
20ms, 30ms, and 40ms corresponding to four seismic zones. Both datasets were prepared,
12
processed, and checked to the highest level of quality. The major sources of uncertainty in
13
wireline log data are depth shifting, hole cleaning condition, borehole size, cycle-skipping,
14
stuck logging tools, and mud filtrate invasion. For the seismic data, a major source of
15
uncertainty is the presence of noise data (coherent, such as ground roll, and incoherent, such
16
as scattering by near-surface sources). Others include the presence of artifacts (such as gain,
17
NMO stretch, and frequency filtering), and wavelet phase changes caused by natural sources
18
such as absorption, deconvolution, and zero phasing. Correcting such uncertainties and
19
anomalies requires estimating all parameters contributing to the uncertainties and then adding
20
them together in a way that displays the confidence level. We ensured that both the seismic
21
and wireline data used in this study are corrected against all these uncertainties and
22
anomalies. Details of the steps taken to ensure the highest quality standard of the datasets are
23
beyond the scope of this paper.
24
Table 1 and 2 show the seismic attributes and wireline logs forming the integrated dataset.
25
Table 3 shows the result of the statistical analysis comprising eight parameters namely
26
average, median, minimum, maximum, kurtosis, skew, standard deviation and variance of the
27
combined seismic-log data.
AC C
EP
TE D
7
28
8
ACCEPTED MANUSCRIPT 1 2 3
Abbreviation IP IF AL HE RMSA
4 5 6 7
Full Meaning Instantaneous Phase Instantaneous Frequency Arc Length Half Energy RMS Amplitude
RI PT
Table 1. Seismic attributes used in this study.
Table 2. Wireline logs used in this study.
M AN U
8 9 10 11
Full Meaning Neutron Porosity Log Density Log Caliper Log Gamma Ray Water Saturation Deep Resistivity
SC
Abbreviation NPL RHOB CT GR SWT DT
IF
HE
AA
AL
Average
33.96
43.99
1.56
1.06
-2.89
0.17
2.48
0.12
18.35
0.48
64.49
Median
32.70
42.85
1.40
1.05
-32.40
0.18
2.47
0.08
15.26
0.44
64.75
Minimum
16.32
16.16
0.63
1.00
-171.19
0.06
2.32
0.01
2.78
0.07
54.37
Maximum
61.84
88.85
2.93
1.16
178.92
0.25
2.67
0.59
160.26
0.99
74.75
Kurtosis
TE D
Table 3. Basic Statistics of the Integrated Data. IP
NPL
RHOB
CT
GR
SWT
DT
0.23
-1.05
-0.95
-1.50
-1.01
-0.90
3.33
51.57
-0.93
-1.01
0.33
0.49
0.37
0.49
0.17
-0.30
0.19
1.82
6.88
0.25
0.03
STD
10.44
15.43
0.65
0.04
110.23
0.05
0.09
0.12
19.14
0.25
5.54
108.90
238.05
0.43
0.00
12150.40
0.00
0.01
0.01
366.26
0.06
30.69
Variance
EP
-0.33
Skew
It should be noted that the seismic attributes were extracted from the seismic cubes of the
15
respective wells using equations that are well established in literature. Details of such
16
equations are not the focus of this paper. Also, both the seismic and wireline log datasets were
17
cleaned and quality-checked by the relevant department before using them for this work.
18
Hence, the quality of the data is assured. It is well known that quality of data affects the
19
performance of models especially those based on machine learning.
20 21
4.2
22
The datasets are integrated in two ways following the two schools of thought on seismic
23
attributes and wireline data integration. One school postulated that only globally averaging
AC C
12 13 14
DATA PRE-PROCESSING
9
ACCEPTED MANUSCRIPT the log data according to the seismic zones without considering the exact well locations is
2
sufficient while the other school suggests that both data should be matched and averaged at
3
the exact points where the well locations intercept the seismic zones. The integrated data
4
arising from the former, with no depth precision, is named Seismic_Log#1. The one arising
5
from the latter led to a loss of a number of samples as those data points fell outside of the
6
seismic zones, hence having no match. This is named Seismic_Log#2. In each case, the
7
stratified sampling approach applied in some previous studies such as El-Sebakhy (2004,
8
2011), Anifowose and Abdulraheem (2010), and Anifowose et al. (2011) is used to divide the
9
dataset into training and validation subsets. With this stratification approach, a randomly
10
selected 70% of the integrated dataset was reserved for training while the remaining was used
11
for validation. The choice of the 70:30 stratification strategy is motivated by the
12
recommendation of Anifowose et al. (2016). This random selection is preferred over the
13
conventional fixed stratification strategy because it gives each data sample an equal chance of
14
being selected for either purpose. This ensures fairness and gives more credence to the results
15
of the study.
M AN U
SC
RI PT
1
16 4.3
CRITERIA USED FOR PERFORMANCE EVALUATION
18
The performance of the proposed ML techniques is evaluated using commonly used statistical
19
model evaluation criteria: correlation coefficient (R2), root mean squared error (RMSE), and
20
mean absolute error (MAE).
21
The R2 is a statistical measure of the strength of a relationship and the trend between n pairs
22
of two variables, x and y. It is mathematically expressed as:
TE D
17
23
EP
25
: =
AC C
24
; ∑=
; ∑ =>
∑=
∑= ∑> ; ∑>
∑>
……………(11)
26
The RMSE is a measure of the degree of spread between the actual x values around the
27
average of the predicted y values. It can be expressed as:
28
:? @ =
∑B ACD =A >A ;
……………..(12)
29
The MAE is the mean of the absolute errors of the predicted y values relative to the actual x
30
values and is given by:
31 32
?E@ = ; ∑;&( |#& − H& |
10
……………..(13)
ACCEPTED MANUSCRIPT 1 4.4
PARAMETRIC STUDIES
3
The six proposed ML techniques were implemented on the integrated seismic and wireline
4
log dataset. The development environment is MATLAB 2018a installed on a 64-bit Core i7
5
2.8 GHz processor running on a Window 10 operating system with a RAM of 16 GB. The
6
general principle employed in the determination of the optimal parameters for the six
7
proposed methods is presented in the following pseudocode:
8 9
RI PT
2
1. For each parameter Pi belonging to a set of parameters P1, P2, …, Pn, do the following. 2. Establish a reasonable search range for the selected parameter, Pi.
11
3. For each value Ni in the search range sequentially
12
4. Note the effects on the model performance using the evaluation criteria.
13
5. Continue the next parameter Ni until Nn.
14
6. Determine the value within the search range that gave the best model performance
16
M AN U
15
SC
10
according to the model evaluation criteria.
7. Using the best value for N, continue the next parameter Pi until Pn.
17
A number of runs were carried out with each technique to ascertain the optimal model
19
parameters as presented in the pseudocode above. Representative results of the various runs
20
with both versions of the dataset are presented in Figures 1 through 9. The objective of this
21
exercise is to obtain optimal models that completely avoid the challenges of overfitting and
22
underfitting. Overfitting is indicated by an increasingly wider gap between the training and
23
testing curves as the search progresses from the low end of the range to the high. Underfitting
24
is indicated by both training and testing curves moving together in low performance in terms
25
of low CC or high RMSE and MAE.
26
Figure 1 shows the results of searching for the optimal number hidden neurons for ANN over
27
a range of 1 to 50. The choice of the limit of 50 neurons follows the work of Beale et al.
28
(2015) which suggested that no more than 50 neurons is sufficient for ANN to solve any
29
problem. This suggestion is found to be valid as there is no sign of convergence after 50
30
neurons but rather a continuation of the same erratic pattern. The two datasets demonstrate
31
and confirm the erratic behavior of ANN as reported in the literature. ANN suffers from local
32
optima as several points (such as 6, 14, 27, and 40 for Run 1 and 6, 23, and 33 for Run 2)
33
could qualify for optimality. With this observed erratic pattern, it would be preferred to go for
34
a lower number of neurons that give the same level of performance than a higher one.
35
For the SVM model, Figure 2 shows the results of the search for the optimal regularization
36
parameter, C, for SVM over a range of 0 to 15000 for Run 1 (a) and Run 2 (b). From the
AC C
EP
TE D
18
11
ACCEPTED MANUSCRIPT plots, it could be confirmed that there was no further improvement in the model performance
2
beyond 15000. Rather there is increased overfitting. It is observed that the point of optimality
3
for each model Run is where the training plot meets the testing. Before this point is
4
underfitting and after it is overfitting. As mentioned earlier, the goal of any machine learning
5
project is to maintain a balance between these two extremes. Figure 3 shows the results of the
6
search for the optimal error goal, lambda, another hyperparameter for SVM. It was also
7
confirmed that going beyond the experimented range does not improve the model
8
performance but rather indicates increased overfitting. Figure 4 shows the results of the search
9
for the optimal penalty for overfitting, epsilon, over a range of 0 to 5. It is also confirmed
10
from the plot that there was no improvement on the model performance beyond 5 as
11
overfitting increases. Figure 5 shows that the SVM model’s kerneloption hyperparameter
12
does not have any effect on the model’s performance.
13
Figure 6 shows the result of the search for the T2FLS’s learning rate, Alpha, over a range of 0
14
to 5. There was no sign of possible improvement in the model performance going beyond 5. It
15
rather indicates that the lower the value, the better for the model. The same was observed for
16
the two Runs 1 and 2. For DT, the results of the search for the optimal number of tree splits
17
are shown in Figure 7. The plots confirmed that there is no gain in performance improvement
18
beyond 50. Similar to the Alpha parameter of SVM, the search result indicates that the lower
19
the value of K, the better for the model.
20
In Figure 8, the results of the search for the optimal number of hidden neurons for the ELM
21
model are presented. Similarly, the plots confirmed that there is no gain in the performance
22
improvement going beyond the limit of 50. Both search results show progressive increase in
23
the training performance but erratic pattern in the testing. This behavior is similar to that of
24
ANN. This is not surprising since ELM is a special case of ANN with a single hidden layer.
25
Several points (such as 24, 31, 36, 45, and 48 for Run 1, and 26 and 43 for Run 2) also quality
26
for optimality. This confirms that ELM overfits with the data used for this study.
27
A summary of the optimal parameters obtained from the two Runs for the six methods is
28
presented in Table 4. The points of optimality were obtained by taking the value of the
29
parameter corresponding to the points on the plots where the R2 (labeled as CC) of the
30
training performance are the closest to those of the testing. This implies points of no or the
31
least overfitting.
AC C
EP
TE D
M AN U
SC
RI PT
1
32 33 34 35 36
12
ACCEPTED MANUSCRIPT 1 2 3
Table 4. Summary of the Optimized Parameters for All Models. Technique
Values of the Optimized Parameters Run 1 • Number of Hidden Layers = 2 • Number of Neurons in each layer = 40 Run 2 • Number of Hidden Layers = 2 • Number of Neurons in each layer = 33 ● Degree of Polynomial = 1 Run 1 • Regularization Parameter, C = 8000 ● Error Allowance, Lambda = 0.008 ● Penalty for Overfitting, epsilon = 0 ● Type of Kernel = Polynomial ● Kernel Step Size = Constant
RI PT
ANN
SC
FN
M AN U
SVM
Run 2 • Regularization Parameter, C = 7000 ● Error Allowance, Lambda = 0.002 ● Penalty for Overfitting, epsilon = 2 ● Type of Kernel = Polynomial ● Kernel Step Size = Constant ● Learning Rate, Alpha = 0.3 ● Learning Algorithm = Gini ● Minimum Number of Tree Splits = 7 Run 1 • Number of Hidden Neurons = 36 ● Activation Function = Sigmoid
DT
Run 2 • Number of Hidden Neurons = 43 ● Activation Function = Sigmoid
AC C
4 5
EP
ELM
TE D
T2FLS
6
After the models have been optimally trained, they could be used to predict the permeability
7
for new/uncored wells. The input data should be composed of the same integrated set of
8
seismic and wireline logs in the same order they were presented to the models during the
9
training. This is to ensure that the coefficients or weights generated for each input feature are
10
used for the same feature during the prediction phase.
11 12
5.0
13
The optimized parameters obtained from the exhaustive search conducted and reported in
14
section 4.3 are used to build the six CI models and their performance results are compared.
RESULTS AND DISCUSSION
13
ACCEPTED MANUSCRIPT 1
The comparative results obtained on the Seismic_Log#1 dataset are presented in Table 5
2
while the representative comparative results of training and validation are shown in Figures 9
3
and 10. Those obtained on the Seismic_Log#2 dataset are presented in Table 6 and similar
4
representative training and validation comparative performances are shown in Figures 11 and
5
12.
6 7 R
RMSE Test
Train
Test
Train
ANN
0.69
0.64
0.68
FN
0.75
0.70
SVM
0.78
0.73
MAE
Min Error Test
Train
Test
Train
0.76
0.52
0.61
-1.70
-1.86
0.59
0.56
0.47
0.42
-1.4
-1.26
0.56
0.51
0.46
0.39
-1.21
-1.16
0.74
0.35
0.63
0.90
0.45
0.67
-2.07
DT
0.95
0.55
0.24
0.78
0.14
0.57
-0.61
ELM
0.45
0.23
0.86
1.09
0.67
0.84
-1.43
8 9
Exec. Time (s) Test
Train
1.86
1.35
1.32
0.001
1.17
0.70
0.06
0.001
1.29
0.58
0.25
0.001
-2.2
0.67
1.33
0.98
0.3
-1.93
1.08
1.30
0.17
0.001
-2.67
2.44
1.88
0.03
0.031
M AN U
T2FLS
Max Err Test
Train
SC
ML Tech.
RI PT
Table 5. Results of Permeability Prediction using Seismic_Log#1. 2
Table 6. Results of Permeability Prediction using Seismic_Log#2. 2
Technique
R Train
RMSE Test
Train
Test
MAE Train
Min Error
Test
Train
Test
Max Err
Train
Test
Exec. Time (s) Train
Test
0.64
0.42
0.70
1.02
0.34
0.82
-3.66
-2.38
0.78
0.81
1.15
0.001
FN
0.84
0.76
0.40
0.567
0.32
0.47
-1.02
-1.16
0.84
0.70
0.062
0.001
0.37
-1.12
-0.14
0.70
0.57
0.33
0.001
TE D
ANN SVM
0.84
0.85
0.41
T2FLS
0.87
0.57
0.45
0.77
0.48
0.34
0.24
0.64
-0.57
-1.44
1.42
1.62
0.655
0.280
DT
1.00
0.73
0.00
0.60
0.00
0.46
ELM
0.71
0.41
0.53
0.60
0.42
0.48
0.00
-1.03
0.00
1.36
0.17
0.001
-1.34
-1.006
0.96
1.64
0.031
0.001
From the results of the permeability prediction using the Seismic_Log#1 dataset shown in
13
Table 5, the DT model exhibit the highest training R2 but gave a poor testing performance
14
(Figure 9). This is a demonstration of overfitting. T2FLS also overfitted the data. However,
15
ELM showed underfitting as both the training and testing R2 values were poor. The SVM
16
model outperformed the ANN and FN models. It shows acceptable performance without
17
overfitting. A similar performance trend was demonstrated by the models in terms of the other
18
performance measures such as the mean absolute error (MAE) (Figure 10). The T2FLS, DT
19
and ELM models gave higher error values relative to the other models. ANN, FN and SVM
20
show more stability with the dataset as they do not show signs of overfitting or underfitting.
21
The performance of SVM in terms of MAE agrees with that of the CC by exhibiting the least
22
errors, thereby emerging as the best model for this dataset.
AC C
EP
10 11 12
14
ACCEPTED MANUSCRIPT In terms of execution time, the dataset could not clearly show the difference among the
2
models as they mostly executed at closely the same amount of time. However, T2FLS showed
3
that it took the most time to execute compared to the ELM model that took the least (Table 5).
4
The results obtained in this work agree perfectly with the performance of each of the models
5
as reported in the literature. FN (Castillo et al 2001; El-Sebakhy 2011) and SVM (Shawe-
6
Taylor & Cristianini 2004) are generally known to be computationally stable and fast in
7
execution. ANN takes more time especially when it has to handle large dataset (Petrus et al.
8
1995). DT (Sherrod 2008) and ELM (Han et al. 2006) have also been described as light-
9
weight techniques but susceptible to overfitting with small datasets while T2FLS (Mendel
10
2003) took the most time due to its complex algorithm involving a gradient descent method of
11
optimizing its performance and the conversion between Type-1 and Type-2 fuzzy sets during
12
its input and output processes.
13
The signs of overfitting/underfitting are further graphically shown in Figure 9 with the wide
14
gap between the training and testing performance curves of the T2FLS, DT, and ELM
15
models. However, the competitive performance of ANN, FN and SVM are also shown with
16
SVM demonstrating the best performance in the overall with the highest R2 (labeled as CC)
17
This relative performance of the models complemented Figure 10 which shows a similar trend
18
with SVM exhibiting the most stability, highest R2 and the lowest error indices.
19
On the Seismic_Log#2 dataset, the performance of the models follows a trend similar to that
20
of the Seismic_Log#1 dataset. The T2FLS, DT and ELM models also exhibit the
21
characteristics of overfitting. ANN showed less performance on this dataset than on the
22
previous. However, FN and SVM improved in their competitive performance with SVM
23
remaining in better performance. Despite that the DT model overfitted the dataset, the
24
performance on this dataset is better than on the former. The error measures are in good
25
agreement with the R2 results as the T2FLS, DT, and ELM models have the highest errors
26
while the FN and SVM models showed the least. In the overall, the SVM model kept the best
27
performance in both data conditions.
28
These performance demonstrations are in perfect agreement with literature as discussed
29
earlier. The reduction in the performance of ANN on this dataset agrees with its reported
30
unstable behavior (Petrus et al. 1995). This is due to its often being trapped in the local
31
optimum. The reduction in the number of samples also affect the performance of the ANN
32
model. ANN has been shown to require more data points for effective generalization. The FN
33
and SVM models, on the other hand, have the capability to handle small data. However, the
34
improvement in the performance of the other models can be attributed to the improvement in
35
the quality of the dataset. It would be recalled that this dataset (Seismic_Log#2) is more
36
rigorously pre-processed than the former (Seismic_Log#1). The Seismic_Log#1 dataset was
37
simply a result of the global averaging of each of the four seismic zones. However,
AC C
EP
TE D
M AN U
SC
RI PT
1
15
ACCEPTED MANUSCRIPT Seismic_Log#2 takes into consideration the proper depth matching ensuring that the same
2
depth locations in the log data correspond to the same locations in the seismic traces.
3
In terms of execution time, the trend follows that of the former dataset as the T2FLS model
4
continued to show its complexity as it took more time to execute for training and testing than
5
the others. Figure 11 confirms the emergence of the SVM model as the best performing
6
technique in terms of training and validation R2 while Figure 12 follows the same trend by
7
showing SVM with the least mean absolute error. Other models either showed signs of
8
overfitting or generally poor performance (underfitting).
9 6.
11
In this study, we have presented the results of a rigorous and comprehensive parametric study
12
to investigate the comparative performance of six traditional and state-of-the-art ML
13
techniques and the effects of tuning hyperparameters on them in the estimation of
14
permeability from integrated seismic attributes and wireline datasets. This study is
15
implemented on the integration of six wireline logs and five seismic attributes extracted from
16
17 wells in a giant reservoir located in the Middle East. The method of integrating the datasets
17
led to the emergence of two different versions: one globally averaged without regard to the
18
exact depth matches with wells; and the other depth-matched at the exact well locations. The
19
objective of this study is to investigate the effect of these datasets on the performance of six
20
ML techniques in the prediction of reservoir permeability.
21
After a rigorous comparative study and analysis of the results, the following conclusions are
22
drawn: •
TE D
M AN U
SC
10
23
CONCLUSION
RI PT
1
Certain techniques have hyperparameters with optimal values beyond which no further performance improvement will be achieved. These techniques are
25
consequently better suited for certain data and other processing scenarios.
27 28 29 30
•
Of all the runs of the study, SVM demonstrated the most acceptable performance with the highest R2 and the lowest error measures.
AC C
26
EP
24
•
Bringing together the best performing techniques from each of the data conditions on
a comparative scale, we found that SVM performed better on the Seismic_Log#2 than the Seismic_Log#1 dataset. This indicates that the effort made on the depth-matching
31
process and the consequent higher precision made a significant difference. On a field-
32
wide scale, this difference could make a big positive impact in exploration efficiency
33
and production capacities.
34 35
•
ANN and FN demonstrated a competitive strength by having a close match to the performance of SVM but with ANN showing more overfitting than FN.
16
ACCEPTED MANUSCRIPT 1 2
•
DT and ELM showed the most sensitivity to the datasets by exhibiting clear overfitting.
3 4 Acknowledgement The authors would like to acknowledge the support provided by King Abdulaziz City for
7
Science and Technology through the Science & Technology Unit at King Fahd University of
8
Petroleum & Minerals for funding this work under Project No. 11-OIL2144-04 as part of the
9
National Science, Technology and Innovation Plan.
10
RI PT
5 6
REFERENCES
12 13 14
Ali JK (1994) Neural networks: a new tool for the petroleum industry? In: Proceedings of the European Petroleum Computer Conference, Aberdeen, U.K., March.
15 16 17 18 19
Anifowose FA, Abdulraheem A (2010) A functional networks-type-2 fuzzy logic hybrid model for the prediction of porosity and permeability of oil and gas reservoirs. Paper #1569334237, In: Proceedings 2nd International Conference on Computational Intelligence, Modeling and Simulation, Bali, September 28 – 30, IEEEXplore, pp. 193 – 198.
20 21 22
Anifowose FA, Abdulraheem A (2011) Fuzzy logic-driven and SVM-driven hybrid computational intelligence models applied to oil and gas reservoir characterization, Journal of Natural Gas Science and Engineering, 3: 505 – 517.
23 24 25 26
Anifowose F, Khoukhi A, Abdulraheem A (2016) Investigating the Effect of Training-Testing Data Stratification on Soft Computing Techniques: An Experimental Study, Journal of Experimental and Theoretical Artificial Intelligence, 29:(3): 517 – 535.
27 28 29 30 31 32 33 34 35
Anifowose FA, Labadin J, Abdulraheem A (2011) A hybrid of functional networks and support vector machine models for the prediction of petroleum reservoir properties. Paper #45, In: Proceedings of 11th International Conference on Hybrid Intelligent Systems, Meleka, Malaysia, 5-8 December, IEEExplore, pp. 85 - 90. Anifowose FA, Adeniye S, Abdulraheem A (2014) Recent advances in the application of computational intelligence techniques in oil and gas reservoir characterization: a comparative study. Journal of Experimental & Theoretical Artificial Intelligence, 26(4): 551 – 570.
36 37 38
Behzadi H, Alvarado V, Padhi A, Mallick S (2011) CO2 saturation, distribution and seismic response in 2D dimensional permeability model. Paper Presented at the SEG 2011 San Antonio Annual Meeting.
AC C
EP
TE D
M AN U
SC
11
17
ACCEPTED MANUSCRIPT Blasingame TA (2008) The characteristic flow behavior of low-permeability reservoir systems. In: Proceeding of the SPE Unconventional Reservoirs Conference held in Keystone, Colorado, U.S.A., 10–12.
4 5 6 7
Brahmantio R, Kusuma A, Bertini F, Dufour J, Paternoster B, Leclair S (2012) Seismic amplitude loss compression for future seismic characterization studies in Peciko Field-Indonesia. Paper IPTC 14412 Presented at the International Petroleum Technology Conference held in Bangkok, Thailand, 7-9 February.
8 9
Beale M, Hagan M, Demuth H (2015) Neural network toolbox user’s guide. The MathWorks, Inc., Apple Hill Drive, Natick, US.
10 11
Brown AR (2001) Calibrate yourself of your data! A vital first step in seismic interpretation. Geophysical Prospection 49:729–733.
12 13
Castillo E, Gutiérrez JM, Hadi AS, Lacruz B (2001) Some applications of functional networks in statistics and engineering. Technometrics 43:10–24.
14 15
Chen Q, Sidney S (1997) Seismic attribute technology for reservoir forecasting and monitoring. The Leading Edge 16:445-456.
16
Darling T (2005) Well Logging and Formation Evaluation. Oxford, UK: Elsevier.
17 18 19
Duch W, Adamczak R, Jankowski N (1997) Initialization and optimization of multilayered perceptrons. In: Proceedings of the Third Conference on Neural Networks and their Applications, Kule, Poland.
20 21 22 23
El-Sebakhy EA (2004) Functional networks training algorithm for statistical pattern recognition. In Proceedings of the Ninth International Symposium on Computers and Communications 2004 Volume 2 (ISCC"04) - Volume 02, IEEE Computer Society Washington, DC, USA.
24 25 26
El-Sebakhy EA (2011) Functional networks as a novel data mining paradigm in forecasting software development efforts. Expert Systems with Applications 38(3):2187-2194.
27 28
Etnyre LM (1989) Finding oil and gas from well logs. Kluwer Academic Publishers. 249 pages.
29 30 31
Fulcher J (2008) Computational Intelligence: An Introduction. In: Fulcher J., Jain L.C. (eds) Computational Intelligence: A Compendium. Studies in Computational Intelligence, vol 115. Springer, Berlin, Heidelberg.
32 33
Guojie L (2004) Radial basis function neural network for speaker verification. A Master of Engineering thesis submitted to the Nanyang Technological University.
34 35 36
Han F, Li X, Lyu MR, Lok T (2006) A modified learning algorithm incorporating additional functional constraints into neural networks. International Journal of Pattern Recognition and Artificial Intelligence 20(2):129-142.
AC C
EP
TE D
M AN U
SC
RI PT
1 2 3
18
ACCEPTED MANUSCRIPT Hosmer D, Lemeshow S (2000) Applied logistic regression. John Wiley and Sons, 392 pages.
3 4 5
Huang GB, Zhu QY, Siew CK (2004) Extreme learning machine: a new learning scheme of feedforward neural networks. In: Proceedings of the international joint conference on neural networks, Budapest, Hungary.
6 7
Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70:489-501.
8 9 10 11 12
Jin L, Alpak FO, Hoek P, Pirmes C, Fehintola T, Tendo F, Olaniyan E (2011) A comparison of stochastic data-integration algorithms for the joint history matching of production and time-lapse seismic data. Paper SPE 14618 Presented at the SPE Annual technical Conference and Exhibition held in Denver, Colorado, USA, 30 October - 2 November.
13 14
Karnik NN, Mendel J (1999) Type-2 fuzzy logic systems. IEEE Transactions on Fuzzy Systems 7(6):643-658.
15 16 17
Kazemi A, Stephen KD (2010) Optimal parameter updating in assisted history matching of the nelson field using streamlines as a guide. In: Proceedings of the SPE EUROPEC/EAGE Annual Conference and Exhibition, 14-17, Barcelona, Spain.
18 19
Lauría EJM, Duchessi P (2006) A bayesian belief network for IT implementation decision support. Decision Support Systems 42(3): 1573-1588.
20 21 22
Lee CC (1990) Fuzzy logic in control systems: fuzzy logic controller. I. In: Proceedings of the IEEE Transactions on Systems, Man, and Cybernetics 20(2): 404418.
23 24 25 26 27
Lertlamnaphakul P, Thanaitit U, Al-Busaidi SM, Russell WR, Seusutthiya K (2012) Integrated well log electrofacies and seismic facies modeling for stratisgrahic trap identification in carbonate reservoirs, North Oman. Paper IPTC 14806 Presented at the International Petroleum Technology Conference held in Bangkok, Thailand, 7-9 February.
28 29 30 31 32
Liang L, Abubakar A, Habashy TM (2011) Improved estimation of permeability from joint inversion of time-lapse crosswell electromagnetic and production data using gradient-based method. Paper SPE 146526 Presented at the SPE Annual technical Conference and Exhibition held in Denver, Colorado, USA, 30 October - 2 November 2011.
33 34 35 36
Maity D, Aminzadeh F (2012) Reservoir characterization of an unconventional reservoir by integrating microseismic, seismic and well log data. SPE Paper 154339 presented at the SPE Western Regional Meeting held in Bakersfield, California, USA, 19-23 March.
37 38
Masters T (1993) Practical neural network recipes in C++. Morgan Kaufmann, 493 pages.
AC C
EP
TE D
M AN U
SC
RI PT
1 2
19
ACCEPTED MANUSCRIPT Mendel JM (2006) Interval type-2 fuzzy logic systems made simple. IEEE Transactions on Fuzzy Systems 14(6):808-821.
3 4
Mendel JM, John RI (2002) Type-2 fuzzy sets made simple. IEEE Trans. on Fuzzy Systems 10:117-127.
5 6
White AP, Liu WZ (1994) Bias in information-based measures in decision tree induction. Machine Learning 15:321-329.
7 8 9
Nalonnil A, Marion B (2010) High resolution reservoir monitoring using crosswell seismic. Paper SPE 132491 Presented at the SPE Asia pacific Oil and Gas Conference and Exhibition, Brisbane, Queensland, Austria, 18-20 October.
10 11 12
Orodu OD, Tang Z, Fei Q (2009) Hydraulic (flow) unit determination and permeability prediction: a case study of block Shen-95, Liaohe oilfield, north-east China. Journal of Applied Sciences 9:1801-1816.
13 14
Peng X, Wang Y (2009) A normal least squares support vector machine (NLS-SVM) and its learning algorithm. Neurocomputing 72:3734-3741.
15 16
Petrus JB, Thuijsman F, Weijters AJ (1995) Artificial neural networks: an introduction to ANN theory and practice. Berlin, Germany, Springer.
17 18
Salah A, Rahman S, Nath K (2005) An enhancement of k-nearest neighbor classification using genetic algorithm. MICS.
19 20 21
Schlumberger Oil Field Glossary (2018) Well logs. Available online: http://www.glossary.oilfield.slb.com/Display.cfm?Term=porosity. Accessed on September 12.
22 23
Sengel EW "Bill" (1981) Handbook on well logging. Institute for Energy Development, Oklahoma City, Oklahoma, 168 pages.
24
Sherrod PH (2008) DTREG: predictive modeling software. User Manual, pp. 197.
25 26 27
Suman A, Fernandez-Martinez JL, Mukerji T (2011) Joint inversion of time-lapse seismic and production data for Norne field. Paper Presented at the SEG San Antonio 2011 Annual Meeting.
28 29
Taner MT, Koehler F, Sheriff RE (1979) Complex seismic trace analysis. Geophysics 44:1041-1063.
30
Vapnik V (1995) The nature of statistical learning theory. Springer-Verlag, London.
31 32 33 34
Walls JD, Taner MT, Guidish T, Taylor G, Dumas D, Derzhi N (1999) North Sea reservoir characterization using rock physics, seismic attributes, and neural networks: a case history. In proceedings of the 69th Ann. International Mtg., Soc. Expl. Geophys., Expanded Abstracts, 1572-1575.
35 36
Zadeh LA (1975) The concept of a linguistic variable and its application to approximate reasoning - I. Information Sciences 8(3):199-249.
AC C
EP
TE D
M AN U
SC
RI PT
1 2
20
ACCEPTED MANUSCRIPT
AC C
EP
TE D
M AN U
SC
RI PT
1 2
21
ACCEPTED MANUSCRIPT List of Figures
2
Figure 1. Optimal Number of Hidden Neurons for ANN with (a) Run 1 and (b) Run 2.
3
Figure 2. Optimal Regularization Parameter for SVM with (a) Run 1 and (b) Run 2.
4
Figure 3. Optimal Error Goal for SVM with (a) Run 1 and (b) Run 2.
5
Figure 4. Optimal Overfitting Penalty for SVM with (a) Run 1 and (b) Run 2.
6
Figure 5. Optimal Kernel Step Size for SVM with (a) Run 1 and (b) Run 2.
7
Figure 6. Optimal Learning Rate for Type-2 Fuzzy Logic.
8
Figure 7. Optimal Number of Splits for DT with (a) Run 1 and (b) Run 2.
9
Figure 8. Optimal Number of Hidden Neurons for ELM with (a) Run 1 and (b) Run 2.
SC
RI PT
1
Figure 9. Comparative R2 of 6 AI Techniques on Data Seismic_Log#1.
11
Figure 10. Comparative Mean Abs. Error of Six AI Techniques on Data Seismic_Log#1.
12
Figure 11. Comparative R2 of Six AI Techniques on Data Seismic_Log#2.
13
Figure 12. Comparative Mean Absolute Error of Six AI Techniques on Data Seismic_Log#2.
M AN U
10
AC C
EP
TE D
14
22
AC C
EP
TE D
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
AC C
EP
TE D
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
AC C
EP
TE D
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
AC C
EP
TE D
(a)
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
(b)
ACCEPTED MANUSCRIPT
Optimal Training/Testing Parameters for SVM 0.8 0.75 0.7
0.6
RI PT
CC
0.65 Training Testing
0.55
SC
0.5 0.45 0
5000
10000
M AN U
0.4
15000
C
(a)
Optimal Training/Testing Parameters for SVM 0.85 0.8 0.75
EP
0.65
AC C
CC
0.7
0.6 0.55 0.5 0.45 0.4
Training Testing
TE D
0.9
0
5000
10000
C
(b)
15000
ACCEPTED MANUSCRIPT
Optimal Training/Testing Parameters for SVM 0.74 0.72 0.7
CC
RI PT
0.68 0.66
0.6
0
0.002
0.004
0.006
0.008
0.01
M AN U
Training Testing
0.62
SC
0.64
0.012
0.014
0.016
0.018
0.02
Lambda (a)
Optimal Training/Testing Parameters for SVM
TE D
0.75
0.74
EP
0.72
AC C
CC
0.73
0.71
0.7
0.69
Training Testing
0
0.002
0.004
0.006
0.008
0.01
0.012
Lambda
(b)
0.014
0.016
0.018
0.02
AC C
EP
TE D
(a)
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
(b)
AC C
EP
TE D
(a)
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
(b)
ACCEPTED MANUSCRIPT
Optimal Training/Testing Number for T2F 1 Training CC Testing CC
0.8
RI PT
0.6
SC
0.2 0
-0.4 -0.6
0
0.5
1
1.5
2
M AN U
-0.2
2.5
EP
TE D
Alpha
AC C
CC
0.4
3
3.5
4
4.5
5
ACCEPTED MANUSCRIPT Optimal Training/Testing Parameters for DT 1 Training Testing
0.9 0.8 0.7
RI PT
CC
0.6 0.5 0.4
SC
0.3 0.2
0
0
5
10
15
20
25
M AN U
0.1
30
35
40
45
50
K, Node Split Threshold
(a)
Optimal Training/Testing Parameters for DT 0.9 0.8
EP
0.7 0.6 0.5
AC C
CC
Training Testing
TE D
1
0.4 0.3 0.2 0.1 0
0
5
10
15
20
25
30
35
K, Node Split Threshold
(b)
40
45
50
ACCEPTED MANUSCRIPT Optimal Training/Testing Parameters for ELM 1
0.8
0.6
RI PT
0.4
CC
0.2
0
-0.2
SC
Training Testing
-0.4
-0.8
0
5
10
15
20
25
M AN U
-0.6
30
35
40
45
50
40
45
50
# of Hidden Neurons
(a)
Optimal Training/Testing Parameters for ELM
TE D
1
0.8
EP
0.6
0.2
AC C
CC
0.4
0
-0.2
Training Testing
-0.4
-0.6
0
5
10
15
20
25
30
# of Hidden Neurons
(b)
35
AC C
EP
TE D
M AN U
SC
RI PT
ACCEPTED MANUSCRIPT
ACCEPTED MANUSCRIPT Parametric study to investigate the comparative performance of ML techniques.
•
Study is applied to the estimation of petroleum reservoir permeability.
•
Seismic and log data are integrated for improved permeability prediction.
•
Outcome assists users to make informed choices on the appropriate techniques.
AC C
EP
TE D
M AN U
SC
RI PT
•