Journal Pre-proofs Research papers Improving Artificial Intelligence Models Accuracy for Monthly Streamflow Forecasting Using Grey Wolf Optimization (GWO) Algorithm Yazid Tikhamarine, Doudja Souag-Gamane, Ali Najah Ahmed, Ozgur Kisi, Ahmed El-Shafie PII: DOI: Reference:
S0022-1694(19)31170-9 https://doi.org/10.1016/j.jhydrol.2019.124435 HYDROL 124435
To appear in:
Journal of Hydrology
Received Date: Revised Date: Accepted Date:
17 October 2019 2 December 2019 3 December 2019
Please cite this article as: Tikhamarine, Y., Souag-Gamane, D., Najah Ahmed, A., Kisi, O., El-Shafie, A., Improving Artificial Intelligence Models Accuracy for Monthly Streamflow Forecasting Using Grey Wolf Optimization (GWO) Algorithm, Journal of Hydrology (2019), doi: https://doi.org/10.1016/j.jhydrol.2019.124435
This is a PDF file of an article that has undergone enhancements after acceptance, such as the addition of a cover page and metadata, and formatting for readability, but it is not yet the definitive version of record. This version will undergo additional copyediting, typesetting and review before it is published in its final form, but we are providing this version to give early visibility of the article. Please note that, during the production process, errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
© 2019 Published by Elsevier B.V.
11. Introduction 2
Broad scale of applications in water resources management and planning require
3
an imperative development to obtain an optimum model for streamflow prediction.
4
It is essential to know that the precise estimation of streamflow is considered a
5
significant stochastic feature in environmental modeling and it has attracted a big
6
deal of importance in different kinds of applications in such as activities of water
7
operation, maintenance, management, agricultural and irrigation management,
8
drought and flood alert systems, etc. as reported in(Bruins, 1990; Amiri, 2015;
9
Vogel et al., 2015). From a practical point of view, the performance of streamflow
10
model necessitates a specific period of time to function, which can either be a long
11
time period (e.g., weekly, monthly and seasonal) or short time period (e.g., hourly
12
and daily), and this matter was the main focus of the researchers in water resources
13
field over the past twenty years(Zealand et al., 1999; Wu et al., 2009a, 2009b;
14
Ismail et al., 2012; Terzi and Ergin, 2014; Zia et al., 2015). Generally, it is difficult
15
to attain a precise and dependable streamflow model because the chaotic properties
16
of streamflow (El-Shafie et al., 2007), including non-linearity, stochasticity and
17
non-stationarity,
18
complexity(Bayazit, 2015). Moreover, several factors are also influencing the
19
stream flow for instance environment climate variability (Narsimlu et al. 2015),
20
local and seasonal patterns, heterogeneity in temperature at the local and regional
21
scale and the frequent of the precipitations, temporal and spatial variability of
22
watershed, the properties of catchments as well as the activities of human
23
beings(Danandeh Mehr et al., 2013; Dehghani et al., 2014; Maity and Kashid,
24
2011; Singh and Cui, 2015).
govern
the
behavior
of
streamflow
and
cause
its
25
Despite of all, water resources scholars are still making marvelous efforts to develop
26
a reliable and precise model for streamflow forecasting(Angelakis and Gikas, 2014),
27
those scholars thrive to develop and acquire original and innovative systems to beat
28
the complexities in their models for stream flow estimating.
291.1. Background 30
The hydrologists are vitally depends on streamflow forecasting in production of
31
sustainable theoretical basis of water infrastructures, in flood measurement control in
32
monitoring the operational projects through observation of river behavior. However,
33
the conventional regression based models are incapable to estimate the streamflow
34
data sufficiently because the relationship between the output and input variables are
35
inherently non-linear. Based on that, it is substantial to improve the capability of the
36
model employed for streamflow data prediction by effectively evaluate and extract the
37
non-linear manner of the relationship between the predictor-predicted.
38
Originally, the conventional statistical approaches mainly the ones that based on
39
theoretical techniques, such as regression methods and Box-Jenkin time series
40
approaches, were applied widely to analyze and model the hydrological time series,
41
mainly the estimation of streamflow (Amisigo et al., 2008; Box and Jenkins, 1970;
42
Chua and Wong, 2011; Valipour et al., 2013). During the last two decades, there are
43
abundant of researches have been developed to investigate the potential of using data-
44
driven models such as Artificial Intelligence (AI) methods. The purpose of data-
45
driven modeling is to use the techniques of artificial intelligence (AI) for extraction of
46
documented data pattern in past to predict streamflow data in future, and it has proven
47
to be highly prevalent and favorable forecasting tool, by generating estimated
48
streamflow data that effectively represent the actual streamflow data (Chen et al.,
49
2015; Deo and Şahin, 2016; Fathian et al., 2019; Zhang et al., 2015).
50
A variety of AI-based approaches, (which involve different variables related to
51
streamflow such as evaporation, drought and temperature) have been applied in a
52
plethora of forecasting studies for wide range of reasons. For instance the models can
53
be employed for local scale processes (e.g. irrigation or farms), they show a relative
54
competitive performance with low complexity comparing to the common physically-
55
based hydrological models, additionally, the data inexpensive nature of AI based
56
models and they can be used easily designing of forecasting models and related
57
applications (Ahmed et al., 2019; Afan et al., 2017). Several of AI models have been
58
established using Artificial Neural Networks (ANN) technique (Bai et at., 2016; El-
59
Shafie et al., 2009; Bahrami et al., 2016; Valipour et al., 2012), fuzzy logic and
60
Adaptive Neuro-Fuzzy Inference System (ANFIS) (El-Shafie et al., 2007b; Kisi,
61
2015; Sharma et al., 2015; Zounemat-Kermani and Teshnehlab, 2008), genetic
62
programing (Danandeh Mehr et al., 2014; Makkeasorn et al., 2008; Turan and
63
Yurdusev, 2014), algorithms of regression and support vector machine (Guo et al.,
64
2011; Rasouli et al., 2012; W. C. Wang et al., 2009).In general, compared to the
65
traditional auto regression and other regression based methods, the performance of
66
AI-based approaches for forecasting of streamflow is proven to be more reliable and
67
effective (Yaseen et al., 2015a; Zaher Mundher Yaseen et al., 2016).
68
Recently, major efforts have been made to review and explore the weakness and
69
merits of different types of AI methods in water resources field (Nourani et al., 2014;
70
Yaseen et al., 2015b).It was concluded that there is no “absolute” AI model
71
appropriate for all kinds of modelling (such as estimation, forecasting, classification,
72
optimization, etc.) and generally there was no individual machine learning approach
73
appropriate for all definite problems. Nevertheless, the precision of AI models (with
74
no data pre-processing techniques) can be improved by using hybrid AI models
75
(consist of coupled models) with data pre-processing methods. The common used
76
method for preprocessing of data in water resources applications is wavelet
77
transformation (WT)(Okkan and Ali Serbes, 2013; Parmar and Bhardwaj, 2014;
78
Pramanik et al., 2010; Wei et al., 2013).
79
From this perspective, it is evident that the certainty of the forecasted streamflow data
80
can be effectively enhanced by the integration of two or more techniques to model
81
and assimilate the data patterns and this is defined as hybrid approach (Fahimi et al.,
82
2016). The incorporation of different optimization models implanted into a separate
83
AI-based model leads to a noticeable improvement in the efficiency of streamflow
84
forecasting model which is demonstrated by reduction of the computational time and
85
by inference of an ideal solution for the valuation problem (Ch et al., 2014; Kavousi-
86
Fard et al., 2014a, 2014b).
87
Although the accuracy of forecasting model can be improved by hybrid models,
88
researchers are still conducting many experiments to generate a sufficient approach
89
that can deduce the optimal solutions in the area of forecasting through the utilization
90
of AI models with preprocessing methods. In general, it has been proved that for
91
engineering optimization application, the achievement of global optima is a vital step
92
in the successfulness of the prediction/forecasting models. A few researches showed
93
that the meta-heuristic algorithms, such as Genetic Algorithm (GA), Particle Swarm
94
Optimization (PSO), Harmony Search (HS), Ante Colony Optimization (ACO), have
95
excellent searching aptitudes to achieve the global optima and elude the local optima
96
compared to the classical optimization algorithms. These meta-heuristic algorithms
97
such as Genetic Algorithm (GA), Particle Swarm Optimization (PSO), Harmony
98
Search (HS), Ante Colony Optimization (ACO). This is due to the fact that these
99
meta-heuristics algorithms might have the nature behave which tolerates the
100
stagnations in the local optima and rapidly switch to searching mode again to achieve
101
the global optima position. In fact, the mentioned optimization algorithms have
102
several advantages and disadvantages. The major advantage in these algorithms is
103
their ability to be adjusted to include several nonlinear systems in parallel or series
104
with different constraints and objective functions. The major disadvantage is the
105
difficulty of addressing the stochastic pattern of the system parameters, slow
106
convergence and lack of ability to distinguish the optimal global solutions. In this
107
context, this study introduces an improved meta-heuristic algorithm to address the
108
stochastic pattern of the prediction methods’ internal parameters and improve the
109
ability of the search procedure to return global optima with relatively faster
110
convergence. Furthermore, the GWO algorithm has the capability of balance between
111
exploration and exploitation efficiently and much better. This avoids a large number
112
of local solutions and provides an assessment of both accuracy and convergence
113
speed. On the other hand, the GWO algorithm with ANN model has been utilized to
114
estimate reference evapotranspiration (ET0) and compared with other four
115
optimization algorithms (PSO, ALO, WOA and MVO) (Tikhamarine et al., 2019a).
116
The results obtained demonstrate the superiority of the GWO algorithm over other
117
optimization algorithms in all cases. Toward this end, this study proposes the Grey
118
Wolf Optimization (GWO) algorithm to be integrated with the prediction model to
119
optimally search for the optimal modelling parameters. The fact that most of the
120
engineering applications of optimization, especially those applications for
121
prediction/forecasting desired variable, the entire searching space is not initially
122
recognized and has quite great numbers of local optima, and hence, the meta-heuristic
123
algorithms could help overcoming such challenges. As a consequence, this work aims
124
to integrate relatively new meta-heuristic optimization algorithm namely; Grey Wolf
125
Optimization (GWO) algorithm(Mirjalili et al., 2014)with ANN models to develop
126
robust integrated systems, as well as to examine the applicability of these hybrid
127
schemes in streamflow forecasting. The effectiveness of the proposed GWO
128
algorithm is suggested to be examined for ANN and AR methods as well.
1291.2. Objective 130
The foremost aim of this research paper is to create an effective hybrid system for
131
streamflow forecasting on monthly timescales by integrating a new meta-heuristic
132
optimization algorithm namely; Grey Wolf Optimization (GWO) algorithm with AI
133
models. In this study, the proposed GWO has been integrated with Support Vector
134
Machine (SVM), Multi-Layer Perceptron Neural Network (MLP-NN) and Auto-
135
Regression (AR) for comparative analysis purposes. Historical natural streamflow for
136
130 years on monthly basis have been collected at Aswan High Dam (AHD) on the
137
Nile River, Egypt to evaluate the performance of the proposed modelling technique.
138
Hydrological and statistical analysis for the collected dataset will be carried out to
139
investigate the practicality of the proposed integrated scheme for forecasting
140
streamflow at semi-arid zone. Moreover, quantitative performance indicators were
141
calculated to evaluate the validity of the integrated models, also the predicted and the
142
observed streamflow date were comprehensively analyzed.
1432. Case study and data description 144
Before introducing the theory and the development of the proposed model, it is
145
necessary to introduce the nature of the selected case study to evaluate the model. In
146
this section, information about the historical natural streamflow of the Nile River at
147
Aswan High Dam (AHD) will be presented. The importance of the choice of this case
148
study to examine the proposed model is due to the fact that AHD is considered as the
149
controller and the regulator for the water resources supply (almost 95%) for Arab
150
Republic of Egypt. In addition, the average amount of the annual streamflow of the
151
Nile River that crossing the AHD is relatively large and equal to 84 Billion Cubic
152
Meter (BCM). Furthermore, the Nile River is considered as one of the longest river
153
and largest catchment in the world and its streamflow is believed as highly stochastic
154
and non-linear, and hence, it is an advantageous and trustworthy illustration for
155
examining any potential method for streamflow forecasting model. Finally, it is vital
156
to accurately develop a model for streamflow forecasting in order to generate a proper
157
operational rule for AHD to have optimal water release policy that meet the water
158
demand downstream the dam.
159
Fortunately, 130 years of monthly natural streamflow at AHD is available during the
160
period between 1871 and 2000 at the Nile River Sector, Ministry of Water Resources
161
and Irrigation, Egypt. Figure 1 shows the natural monthly streamflow at AHD in
162
BCM. With careful visualization for Figure 1, it could be depicted that during these
163
130 years, there are two major seasons for hydrological regime for streamflow, wet
164
season during months between July and December while dry season during the period
165
between January and June. In addition, the fluctuation in the streamflow for the same
166
month along these 130 years showed that the streamflow is greatly stochastic and
167
extremely non-linear. For detailed analysis, it could be noticed that the maximum
168
annual streamflow was happened in 1877-78 water-year (an amount is almost equal to
169
150.33 BCM) and the annual minimum streamflow has been experienced in 1913-14
170
water year with an amount of 42.09 BCM.
171
It is essential to assure the reliabilty of the recorded data before the year 1900 because
172
during this time the natural streamflow were recorded by using manual measurment.
173
During this time, the streamflow records have been recorded by monitoring the water
174
level, and hence, using the rating curve to calculate the coresponding streamflow. In
175
order to verify the streamflow records before 1900, Hurst et al(1966)has been
176
reviewed. It has been concluded in Hurst et al (1966)that during this period, double
177
verification with another monitoring station namely, Halfa gauge station (south of
178
AHD) has been acquired and examined to assure the matching with those streamflow
179
records at AHD. In addition, it has been found out that the recent streamflow records
180
during the period between 1900 and 1960 maintain the same means and standard
181
deviation as the period before 1900, Hurst et al. (1966).
182
Generally, one of the major step that should be carried out before using such historical
183
data is to assure its reliability. In this context, prior statistical analysis of the collected
184
data have been accomplished. To substantiate the realibilty and the accuracy of the
185
collected data, two main indices namely; the mean and the standard deviation have
186
been evaluated for different 30 sets of 100 annual streamflow records for these 130
187
years data. It has been found that the mean values are ranged between 82.1 and 84,2
188
BCM while the standard deviation is ranged between 17.95 and 19.25 BCN. With this
189
pre-analysis for the collected data that showed a narrow range of the mean and the
190
standard deviation of the collected data could result in high confidenant of the
191
collected data that will be used in the current study.
192
Further analysis for the collected data have been performed at the monthly basis to
193
show that the data is highly frequent from one month to another and its wide range for
194
each month to demonstrate the difficulty modelling it. Table 1 shows the simple
195
statistical indices such as the minimum, maximum and the mean monthly streamflow
196
for each month during the period between the 1871 and 2000. As it could be noticed
197
in Table 1, the maximum and the maximum average streamflow have been
198
experienced in September with an amount of 32.79 BCM and 21.51 BCM,
199
respectively. . In addition, it could be depicted that the minimum streamflow has been
200
occurred in May with an amount equal to 0.80 BCM. In addition, the maximum and
201
the minimum standard deviation is ranged between 0.91 and 5.24 BCM for March and
202
September, respectively. Finally, in order to demonstrate how the range of the
203
streamflow is wide for each month, it is obvious that the largest range is occurred in
204
July (the maximum streamflow is almost six times the minimum streamflow), while
205
the smallest range is experienced in November (the maximum streamflow is almost
206
three times the minimum streamflow).
207
In order to recognize the wide range of the possible streamflow at the AHD for each
208
month, the streamflow data for each month has been discretized into different classes.
209
If the data for each month has been individually analyzed, it could be determined that
210
the streamflow for each month could categorized into three different classes high,
211
medium and low streamflow. In order to carry out this analysis, a straightforward
212
process has been performed to the experienced range of the streamflow for each
213
month (the difference between the maximum value of the streamflow events and the
214
minimum value during these 130 years records). The upper limit of the range of the
215
high class for any month is the maximum streamflow record while the lower limit is
216
70% of the difference between the minimum and the maximum streamflow records.
217
The lower limit of the range of the low class is the minimum streamflow record while
218
the upper limit is 30% of the difference between the minimum and the maximum
219
streamflow records. Finally, the upper limit of the range of the medium class is the
220
lower limit of the high class and lower limit is the upper limit of the lower class range.
221
It could be observed that there is a quite high variation of the ranges of the streamflow
222
class for each month, which reflects the highly stochastic of the streamflow pattern, as
223
shown in Figure 2.
224 225 2263. Methodology 2273.1. Methods 2283.1.1. Support Vector Regression (SVR) 229
Support vector regression (SVR) is a kind of regression intelligent model based on
230
support vector machine (SVM) which is developed by Smola (1996). Vapnik (1995)
231
developed the SVM technique for the first time on the basis of statistical learning
232
theory and structural risk minimization principle. The main objective of SVR is to
233
identify a function f (x) that uses all pairs (training data xi, / observed targets yi) with
234
the most minimum (ε) precision and became to be as linear as possible(Smola, 1996).
235
The SVR regression function is declaredas:
236
𝑓(𝑥) = 𝑤 × ∅(x) + 𝑏
237
Where; b is the bias term w is the weight vector in the feature space, ϕ is the transfer
238
function.
239
In order to find a suitable SVR function f (x), the problem of regression can be
240
expressed:
241
Minimize
1
2 2‖𝑤‖
(1)
𝑁
+ 𝐶∑𝑖 = 1(𝜉𝑖 + 𝜉𝑖∗ )
{
𝑦𝑖 ― f(x) ≤ ε + 𝜉𝑖 f(x) ― 𝑦𝑖 ≤ ε + 𝜉𝑖∗ 𝜉𝑖,𝜉𝑖∗ ≥ 0, 𝑖 = 1,2,3,….,𝑁
(2)
242
Subject to the condition:
(3)
243
Where C > 0 is a penalty parameter, 𝜉𝑖 and 𝜉𝑖∗ are the two slack variables to specify
244
the distance from observed values to the ε that corresponding boundary values. By
245
using the Lagrangian multipliers, the optimization issue is largely converted into a
246
quadratic programming and the solution of the nonlinear regression function can be
247
given as follows:
248
𝑓(𝑥) = ∑𝑖 = 1(𝛼𝑖 ― 𝛼𝑖∗ )𝐾(𝑥,𝑥𝑖) + 𝑏
249
Where K(x, xi) is the Kernel function, αi, αi* ≥ 0 are dual variables.
250
Steve Gunn (1998) describes the details of the use of SVM and SVR techniques in the
251
technical report: Vector Machinery Support for Classification and Regression.
252
The kernel function can be selected by different options including: linear, polynomial,
253
sigmoid and radial basis function (RBF) kernels. However, choosing the appropriate
254
kernel function is an important step in using SVR model. In this research, the RBF
255
was adopted as the kernel function because of its performance compared with other
256
kernel functions and the most popular used in literatures (Liong and Sivapragasam,
257
2002; Asefa et al., 2006; Kisi and Cimen, 2011; Tikhamarine et al., 2019b; He et al.,
258
2014).
259
The RBF kernel is defined as:
260
𝐾(𝑥,𝑥𝑖) = exp ( ― γ‖𝑥𝑖 ― 𝑥‖2)
261
Where γ is the kernel parameter, which means that C, γ, and ε are the three parameters
262
that are responsible for the SVR performance. In this study, SVR models were
263
implemented using Matlab software and LIBSVM (version 3.23) developed by Chang
264
and Lin (2011) and the default parameters were selected as follows (C = 1, γ = 0.01
265
and ε = 0.001) which are used in standard SVR.
𝑁
(4)
(5)
2663.1.2. Multiple linear regression (MLR) 267
Multiple linear regression (MLR) is a simple regression equation and one of the most
268
common methods used to solve classical regression problems in statistical analysis
269
(Tabari et al., 2011). In general, MLR is used to find an appropriate relationship
270
between the dependent variable (Y) and one or a set of independent variables (Xi).
271
MLR can generate a dependent relationship by building a linear equation calculated
272
based on the following formula:
273
𝑌 = 𝛼0 + 𝛼1𝑋1 + 𝛼2𝑋2 + 𝛼3𝑋3 +… + 𝛼𝑁𝑋𝑁
274
Where Y is the dependent variable (Qt) and Xi are the independent variables; and α0
275
to αN are the regression coefficients of MLR.
(6)
2762.3.1 Artificial neural network (ANN) 277
Artificial neural network (ANN) is a mathematical model inspired by the function of
278
the biological nervous system of the human being. The ANN has been presented for
279
the first time by McCulloch and Pitts (1943) and can be considered as a mathematical
280
model to solve the complex relationships between variables (Haykin, 1994).
281
The most commonly used ANN model is the multilayer perceptron neural network
282
(MLP) consisting of input and output layers and only one hidden layer in the middle
283
connected to each other with weights and biases. Standard ANN can be reported as an
284
MLP with Levenberg-Marquardt (LM) training algorithm.
285
The explicit mathematical expression to calculate the predicted streamflow can be
286
expressed using the artificial neural network as follows:
287
𝑌 = 𝐹2 ∑𝑖 = 1𝑊𝑘𝑗 × 𝐹1(𝐴𝑗) + 𝑏𝑜
288
𝐴𝑗 = ∑𝑖 = 1𝑋𝑖𝑊𝑗𝑖 + 𝑏𝑗
[
𝑛
𝑚
]
(7) (8)
289 290
Where;𝑌 is the output variable which calculated by the ANN model (the streamflow
291
predicted reported as Y), Xi is the input variable, 𝐴𝑗is the summation of the inputs and
292
their weights represented by equation (8).F1 is the activation function for the hidden
293
layer represented by equation (9), F2 is the activation function for the output layer Wij
294
is the weight between the input i and the hidden node j. bj is the bias of the hidden
295
neuron j, Wjk is the weight of connection of neuron j in the hidden layer to neuron k in
296
the output layer and bo is the bias of the output node k.
297
𝐹1(𝐴𝑗) = 1 + exp ( ― 𝐴𝑗)
298
Because there is no fixed way for selecting the appropriate number of hidden nodes in
299
ANN and in order to avoiding drawback in the large numbers of trial and error
300
process from other side the number of hidden nodes was calculated based on the
301
equation (10) used in literatures (Mirjalili, 2015; Faris et al., 2016; Aljarah et al.,
302
2018, 2019).
303
𝑚=2∗𝑛+1
304
Where; m is the number of neurons; n is the number of inputs.
1
(9)
(10)
305 3.1.4. Grey wolf optimizer (GWO) algorithm 306
The GWO algorithm is a new intelligent algorithm mimics the hierarchy and social
307
hunting of grey wolves proposed by Mirjalili et al.(2014). Generally, the pack of
308
wolves are divided into four groups; Alpha (α), Beta (β), Delta (δ) and the rest of
309
wolves are the Omega (ω). The most dominated wolf is Alpha and can be considered
310
as a leader of the pack. The domination level decrease from alpha to omega is shown
311
in Figure 4 (a).The GWO mechanism is carried out by splitting a set of solutions to
312
the given optimization problem into four groups. The first three solutions are the best
313
α, β and δ. The remaining solutions belong to ω wolves. To implement this
314
mechanism, the hierarchy in each iteration is updated is updated based on the three
315
best solutions. The illustration of the update location is shown in Figure 4 (b).
316
The main principal in GWO algorithm is searching, encircling, hunting, and attacking
317
the prey.
318
Before hunting process, the grey wolves are encircling the prey. The following
319
equations represent encircling behaviour of grey wolves:
320
𝑋(𝑡 + 1) = 𝑋𝑃(𝑡) ― 𝐴 ⋅ 𝐷
321
Where;𝑋(𝑡 + 1) is the next location of any wolf,𝑋𝑃(𝑡)is position vector of the grey
322
wolf, t is the current iteration, 𝐴 is matrix coefficient and 𝐷 is the distance separating
323
the grey wolf and the prey which can be estimated as follows:
324
𝐷 = |𝐶 ⋅ 𝑋𝑃(𝑡) ― 𝑋(𝑡)|
(12)
325
𝐴 = 2𝑎 ⋅ 𝑟1 ― 𝑎
(13)
326
𝐶 = 2 ⋅ 𝑟2
(14)
327
Where; 𝑟1and 𝑟2 are randomly generated from (0 to 1).
328
The previous equations permit a solution to relocate around the prey in a hyper-sphere
329
form (figure 4 (b)). This is not sufficient, nevertheless, to simulate the social
330
intelligence of grey wolves. In order to simulate the prey, the best solution obtained so
331
far considered as the alpha wolf is closer to the prey position, but the global optimal
332
solution is unknown, so it is assumed that the top three solutions have a good idea of
333
their location, therefore other wolves should be obliged to update their locations by
334
using the following equations:
335
𝑋(𝑡 +1) = 3𝑋1 + 3𝑋2 + 3𝑋3
336
Where;𝑋1,𝑋2 and 𝑋3 are calculated using the following equations:
337
𝑋1 = 𝑋𝛼(𝑡) ― 𝐴1 ∗ 𝐷𝛼
(16)
338
𝑋2 = 𝑋𝛽(𝑡) ― 𝐴2 ∗ 𝐷𝛽
(17)
339
𝑋2 = 𝑋𝛽(𝑡) ― 𝐴2 ∗ 𝐷𝛽
(18)
340
Where; 𝐷𝛼, 𝐷𝛽 and𝐷𝛿are given by:
341
𝐷𝛼 = |𝐶1 ⋅ 𝑋𝛼(𝑡) ― 𝑋|
1
1
1
(11)
(15)
(19)
342
𝐷𝛽 = |𝐶2 ⋅ 𝑋𝛽(𝑡) ― 𝑋|
(20)
343
𝐷𝛿 = |𝐶3 ⋅ 𝑋𝛿(𝑡) ― 𝑋|
(21)
344
The prey encircling and attacking are repeated until an optimum solution is obtained
345
or it reaches the maximum number of iterations.
3463.2. Inputs selection and model development 3473.2.1. Inputs selection 348
Choosing proper input variables is very important for the development of SVR, ANN
349
and MLR models since it gives the essential information about the designed system.
350
In the present study, eight different input combinations including the previous values
351
of streamflow have been selected based on simple autocorrelation function (ACF) and
352
partial autocorrelation function (PCF). ACF and PCF are frequently used to determine
353
the appropriate inputs in the time series prediction field. The ACF and PCF were
354
utilized to identify delays that clarify the variance in the predicted streamflow. Figure
355
3 represents the ACF and PACF curves for the Nile River at Aswan High Dam. The
356
lag that shows the great correlation have been chosen to be an input to the selected
357
model. The optimal input sets derived examining ACF and PACF in terms of model
358
number, inputs and output for each model are reported in Table 3.
3593.2.2. Model development 360
The optimization algorithm used in this study is the grey wolf optimize(Mirjalili et al.,
361
2014), because it is one of the modern swarm intelligence algorithm and used
362
successfully in the engineering field (Yu and Lu, 2018; Al Shorman et al., 2019;
363
Tikhamarine et al., 2019a; Maroufpoor et al., 2019).
364
Support vector regression and artificial neural network are belongs to the models of
365
artificial intelligence and shows high performance accuracy in modeling the nonlinear
366
relationship between predictors and predictors (Asefa et al., 2006; Lin et al., 2006;
367
W.-C. Wang et al., 2009). However, the performance of the SVR depends on its
368
parameters and the choice of kernel function. As with the artificial neural network, the
369
performance of the ANN is also depends on the correct selection of weights and
370
biases. The MLR is a linear regression models used to find an appreciated relationship
371
between variables. Nevertheless, the accuracy of MLR also depends on the correct
372
choice of regression coefficients (αi). Consequently, the correct selection of
373
parameters can be considered an optimization problem and need a high optimization
374
algorithm to resolve this problem. Therefore, SVR was coupledwith GWO to build
375
the SVR-GWO model, ANN was coupled with GWO to build the ANN-GWO and the
376
MLR embedded with GWO to construct the MLR-GWO model to predict the
377
streamflow for the Nile River at Aswan High Dam. The SVR-GWO, ANN-GWO and
378
MLR-GWO models were trained and tested for each of eight combination. The
379
flowcharts of the proposed hybrid models are illustrated in Fig. 4: (a) for the hybrid
380
ANN-GWO, (b) for MLR-GWO and (c) for SVR-GWO.
381
For the MLR-GWO, the search agent number is the number of regression coefficients
382
(equation 6) while in ANN-GWO the search agent number can be obtained using the
383
following equation:
384
Number of wights and bieses = (n × m) + (1 × m) + (m × 1) + 1 (22)
385
Where; the search agent number is the number of weights and biases for ANN model,
386
n is the number of inputs and m is the number of hidden neurons in the hidden layer.
387
Besides, as prescribed in literatures, we utilized the most parameter settings for GWO
388
algorithm. Table 4 gives initial parameters of GWO algorithm in detail for the SVR,
389
ANN and MLR models.
3903.3. Performance indicators
391
In this study, the performance criteria used to assess the model’s performance are:
392
Root mean squared error (RMSE):
393
RMSE =
394
Mean absolute error (MAE):
395
MAE = 𝑁∑𝑖 = 1|QO,𝑖 ― QP, 𝑖|
396
Coefficient of correlation (R):
1 𝑁
1
𝑁
∑𝑖 = 1(QP, 𝑖 ― QO,𝑖)2
𝑁
(0 < RMSE < ∞)
(23)
(0 < MAE < ∞)
(24)
(–1 < R < 1)
(25)
𝑁
∑𝑖 = 1(QO,𝑖 ― QO )(QP,𝑖 ― QP)
397
R=
398
Nash-Sutcliffe efficiency coefficient (NSE):
2 𝑁
𝑁
∑𝑖 = 1(QO,𝑖 ― QO ) ∑𝑖 = 1(QP,𝑖 ― QP)
[
𝑁
2
]
∑𝑖 = 1(QO, 𝑖 ― QP, 𝑖)2
(–∞ < NSE < 1)
399
NSE = 1 ―
400
Willmott Index (WI) (Willmott, 1981 and 1984):
[
𝑁
∑𝑖 = 1(QO, 𝑖 ― QO )
𝑛
2
∑𝑖 = 1(QO,𝑖 ― QP,𝑖)2
]
(0< WI < 1)
(26)
(27)
401
WI = 1 ―
402
Where; QO, 𝑖is the value of the streamflow observed in the current observed (i),QP, 𝑖is
403
the predicted value, QO is the average value of observations andQP is the average value
404
of the predictions and N is the number of the data.
𝑛
∑𝑖 = 1(|QO,𝑖 ― QO | + |QP,𝑖 ― QP|)2
4054. Results and discussions 406
In this research, the abilities of hybrid ANN-GWO, SVR-GWO and MLR-GWO data
407
driven methods are investigated in monthly inflow forecasting and their results are
408
compared with those of standard ANN and SVR approaches. Before applying the
409
mentioned methods, whole data are split in two data sets, 70% for training and 30%
410
for testing. Table 2 sums up the brief statistical parameters of the utilized data sets. It
411
is clear that both periods have generally similar characteristics and test data have a
412
little more skewed distribution compared to that of training. Previous inflow values
413
were utilized as inputs to the implemented methods and optimal inputs were decided
414
by applying correlation analysis. Autocorrelation function (ACF) and partial
415
autocorrelation function (PACF) are demonstrated in Figure 2 and the optimal input
416
sets derived examining ACF and PACF are reported in Table 3. Totally 8 input
417
combinations were selected considering the correlation magnitude of inflow data and
418
as can be observed, all previous values up to 13 lags were also utilized as input to
419
better see the usefulness of such analysis in modeling stage of data driven methods.
420
Table 4givesinitial parameters of GWO algorithm in detail for the SVR, ANN and
421
MLR methods. Some control parameters were decided considering the past literature.
422
The performance of the SCR-GWO is summarized in Table 5 for both training and
423
test stages. The optimized SVR parameters utilizing meta heuristic GWO algorithm
424
are also provided in the table for each input case. Table 5 clearly shows that the
425
simulation (training stage) and forecasting (testing stage) accuracy of the SVR-GWO
426
method varies with respect to input combination and the lowest training and testing
427
RMSE (1.5578 m3/s and 2.0570 m3/s), MAE (1.0762 m3/s and 1.2005 m3/s) and the
428
highest training and testing R (0.9682 and 0.9363), NSE (0.9566 and 0.8728) and WI
429
(0.9886 and 0.9671) belong to the third model with input of Qt-1, Qt-2, Qt-11, Qt-12, Qt-
430
13.Test
431
RMSE from 2.7808 m3/s to 2.0570 m3/s and increase NSE from 0.7676 to 0.8728) and
432
then decrease till 5th input combination (increase in RMSE from 2.1937 m3/s to
433
2.2880m3/s and decrease in NSE from 0.8554 to 0.8427). As evident from the results,
434
including high number inputs do not guarantee better forecasting accuracy (e.g.,
435
Zhang et al. 2019; Adnan et al. 2019; Shi et al. 2012) because increasing inputs
436
number may increase variance and add more complexity to the implemented model
accuracy of the SVR-GWO method increases up to 3rd input case (decrease in
437
and this deteriorate the model accuracy in forecasting. The other important issue
438
which is worthy to mention here that the input combination recommended by the
439
correlation analysis should not be directly applied to the data driven methods and it is
440
better to try several input cases by considering ACF and PACF.
441
Table 6sums up the performance of hybrid ANN-GWO method in training and testing
442
stages of different input cases. The optimal model structures are also provided in the
443
table. In the structure column, first number indicates input number, second number is
444
hidden node number and last number indicates the output Qt-1 (inflow value of current
445
month). For example, the ANN with 2-5-1structure has 2 inputs (Qt-1 and Qt-2), 5
446
hidden nodes and one output. Similar to the SVR-GWO, among the ANN-based
447
models, the third one having 5-11-1 structure with the input of Qt-1, Qt-2, Qt-11, Qt-12, Qt-
448
13has
449
(1.0168m3/s and 1.2999m3/s) and the highest training and testing R (0.9717 and
450
0.9314), NSE (0.9441 and 0.8561) and WI (0.9854 and 0.9636). An increase in
451
accuracy (decrease in RMSE from 3.1324 m3/s to 2.1883 m3/s and increase NSE from
452
0.7051 to 0.8561) from 1st input to 3rd input cases and then an increase in accuracy
453
(increase in RMSE from 2.3782m3/s to 2.3850m3/s and decrease in NSE from 0.8300
454
to 0.8290) is seen from 4th input to 5th input cases for the testing stage of ANN-GWO
455
method. For this method, also less number of inputs provides better accuracy
456
(compare the M3 with M2, M5, M6, M7 and M8).
457
Tables 7-8 give the accuracy of standard SVR and ANN methods in forecasting
458
monthly inflows. Alike to hybrid SVR-GWO and SVR-GWO, among the SVR-based
459
models, the third SVR-based model has the lowest performance in both training
460
(RMSE: 1.8253 m3/s, MAE: 0.9780 m3/s, R: 0.9724,NSE: 0.9404, WI: 0.9834) and
461
testing (RMSE: 2.2248 m3/s, MAE: 1.3190 m3/s, R: 0.9244, NSE: 0.8512, WI:
the lowest training and testing RMSE (1.7663m3/s and 2.1883m3/s), MAE
462
0.9605)stages (Table 7). Similarly, among the ANN-based models, the third ANN-
463
based model has the lowest performance in training (RMSE: 1.3110m3/s, MAE:
464
0.8006m3/s, R: 0.9845, NSE: 0.9692, WI: 0.9921) and testing (RMSE: 2.4214m3/s,
465
MAE: 1.3105m3/s, R: 0.9118, NSE: 0.8238, WI: 0.9539) stages (Table 8).In both
466
SVR and ANN methods, accuracy of the models increase from 1st input to 3rd input
467
and decrease from 4th to 5th input similar to hybrid versions. Also, for these methods,
468
better accuracy can be obtained utilizing the models with a smaller number of inputs.
469
The accuracy of hybrid MLR-GWO method is provided in Table 9 for training and
470
test periods. Parallel results were also obtained from this method so that the third
471
model had the best accuracy (RMSE: 2.0366 m3/s and2.3883 m3/s, MAE: 1.1255m3/s
472
and 1.3370 m3/s, R: 0.9623 and 0.9121, NSE: 0.9257 and 0.8286, WI: 0.9806 and
473
0.9538 for training and testing, respectively) in both periods and accuracy variation
474
with respect to input cases is similar to previous methods.
475
Comparison of hybrid methods (Table 5, 6 and 9) reveals that the both hybrid SVR-
476
GWO and ANN-GWO methods have better approximation accuracy in training stage
477
and performs superior to the MLR-GWOin forecasting (testing) stage in all input
478
cases. The best model is SVR-GWO-based model with 3rd input combination and
479
followed by the ANN-GWO-based model with the same input. This indicates the
480
nonlinearity of the investigated phenomenon and therefore, linear MLR-GWO-based
481
models cannot adequately map monthly inflows.The other advantages of hybrid SVR
482
and ANN methods are involving much more model parameters compared to hybrid
483
MLR approach. This can be clearly seen from the Tables 10-11 which present the
484
optimized parameters of ANN and MLR methods.High number of parameters adds
485
flexibility to the data driven methods but if they are adequately calibrated. The
486
relative RMSE, MAE, R, NSE and WI differences between the optimal hybrid SVR-
487
GWO (model M3) and ANN-GWO (model M3) in the testing stage are 6%, 7.6%,
488
0.5%, 2%, 0.4%, respectively.
489
Comparison of hybrid SVR-GWO and standard SVR methods (Table 5 and 7)
490
indicates that the hybrid SVR-GWO method can make better approximation in
491
training stage and produce better forecasts than the single SVR in forecasting monthly
492
inflow in all input cases. This indicates the superiority of GWO meta-heuristic
493
algorithm compared to training algorithm of standard SVR. The relative RMSE,
494
MAE, R, NSE and WI differences between the optimal hybrid SVR-GWO (model
495
M3) and SVR (model M3) in the testing stage are 7.5%, 9%, 1.3%, 2.5%, 0.7%,
496
respectively. Similarly, comparison of hybrid ANN-GWO and standard ANN
497
methods (Table 6 and 8) indicates that the GWO algorithm makes better
498
approximation in training stage and provides better forecasts than the single ANN in
499
forecasting inflow in all input cases. This reveals the superiority of GWO meta-
500
heuristic algorithm compared to the training algorithm (Levenberg-Marquardt) of
501
standard ANN. The relative RMSE, MAE, R, NSE and WI differences between the
502
optimal hybrid ANN-GWO (M3) and ANN (M3) in the testing stage are 10.7%,
503
0.8%, 2.1%, 3.8% and 1%, respectively. From the results obtained from Table 5-8, we
504
can say that the SVR method performs superior to the ANN method in forecasting
505
monthly inflows.Kernel function add flexibility to the SVR and thus it can implicitly
506
simulate the input data to a higher dimensional (HD) space. Linear solutions of the
507
problems in the feature space of HD correspond to non-linear solutions in the original
508
lower dimensional input space. These characteristicsmake SVR a feasible alternative
509
for the solution of several problems in water resources and hydrology, which are
510
naturally non-linear (Naganna and Deka, 2014).
511
Figure 6 illustrates time variation graph of observed and forecasted monthly inflows
512
by the optimal models. It also includes detail graphs of two sections. It is clearly seen
513
from detail graphs that the forecasts of the SVR-GWO-3 model are closer to the
514
corresponding observed inflow values. The forecasts of the optimal models are
515
compared in Figure 7 in the scatter plot form. It is apparent from the graphs that the
516
SVR-GWO-3 (model M3) has less scattered forecasts compared to ANN-GWO-3,
517
MLR-GWO-3, ANN-3 and SVR-3 models. The time variation of relative errors is
518
shown in Figure 8 for each method. As observed from the figure, the range of relative
519
errors is less for the SVR-GWO-3 model in comparison to the other models.
5205. Conclusion 521
A new integrated model, which combines the Grey Wolf Optimization (GWO)
522
algorithm and Support Vector Regression, was proposed to predict monthly
523
streamflow in Aswan High Dam. GWO was developed for optimizing the hyper-
524
parameters of SVR such as the global solution parameters and then it compared with
525
two more integrated algorithms such as Artificial Neural Network (ANN) and
526
Multiple Linear Regression (MLR). The results reveal, the integrated AI models with
527
GWO were proven to be more accurate and effective compared with the standard AI.
528
Moreover, the integrated SVR-GWO outperformed, in terms of performance criteria
529
convergence, ANN-GWO and MLR-GWO in forecasting the monthly streamflow.
530
This reveals the superiority of GWO meta-heuristic algorithm in optimizing the
531
parameter of the standard SVR to improve its accuracy in forecasting the streamflow.
532
In future researches, the improved revised algorithm of Grey Wolf Optimization can
533
be introduced to enhance the hyper-parameters optimization in order to avoid trapping
534
in local optima. In addition, in a few cases studies, there is a need to develop real-time
535
forecasting model, and hence, the time consuming for the model execution should be
536
lessen. Therefore, there is a need to improve the searching algorithm for the GWO to
537
achieve fast convergence procedure as well. Moreover, other advanced meta-heuristic
538
algorithms could be investigated to improve streamflow forecasting.
539 540
Reference
541
Afan, H.A., Keshtegar, B., Mohtar, W.H.M.W., El-Shafie, A., 2017. Harmonize input
542
selection
543
https://doi.org/10.1016/j. jhydrol.2017.07.008
544 545
for
sediment
transport
prediction.
J.
Hydrol.
552.
Ahmed, A.N., Othman, F.B., Afan, H.A., Elsha, A., 2019. Machine learning methods for better water quality prediction. J. Hydrol. 578
546
Al Shorman, A., Faris, H., Aljarah, I., 2019. Unsupervised intelligent system based on
547
one class support vector machine and Grey Wolf optimization for IoT botnet
548
detection. J Ambient Intell Human Comput. https://doi.org/10.1007/s12652-
549
019-01387-y
550
Aljarah, I., Faris, H., Mirjalili, S., 2018. Optimizing connection weights in neural
551
networks using the whale optimization algorithm. Soft Computing 22, 1–15.
552
https://doi.org/10.1007/s00500-016-2442-1
553
Aljarah, I., Faris, H., Mirjalili, S., Al-Madi, N., Sheta, A., Mafarja, M., 2019.
554
Evolving neural networks using bird swarm algorithm for data classification
555
and
556
https://doi.org/10.1007/s10586-019-02913-5
557
regression
applications.
Cluster
Computing.
Amiri, E., 2015. Forecasting daily river flows using nonlinear time series models.
558
Journal
of
Hydrology
559
https://doi.org/10.1016/j.jhydrol.2015.05.048
527,
1054–1072.
560
Amisigo, B.A., van de Giesen, N., Rogers, C., Andah, W.E.I., Friesen, J., 2008.
561
Monthly streamflow prediction in the Volta Basin of West Africa: A SISO
562
NARMAX polynomial modelling. Physics and Chemistry of the Earth, Parts
563
A/B/C 33, 141–150. https://doi.org/10.1016/j.pce.2007.04.019
564
Angelakis, A.N., Gikas, P., 2014. Water reuse: Overview of current practices and
565
trends in the world with emphasis on EU states. Water Utility Journal 8, 67–
566
78.
567
Asefa, T., Kemblowski, M., McKee, M., Khalil, A., 2006. Multi-time scale stream
568
flow predictions: The support vector machines approach. Journal of
569
Hydrology 318, 7–16. https://doi.org/10.1016/j.jhydrol.2005.06.001
570
Bahrami, S., Ardejani, F.D., Baafi, E., 2016. Application of artificial neural network
571
coupled with genetic algorithm and simulated annealing to solve groundwater
572
inflow problem to an advancing open pit mine. J. Hydrol. 536, 471–484.
573
Bai, Y., Chen, Z., Xie, J., Li, C., 2016. Daily reservoir inflow forecasting
574
usingmultiscale deep feature learning with hybrid models. J. Hydrol. 532,
575
193–206.
576
Bayazit, M., 2015. Nonstationarity of Hydrological Records and Recent Trends in
577
Trend Analysis: A State-of-the-art Review. Environ. Process. 2, 527–542.
578
https://doi.org/10.1007/s40710-015-0081-7
579 580
Box, G.E.P., Jenkins, G.M., 1970. Time Series Analysis, Forecasting and Control, 1st editio. ed. Holden-Day, San Francisco, CA.
581
Bruins, H.J., 1990. Water harvesting for plant production: Reij, C., Mulder, P. and
582
Begemann, L. Washington, D.C.: The World Bank, World Bank Technical
583
Paper
584
https://doi.org/10.1016/0143-6228(90)90043-O
Number
91,
1988,
123
pp.
Applied
Geography
10,
359.
585
Ch, S., Sohani, S.K., Kumar, D., Malik, A., Chahar, B.R., Nema, A.K., Panigrahi,
586
B.K., Dhiman, R.C., 2014. A Support Vector Machine-Firefly Algorithm
587
based forecasting model to determine malaria transmission. Neurocomputing
588
129, 279–288. https://doi.org/10.1016/j.neucom.2013.09.030
589
Chang, C.-C., Lin, C.-J., 2011. LIBSVM: A Library for Support Vector Machines.
590
ACM
Trans.
Intell.
Syst.
591
https://doi.org/10.1145/1961189.1961199
Technol.
2,
27:1–27:27.
592
Chen, X., Chau, K., Wang, W., 2015. A novel hybrid neural network based on
593
continuity equation and fuzzy pattern-recognition for downstream daily river
594
discharge
595
https://doi.org/10.2166/hydro.2015.095
forecasting.
Journal
of
Hydroinformatics
17,
733–744.
596
Chua, L.H.C., Wong, T.S.W., 2011. Runoff forecasting for an asphalt plane by
597
Artificial Neural Networks and comparisons with kinematic wave and
598
autoregressive moving average models. Journal of Hydrology 397, 191–201.
599
https://doi.org/10.1016/j.jhydrol.2010.11.030
600
Danandeh Mehr, A., Kahya, E., Bagheri, F., Deliktas, E., 2013. Successive-station
601
monthly streamflow prediction using neuro-wavelet technique. Earth Science
602
Informatics 1–13. https://doi.org/10.1007/s12145-013-0141-3
603
Danandeh Mehr, A., Kahya, E., Yerdelen, C., 2014. Linear genetic programming
604
application for successive-station monthly streamflow prediction. Computers
605
and Geosciences 70, 63–72. https://doi.org/10.1016/j.cageo.2014.04.015
606
Dehghani, M., Saghafian, B., Nasiri Saleh, F., Farokhnia, A., Noori, R., 2014.
607
Uncertainty analysis of streamflow drought forecast using artificial neural
608
networks and Monte-Carlo simulation. International Journal of Climatology
609
34, 1169–1180. https://doi.org/10.1002/joc.3754
610
Deo, R.C., Şahin, M., 2016. An extreme learning machine model for the simulation of
611
monthly mean streamflow water level in eastern Queensland. Environmental
612
monitoring and assessment 188. https://doi.org/10.1007/s10661-016-5094-9
613
El-Shafie, A., Abdin, A.E., Noureldin, A., Taha, M.R., 2009. Enhancing inflow
614
forecasting model at aswan high dam utilizing radial basis neural network and
615
upstream monitoring stations measurements. Water Resources Management
616
23, 2289–2315. https://doi.org/10.1007/s11269-008-9382-1
617
El-Shafie, A., Taha, M.R., Noureldin, A., 2007a. A neuro-fuzzy model for inflow
618
forecasting of the Nile river at Aswan high dam. Water Resour Manage 21,
619
533–556. https://doi.org/10.1007/s11269-006-9027-1
620
El-Shafie, A., Taha, M.R., Noureldin, A., 2007b. A neuro-fuzzy model for inflow
621
forecasting of the Nile river at Aswan high dam. Water resources management
622
21, 533–556.
623
Fahimi, F., Yaseen, Z.M., El-shafie, A., 2016. Application of soft computing based
624
hybrid models in hydrological variables modeling: a comprehensive review.
625
Theoretical and Applied Climatology 1–29. https://doi.org/10.1007/s00704-
626
016-1735-8
627
Fathian, F., Mehdizadeh, S., Kozekalani Sales, A., Safari, M.J.S., 2019. Hybrid
628
models to improve the monthly river flow prediction: integrating artificial
629
intelligence and nonlinear time series models. J. Hydrol. 575, 1200–1213
630
Faris, H., Aljarah, I., Mirjalili, S., 2016. Training feedforward neural networks using
631
multi-verse optimizer for binary classification problems. Applied Intelligence
632
45, 322–332. https://doi.org/10.1007/s10489-016-0767-1
633
Guo, J., Zhou, J., Qin, H., Zou, Q., Li, Q., 2011. Monthly streamflow forecasting
634
based on improved support vector machine model. Expert Systems with
635
Applications 38, 13073–13081. https://doi.org/10.1016/j.eswa.2011.04.114
636 637
Haykin, 1994. Neural Networks: A Comprehensive Foundation. Prentice Hall PTR Upper Saddle River NJ, USA.
638
He, Z., Wen, X., Liu, H., Du, J., 2014. A comparative study of artificial neural
639
network, adaptive neuro fuzzy inference system and support vector machine
640
for forecasting river flow in the semiarid mountain region. Journal of
641
Hydrology 509, 379–386. https://doi.org/10.1016/j.jhydrol.2013.11.054
642
Hurst, H.E., Black, R.P., Simaika, Y.M., 1966. The Major Nile Projects. The Nile
643
Basin, vol. X. Government Press, Cairo.
644
Ismail, S., Shabri, A., Samsudin, R., 2012. A hybrid model of self organizing maps
645
and least square support vector machine for river flow forecasting. Hydrology
646
and Earth System Sciences 16, 4417–4433. https://doi.org/10.5194/hess-16-
647
4417-2012
648
Kavousi-Fard, A., Samet, H., Marzbani, F., 2014a. A new hybrid Modified Firefly
649
Algorithm and Support Vector Regression model for accurate Short Term
650
Load Forecasting. Expert Systems with Applications 41, 6047–6056.
651
https://doi.org/10.1016/j.eswa.2014.03.053
652
Kavousi-Fard, A., Samet, H., Marzbani, F., 2014b. A new hybrid Modified Firefly
653
Algorithm and Support Vector Regression model for accurate Short Term
654
Load Forecasting. Expert Systems with Applications 41, 6047–6056.
655
https://doi.org/10.1016/j.eswa.2014.03.053
656
Kisi, O., 2015. Streamflow Forecasting and Estimation Using Least Square Support
657
Vector Regression and Adaptive Neuro-Fuzzy Embedded Fuzzy c-means
658
Clustering 5109–5127. https://doi.org/10.1007/s11269-015-1107-7
659
Kisi, O., Cimen, M., 2011. A wavelet-support vector machine conjunction model for
660
monthly streamflow forecasting. Journal of Hydrology 399, 132–140.
661
https://doi.org/10.1016/j.jhydrol.2010.12.041
662
Lin, J.-Y., Cheng, C.-T., Chau, K.-W., 2006. Using support vector machines for long-
663
term discharge prediction. Hydrological Sciences Journal 51, 599–612.
664
https://doi.org/10.1623/hysj.51.4.599
665
Liong, S.-Y., Sivapragasam, C., 2002. FLOOD STAGE FORECASTING WITH
666
SUPPORT VECTOR MACHINES. Journal of the American Water Resources
667
Association 38, 173–186. https://doi.org/10.1111/j.1752-1688.2002.tb01544.x
668
Maity, R., Kashid, S.S., 2011. Importance analysis of local and global climate inputs
669
for basin-scale streamflow prediction. Water Resources Research 47, 1–17.
670
https://doi.org/10.1029/2010WR009742
671
Makkeasorn, a., Chang, N.B., Zhou, X., 2008. Short-term streamflow forecasting
672
with global climate change implications - A comparative study between
673
genetic programming and neural network models. Journal of Hydrology 352,
674
336–354. https://doi.org/10.1016/j.jhydrol.2008.01.023
675
Maroufpoor, S., Maroufpoor, E., Bozorg-Haddad, O., Shiri, J., Mundher Yaseen, Z.,
676
2019. Soil moisture simulation using hybrid artificial intelligent model:
677
Hybridization of adaptive neuro fuzzy inference system with grey wolf
678
optimizer
679
https://doi.org/10.1016/j.jhydrol.2019.05.045
algorithm.
Journal
of
Hydrology
575,
544–556.
680
McCulloch, W.S., Pitts, W., 1943. A logical calculus of the ideas immanent in
681
nervous activity. Bulletin of Mathematical Biophysics 5, 115–133.
682
https://doi.org/10.1007/BF02478259
683
Mirjalili, S., 2015. How effective is the Grey Wolf optimizer in training multi-layer
684
perceptrons.
685
https://doi.org/10.1007/s10489-014-0645-7
686
Applied
Intelligence
43,
150–161.
Mirjalili, S., Mirjalili, S.M., Lewis, A., 2014. Grey Wolf Optimizer. Advances in
687
Engineering
Software
688
https://doi.org/10.1016/j.advengsoft.2013.12.007
69,
46–61.
689
Nourani, V., Hosseini Baghanam, A., Adamowski, J., Kisi, O., 2014. Applications of
690
hybrid wavelet–Artificial Intelligence models in hydrology: A review. Journal
691
of Hydrology 514, 358–377. https://doi.org/10.1016/j.jhydrol.2014.03.057
692
Okkan, U., Ali Serbes, Z., 2013. The combined use of wavelet transform and black
693
box models in reservoir inflow modeling. Journal of Hydrology and
694
Hydromechanics 61, 112–119. https://doi.org/10.2478/johh-2013-0015
695
Parmar, K.S., Bhardwaj, R., 2014. River Water Prediction Modeling Using Neural
696
Networks, Fuzzy and Wavelet Coupled Model. Water Resources Management
697
29, 17–33. https://doi.org/10.1007/s11269-014-0824-7
698
Pramanik, N., Panda, R.K., Singh, A., 2010. Daily river flow forecasting using
699
wavelet ANN hybrid models. Journal of Hydroinformatics 13, 49.
700
https://doi.org/10.2166/hydro.2010.040
701
Rasouli, K., Hsieh, W.W., Cannon, A.J., 2012. Daily streamflow forecasting by
702
machine learning methods with weather and climate inputs. Journal of
703
Hydrology 414–415, 284–293. https://doi.org/10.1016/j.jhydrol.2011.10.039
704
Sharma, S., Srivastava, P., Fang, X., Kalin, L., 2015. Performance comparison of
705
Adoptive Neuro Fuzzy Inference System (ANFIS) with Loading Simulation
706
Program C++ (LSPC) model for streamflow simulation in El Niño Southern
707
Oscillation (ENSO)-affected watershed. Expert Systems with Applications 42,
708
2213–2223. https://doi.org/10.1016/j.eswa.2014.09.062
709
Singh, V.P., Cui, H., 2015. Entropy Theory for Streamflow Forecasting.
710
Environmental Processes 2, 449–460. https://doi.org/10.1007/s40710-015-
711
0080-8
712
Smola, 1996. Regression Estimation with Support Vector Learning Machines.
713
Steve Gunn, 1998. Support Vector Machiens for Classification and Regression, Image
714
Speech & Intelligent Systems Research Group, University of Southapton.
715
Tabari, H., Sabziparvar, A.-A., Ahmadi, M., 2011. Comparison of artificial neural
716
network and multivariate linear regression methods for estimation of daily soil
717
temperature in an arid region. Meteorol Atmos Phys 110, 135–142.
718
https://doi.org/10.1007/s00703-010-0110-z
719
Terzi, Ö., Ergin, G., 2014. Forecasting of Monthly River Flow with Autoregressive
720
Modeling and Data-driven Techniques. Neural Comput. Appl. 25, 179–188.
721
https://doi.org/10.1007/s00521-013-1469-9
722
Tikhamarine, Y., Malik, A., Kumar, A., Souag-Gamane, D., Kisi, O., 2019a.
723
Estimation of monthly reference evapotranspiration using novel hybrid
724
machine
725
02626667.2019.1678750. https://doi.org/10.1080/02626667.2019.1678750
learning
approaches.
Hydrological
Sciences
Journal
726
Tikhamarine, Y., Souag-Gamane, D., Kisi, O., 2019b. A new intelligent method for
727
monthly streamflow prediction: hybrid wavelet support vector regression
728
based on grey wolf optimizer (WSVR–GWO). Arab J Geosci 12, 540.
729
https://doi.org/10.1007/s12517-019-4697-1
730
Turan, M.E., Yurdusev, M.A., 2014. Predicting Monthly River Flows by Genetic
731
Fuzzy
Systems.
Water
Resources
Management
732
https://doi.org/10.1007/s11269-014-0767-z
28,
4685–4697.
733
Valipour, M., Banihabib, M., Behbahani, S., 2012. Monthly Inflow Forecasting using
734
Autoregressive Artificial Neural Network. Journal of Applied Sciences 12,
735
2139–2147.
736
Valipour, M., Banihabib, M.E., Behbahani, S.M.R., 2013. Comparison of the ARMA,
737
ARIMA, and the autoregressive artificial neural network models in forecasting
738
the monthly inflow of Dez dam reservoir. Journal of Hydrology 476, 433–441.
739
https://doi.org/10.1016/j.jhydrol.2012.11.017
740
Vapnik, V.N., 1995. The nature of statistical learning theory. Springer, New York.
741
Vogel, R.M., Lall, U., Cai, X., Rajagopalan, B., Weiskel, P.K., Hooper, R.P., Matalas,
742
N.C., 2015. Hydrology: The interdisciplinary science of water. WAT.
743
RESOUR. RES. 51, 4409–4430. https://doi.org/10.1002/2015WR017049
744
Wang, W.C., Chau, K.W., Cheng, C.T., Qiu, L., 2009. A comparison of performance
745
of several artificial intelligence methods for forecasting monthly discharge
746
time
747
https://doi.org/10.1016/j.jhydrol.2009.06.019
series.
Journal
of
Hydrology
374,
294–306.
748
Wang, W.-C., Chau, K.-W., Cheng, C.-T., Qiu, L., 2009. A comparison of
749
performance of several artificial intelligence methods for forecasting monthly
750
discharge
751
https://doi.org/10.1016/j.jhydrol.2009.06.019
time
series.
Journal
of
Hydrology
374,
294–306.
752
Wei, S., Yang, H., Song, J., Abbaspour, K., Xu, Z., 2013. A wavelet-neural network
753
hybrid modelling approach for estimating and predicting river monthly flows.
754
Hydrological
755
https://doi.org/10.1080/02626667.2012.754102
Sciences
Journal
58,
374–389.
756
Willmott, C.J., 1984. On the Evaluation of Model Performance in Physical
757
Geography, in: Gaile, G.L., Willmott, C.J. (Eds.), Spatial Statistics and
758
Models.
759
https://doi.org/10.1007/978-94-017-3048-8_23
Springer
Netherlands,
Dordrecht,
pp.
443–460.
760
Willmott, C.J., Wicks, D.E., 1980. An Empirical Method for the Spatial Interpolation
761
of Monthly Precipitation within California. Physical Geography 1, 59–73.
762
https://doi.org/10.1080/02723646.1980.10642189
763
Wu, C.L., Chau, K.W., Li, Y.S., 2009a. Predicting monthly streamflow using data-
764
driven models coupled with data-preprocessing techniques. Water Resources
765
Research 45. https://doi.org/10.1029/2007WR006737
766
Wu, C.L., Chau, K.W., Li, Y.S., 2009b. Methods to improve neural network
767
performance in daily flows prediction. Journal of Hydrology 372, 80–93.
768
https://doi.org/10.1016/j.jhydrol.2009.03.038
769
Yaseen, Z.M., El-shafie, A., Jaafar, O., Afan, H.A., Sayl, K.N., 2015a. Artificial
770
intelligence based models for stream-flow forecasting: 2000–2015. Journal of
771
Hydrology 530, 829–844. https://doi.org/10.1016/j.jhydrol.2015.10.038
772
Yaseen, Z.M., El-shafie, A., Jaafar, O., Afan, H.A., Sayl, K.N., 2015b. Artificial
773
intelligence based models for stream-flow forecasting: 2000–2015. Journal of
774
Hydrology 530, 829–844. https://doi.org/10.1016/j.jhydrol.2015.10.038
775
Yaseen, Zaher M, Jaafar, O., Deo, R.C., Kisi, O., Adamowski, J., Quilty, J., El-shafie,
776
A., 2016. Stream-flow forecasting using extreme learning machines: A case
777
study
in
a
semi-arid
region
in
Iraq.
778
https://doi.org/10.1016/j.jhydrol.2016.09.035
Journal
of
Hydrology.
779
Yaseen, Zaher Mundher, Kisi, O., Demir, V., 2016. Enhancing Long-Term
780
Streamflow Forecasting and Predicting using Periodicity Data Component:
781
Application of Artificial Intelligence. Water Resources Management.
782
https://doi.org/10.1007/s11269-016-1408-5
783
Yu, S., Lu, H., 2018. An integrated model of water resources optimization allocation
784
based on projection pursuit model – Grey wolf optimization method in a
785
transboundary
786
https://doi.org/10.1016/j.jhydrol.2018.02.033
river
basin.
Journal
of
Hydrology
559,
156–165.
787
Zealand, C.M., Burn, D.H., Simonovic, S.P., 1999. Short term streamflow forecasting
788
using artificial neural networks. Journal of Hydrology 214, 32–48.
789
https://doi.org/10.1016/S0022-1694(98)00242-X
790
Zhang, X., Peng, Y., Zhang, C., Wang, B., 2015. Are hybrid models integrated with
791
data preprocessing techniques suitable for monthly streamflow forecasting?
792
Some
793
https://doi.org/10.1016/j.jhydrol.2015.09.047
experiment
evidences.
Journal
of
Hydrology
530,
137–152.
794
Zia, H., Harris, N., Merrett, G., Rivers, M., 2015. Predicting Discharge Using a Low
795
Complexity Machine Learning Model. Comput. Electron. Agric. 118, 350–
796
360. https://doi.org/10.1016/j.compag.2015.09.012
797
Zounemat-Kermani, M., Teshnehlab, M., 2008. Using adaptive neuro-fuzzy inference
798
system for hydrological time series prediction. Applied Soft Computing
799
Journal 8, 928–936. https://doi.org/10.1016/j.asoc.2007.07.011
800 801
802 803 35.00
AUG.
SEP.
OCT.
NOV.
DEC.
FEB.
MAR.
APR.
MAY
JUN.
JAN.
Natural Inflow (BCM)
30.00 25.00 20.00 15.00 10.00 5.00
1872-73 1876-77 1880-81 1884-85 1888-89 1892-93 1896-97 1900-01 1904-05 1908-09 1912-13 1916-17 1920-21 1924-25 1928-29 1932-33 1936-37 1940-41 1944-45 1948-49 1952-53 1956-57 1960-61 1964-65 1968-69 1972-73 1976-77 1980-81 1984-85 1988-89 1992-93 1996-97 2000-1
0.00
Year
804 805 806
Fig. 1 Monthly Natural Nile river flow at Aswan High Dam for the period between 1871 and 2000
807 808 809 810 35
High Inflow Mediam Inflow Low Inflow
30
Inflow (BCM)
25 20 15 10 5 0
811 812 813 814
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Sep
Oct
Nov
Dec
Fig. 2 Monthly hydrological characteristics at AHD. a) Monthly natural inflow classes (box-and-whisker plot)
815 816 817 818 819 820 821 822
Lag (Month)
Lag (Month)
823 824 825 826 827
Fig. 3 Autocorrelation and partial autocorrelation functions of Aswan station time series
828
a
829 830 831 832 833 834 835 836 837 838 839 840 841 842
843
b
844 845 846 847
Fig. 4 (a) The social hierarchy of grey wolves, (b) Illustration of position updating mechanism of ω wolves according to positions of α, β and δ wolves (Source Al Shorman et al., 2019).
848 849 850 851 852
Fig. 5 Flowchart of the proposed hybrid models. a: hybrid ANN-GWO. b: hybrid MLR-GWO and c: hybrid SVR-GWO
853
854 855
Fig. 6 Tim e vari atio n grap hs of the obs erve d and esti mat ed stre amf low s by
856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882
-3
883 884
Fig. 7 Scatterplots for the developed models in the testing period.
885 886 100 887 R el 50 888 at 0 889 iv -50 e 890 er -100 ro 891 0 r
SVR-GWO-3 Test
50
100
150
200
250
300
350
400
450
500
100
892 R el 50 893 at 0 iv 894-50 e er -100 895 ro 0 r 896
ANN-GWO-3 Test
50
100
150
200
250
300
350
400
100
897 R el 50 898 at 0 iv 899-50 e er -100 900 ro 0 901 r
450
500
MLR-GWO-3 Test
50
100
150
200
250
300
350
400
450
500
100
902 R el 50 903 at 0 iv 904-50 e -100 er 905 ro 0 906 r
SVR-3 Test
50
100
150
200
250
300
350
400
450
500
100
907 R el 50 908 at 0 iv 909-50 e er -100 ro 0 r
ANN-3 Test
50
100
150
200
250
300
350
400
450
500
910
Statistical index Min. Mean
Max.
Streamflow classes (Median value) High Medium Low
Month
(BCM/month)
(BCM/month)
(BCM/month)
(BCM/month)
(BCM/month)
(BCM/month)
Aug Sep Oct Nov Dec Jan Feb Mar Apr May Jun Jul
29.10 32.79 27.40 14.40 11.00 7.70 6.04 5.81 5.26 4.72 5.16 11.03
6.50 7.31 5.97 4.12 2.83 1.72 1.15 1.07 0.95 0.80 0.90 1.74
18.96 21.51 14.52 8.11 5.54 4.31 3.16 2.75 2.52 2.33 2.25 5.24
27.5 31 21.2 10.9 6.5 4.8 3.7 3.5 2.7 2.5 2.8 7.7
20.4 24.05 15.6 7.3 4.3 3.15 1.95 1.7 1.15 1.35 1.65 4.75
15.05 18.55 11.3 4.75 2.7 1.9 0.8 0.55 0.3 0.65 0.9 2.8
911
Variable Sets Minimum Maximum (BCM/month) (BCM/month)
Training Testing Whole data
0.80 2.09 0.80
32.00 32.79 32.79
Statistical parameters Mean (BCM/month) Standard deviation 7.76 7.16 7.58
Skewness Kurtosis
7.48 5.77 7.01
1.25 1.79 1.39
0.37 2.52 0.84
912 913 914 915
Fig. 8 Relative error between observed and predicted models in the testing period.
916 917 918 919 920 921 922 923 924 925 926
Table 1. Statistical Summary and streamflow classes for the natural streamflow at AHD based the period between 1871 and 2000 Table 2. Statistical parameters of training and testing datasets for study station
Table 3. Different input combinations for forecasting future river inflow Models
INPUTS
OUTPUT
M1 M2 M3 M4 M5 M6 M7 M8
Qt-1, Qt-2 Qt-1, Qt-2, Qt-10, Qt-11, Qt-12, Qt-13 Qt-1, Qt-2, Qt-11, Qt-12, Qt-13 Qt-1, Qt-11, Qt-12, Qt-13 Qt-1, Qt-4, Qt-5 Qt-6, Qt-7, Qt-8, Qt-11, Qt-12, Qt-13 Qt-1, Qt-5 Qt-6, Qt-7, Qt-11, Qt-12, Qt-13 Qt-1, Qt-2, Qt-4, Qt-5, Qt-6, Qt-7, Qt-8, Qt-11, Qt-12, Qt-13 Qt-1, Qt-2, Qt-3, Qt-4, Qt-5 Qt-6, Qt-7, Qt-8, Qt-9, Qt-10, Qt-11, Qt-12, Qt-13
Qt Qt Qt Qt Qt Qt Qt Qt
927
Table 4. Initial parameters of the GWO meta-heuristic algorithm
928 Models
Number of search agent
SVR-GWO
03
ANN-GWO
MLR-GWO
929
Equation (23)
Equation (6)
Parameters
Value
Α
Min = 0 and max = 2
Number of agents
100
Iterations number
1000
Range partitions (C)
[10-5, 105] (Sudheer et al., 2014)
Range partitions (γ)
[10-5, 101] (Sudheer et al., 2014)
Range partitions (ε)
[10-5, 101] (Sudheer et al., 2014)
Α
Min = 0 and max = 2
Number of agents
100
Iterations number
1000
Range partitions (αi)
[-3, +3]
Α
Min = 0 and max = 2
Number of agents
100
Iterations number
1000
Range partitions (αi)
[-10-3, 103]
Table 5. RMSE, MAE, R, and NSE values during training and testing periods of
930
SVR-GWO parameters
Training 70% (1083 Samples)
Testing 30% (464 Samples)
γ
C
ε
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
M1
0.06151
2.978
0.01187
1.8960
1.0762
0.9682
0.9356
0.9828
2.7808
1.8187
0.8910
0.7676
0.9420
M2
0.00060
110.960
0.31953
1.6010
0.8881
0.9769
0.9541
0.9880
2.0756
1.1559
0.9332
0.8705
0.9649
M3
0.00328
6.706
0.43790
1.5578
0.8521
0.9782
0.9566
0.9886
2.0570
1.2005
0.9363
0.8728
0.9671
M4
0.00560
1.870
0.04146
1.7769
0.9753
0.9719
0.9435
0.9849
2.1937
1.2751
0.9269
0.8554
0.9613
M5
0.00046
24.147
0.25284
1.8461
0.9870
0.9693
0.9390
0.9837
2.2880
1.3038
0.9189
0.8427
0.9561
M6
0.00461
3.417
0.08112
1.6949
0.9113
0.9744
0.9486
0.9863
2.2376
1.3184
0.9245
0.8495
0.9583
M7
0.00068
39.065
0.02235
1.6395
0.8651
0.9758
0.9519
0.9874
2.1126
1.2318
0.9317
0.8659
0.9632
M8
0.00201
4.332
0.06883
1.5849
0.8495
0.9775
0.9550
0.9882
2.1315
1.2830
0.9303
0.8635
0.9637
SVR-GWO model
931 932 933 934 935 936 937
Table 6. RMSE, MAE, R, and NSE values during training and testing periods of
938
Training 70% (1083 Samples)
Testing 30% (464 Samples)
ANN Structure
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
M1
2-5-1
2.4293
1.7153
0.9458
0.8944
0.9711
3.1324
2.3454
0.8565
0.7051
0.9192
M2
6-13-1
1.8122
1.0781
0.9702
0.9412
0.9846
2.2625
1.3041
0.9246
0.8461
0.9608
M3
5-11-1
1.7663
1.0168
0.9717
0.9441
0.9854
2.1883
1.2999
0.9314
0.8561
0.9636
M4
4-9-1
1.9651
1.1333
0.9648
0.9309
0.9818
2.3782
1.3947
0.9118
0.8300
0.9520
M5
9-19-1
1.9280
1.0994
0.9662
0.9335
0.9825
2.3850
1.5014
0.9167
0.8290
0.9556
M6
7-15-1
1.8104
1.0648
0.9702
0.9413
0.9847
2.2560
1.3595
0.9228
0.8470
0.9598
M7
10-21-1
1.8680
1.1041
0.9683
0.9375
0.9836
2.2126
1.3393
0.9278
0.8529
0.9626
M8
13-27-1
1.8485
1.1337
0.9689
0.9388
0.9839
2.3845
1.5187
0.9177
0.8291
0.9556
Model
ANN-GWO
SVR-GWO
Model
939
ANN-GWO model
940 941
Table 7. RMSE, MAE, R, and NSE values during training and testing periods of SVR model
SVR parameters
SVR
Model
Training 70% (1083 Samples)
Testing 30% (464 Samples)
γ
C
ε
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
RMSE (m3/s)
MAE (m3/s)
R
NSE
WI
M1
0.01
1
0.001
3.1180
1.6409
0.9178
0.8260
0.9481
3.2319
1.9335
0.8361
0.6861
0.9099
M2
0.01
1
0.001
1.8603
0.9934
0.9718
0.9380
0.9826
2.3335
1.3235
0.9148
0.8363
0.9543
M3
0.01
1
0.001
1.8253
0.9780
0.9724
0.9404
0.9834
2.2248
1.3190
0.9244
0.8512
0.9605
M4
0.01
1
0.001
1.8635
1.0213
0.9702
0.9378
0.9829
2.2403
1.3216
0.9239
0.8492
0.9587
M5
0.01
1
0.001
2.1084
1.0996
0.9654
0.9204
0.9770
2.5427
1.4786
0.9111
0.8057
0.9405
M6
0.01
1
0.001
2.0247
1.0763
0.9670
0.9266
0.9791
2.4034
1.4161
0.9167
0.8264
0.9491
M7
0.01
1
0.001
2.0979
1.0703
0.9668
0.9212
0.9770
2.4938
1.4211
0.9090
0.8131
0.9432
M8
0.01
1
0.001
2.1878
1.1012
0.9641
0.9143
0.9748
2.7126
1.5005
0.8927
0.7788
0.9271
Table 8. RMSE, MAE, R, and NSE values during training and testing periods of ANN model
942 943
ANN
Model
ANN Structure
Training 70% (1083 Samples) RMSE MAE R NSE WI (m3/s) (m3/s)
Testing 30% (464 Samples) RMSE MAE R NSE (m3/s) (m3/s)
M1
2-5-1
1.9242
1.2552
0.9663
0.9337
0.9826
3.2161
2.2878
0.8640
0.6891
0.9230
M2
6-13-1
1.2031
0.7478
0.9870
0.9741
0.9934
2.4954
1.4227
0.9103
0.8128
0.9529
M3
5-11-1
1.3110
0.8006
0.9845
0.9692
0.9921
2.4214
1.3105
0.9118
0.8238
0.9539
M4
4-9-1
1.4532
0.8774
0.9809
0.9622
0.9903
2.5653
1.4181
0.9019
0.8022
0.9485
M5
9-19-1
1.1254
0.6876
0.9886
0.9773
0.9942
2.8792
1.7276
0.8776
0.7508
0.9348
M6
7-15-1
1.2462
0.7592
0.9860
0.9722
0.9929
2.8793
1.6991
0.8747
0.7508
0.9304
M7
10-21-1
1.0304
0.6328
0.9905
0.9810
0.9952
2.5973
1.5167
0.9149
0.7972
0.9526
M8
13-27-1
0.7334
0.4738
0.9952
0.9904
0.9976
3.5392
2.0917
0.8605
0.6235
0.9170
Table 9. RMSE, MAE, R, and NSE values during training and testing periods of MLR-GWO model
944 945
Training 70% (1083 Samples) Model
M1 M2
MLR-GWO
WI
M 3 M4 M5 M6 M7 M8
RMS E (m3/s) 4.011 9 2.170 9 2.036 6 2.087 4 3.963 0 2.844 7 2.067 7 2.751
Testing 30% (464 Samples)
MAE (m3/s)
R
NSE
WI
2.729 7 1.275 9 1.125 5 1.164 2 2.682 3 1.749 3 1.180 3 1.816
0.843 7 0.957 2 0.962 3 0.960 2 0.848 2 0.925 1 0.961 0 0.929
0.711 9 0.915 6 0.925 7 0.922 0 0.718 8 0.855 1 0.923 5 0.864
0.909 0 0.977 9 0.980 6 0.979 4 0.915 2 0.960 3 0.979 8 0.962
RMS E (m3/s) 3.916 5 2.518 1 2.388 3 2.445 0 4.066 4 3.046 7 2.417 6 2.952
MAE (m3/s)
R
NSE
WI
2.741 7 1.459 7 1.337 0 1.359 2 2.761 0 1.694 0 1.365 0 1.815
0.751 1 0.902 9 0.912 1 0.907 7 0.721 7 0.853 5 0.909 6 0.861
0.539 0 0.809 4 0.828 6 0.820 3 0.503 0 0.721 0 0.824 3 0.738
0.856 3 0.948 9 0.953 8 0.951 5 0.836 2 0.920 2 0.952 5 0.923
1
7
8
5
4
4
2
6
0
8
946
Table 10. Optimal MLR-GWO Parameters.
947
Model
MLR Equation Qt = 1.17417 *Qt-1 - 0.62788 Qt-2 + 3.50692
M1
M2 Qt = 0.66807*Qt-1 + 0.03843*Qt-2 + 0.0651*Qt-10 - 0.08724*Qt-11 + 0.97596*Qt-12 - 0.62635*Qt-13 0.61387 M3
Qt = 0.73655*Qt-1- 0.0.05822*Qt-2+ 0.07324*Qt-11 + 0. 79472*Qt-12 - 0. 54969*Qt-13 - 0.06750
M4
Qt = 0. 67595*Qt-1+ 0. 01034*Qt-11+ 0. 92619*Qt-12+ 0. 61721*Qt-13 -17546
MLR-GWO
M5 -0.03010 0.60628
-0.12549 0.12141 0.09536
-0.15357 0.09885
-0.02450
-1.42590 0.59310 Qt = -02450*Qt-1+0. 09885*Qt-4 -0. 15357*Qt-5 +0. 09536*Qt-6-0. 12141*Qt-7 -0.12549*Qt-8 0.60628*Qt-11 -03010*Qt-12 + 59310*Qt-13 -1.42590 M6
Qt = -0. 00085*Qt-1+0. 02141*Qt-5 -0. 02274*Qt-6 +0. 01284*Qt-7 +0.04802*Qt-11 +87295*Qt-12 +0.05226*Qt-13 -0.00142
M7
Qt =0. 68348*Qt-1 -0. 01288*Qt-2 - 0. 03845*Qt-4+ 0. 03824*Qt-5 -0. 04732*Qt-6+ 0. 01018*Qt-7+ 0. 00571*Qt-8- 0. 00068*Qt-11+ 0. 89217*Qt-12 -0. 60036*Qt-13+ 0. 53098
M8
Qt =0. 0000003*Qt-1 +0. 03549*Qt-2 -0. 08575*Qt-3+ 0. 01157*Qt-4 -0. 04458*Qt-5- 0. 05618*Qt6+
0. 01988*Qt-7- 0. 07905*Qt-8+ 0. 02019*Qt-9- 0. 10697*Qt-10+ 0. 09547*Qt-11+ 0. 79328*Qt-12+ 0. 00602*Qt-13+ 3.01200
948 949 950
Table 11. ANN Parameters versus ANN-GWO parameters for the best model (M3).
951
ANN
Model
Parameters IW1_1 =[0.645548352588935 0.490849210105060 -0.636814115741003 -0.197023018606021 1.15095937712121;-0.374741951566058 2.09602351895915 -0.948572086483014 0.448172347550920 -0.390243232966560;-0.508664569672024 1.42473219848372 1.81346308646462 -0.422220219074745 0.307797907505390;1.69099385336773 1.91762647812312 -1.37118240411381 0.991561905462959 -0.0645409927642741;1.16667356534635 1.03770562438892 -0.655038460261841 0.365195799970887 1.39859811955597;-0.436086805954856 -0.402680235576198 -0.674915728267535 0.573879996013445 0.970907264566725;1.40043825812518 -0.894001252195594 0.387858261897225 -0.235609833166580 1.22505735374624;0.157337536509532 0.440907790543129 2.09832896723230 -3.05274752078594 1.20295125926541;1.94877786008619 0.653535998773184 0.747141370532021 1.24502310208271 0.0407935886488361;0.995475357888400 0.663088707022791 -1.23122998045925 0.850028887499989 -0.420747514818101;0.507805643987633 1.25112356595343 0.347455862427703 -1.53293583422575 -0.415921708344004]
b1=[0.364567778992709;-0.253277415388278;0.0776162521275670;-0.335489901278488;1.49190695238374;1.23261751911969;0.766095010968687;-1.12821463276292;1.53241532465276;0.262077228973190;-1.24795736214157] LW2_1= [-1.27781252764868 0.750943129520385 -1.16056523012868 0.711855144075440 1.66920057928637 -1.07462179886860 -1.81513806044960 -0.706292165764873 1.35623649635319 1.27555964001351 0.568715308486404]
ANN-GWO
b2= [-0.523926338451804] IW1_1=[0.679633208818835 -0.0598736973233613 0.0290671638159691 -0.118898973732041 1.19996763002510;0.00969730241328533 -0.119191575741640 0.0194353130324328 1.07549721717923 0.652667379286810;-4.98948316200092e-05 -0.285414929657196 -1 0.346887454209408 0.722057669039339;-0.00385812223558482 -0.704007857206319 0.00969361579187186 0.00202473907503717 0.102333284563727;-0.0530813631767349 0.00874038005562050 0.940658603673713 1.88808217020703 0.558951907069030;1.45027460136395 0.0369561267210984 0.0938558513359091 0.0202778143417594 -0.0287463129026476;-0.129114564501272 -0.999595557587754 0.559480176726186 1.93021116801354 -0.0427165998042683;-0.0432015511228707 0.425699113581515 0.0390928538728479 -0.0269531954052187 0.653111091749518;0.0208566152618470 0.198693391722621 -0.0272336852909195 0.000128014264935471 0.691879316989853;0.927288538383699 -0.486089339990834 0.00749547491297705 0.532970735519229 -0.805370703363221;-0.455084554781621 0.157062590076071 -0.0106175418622636 -0.568214708185782 -0.0186313139069768] b1=[-0.999544626417181;-0.0257826944723780;1.54804395315099;0.419551508628818;0.201266019663968;1.33194412312976;-0.223980124713737;-0.355168069400448;0.0536968092305320;-0.138026029294201;-0.999752148603694] LW2_1=[0.519446674804311 -0.308173183218543 0.214442030218500 -0.0186771941202023 0.0485869393744300 0.0251913808893511 0.243068491697808 -0.0542094341107475 0.184388522291043 0.581590936122574 -0.895940329672519] b2=[-0.564570091202907]
952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971
Where: IW1_1: is the weight matrix between the input layer and the hidden layer. b1: is the bias vector between the input layer and the hidden layer. LW2_1 is the weight vector between the hidden layer and the output layer. b2: is the bias vector between the hidden layer and the output layer.
Improving Artificial Intelligence Models Accuracy for Monthly Streamflow Forecasting Using Grey Wolf Optimization (GWO) Algorithm Yazid Tikhamarine1, Doudja Souag-Gamane1, Ali Najah Ahmed2, Ozgur Kisi3 and Ahmed El-Shafie4 1LEGHYD Laboratory, Department of Civil Engineering, University of Scienceand Technology Houari Boumediene, BP 32, BabEzzouar, Algiers, Algeria 2Institute for Energy Infrastructure (IEI), Universiti Tenaga Nasional (UNITEN), Kajang 43000, Selangor Darul Ehsan, Malaysia 3School of Technology, Ilia State University, Tbilisi, Georgia 4Department of Civil Engineering, Faculty of Engineering, University of Malaya (UM), 50603 Kuala Lumpur, Malaysia
Abstract
972
Monthly streamflow forecasting is required for short- and long-term water resources
973
management especially in extreme events such as flood and drought. Therefore, there
974
is need to develop a reliable and precise model for streamflow forecasting. The
975
precision of Artificial Intelligence (AI) models can be improved by using hybrid AI
976
models which consist of coupled models. Therefore, the chief aim of this study is to
977
propose efficient hybrid system by integrating Grey Wolf Optimization (GWO)
978
algorithm with Artificial Intelligence (AI) models. 130 years of monthly historical
979
natural streamflow data will be used to evaluate the performance of the proposed
980
modelling technique. Quantitative performance indicators will be introduced to
981
evaluate the validity of the integrated models; in addition to that, comprehensive
982
analysis will be conducted between the predicted and the observed streamflow.The
983
results show the integrated AI with GWO outperform the standard AI methods and
984
can make better forecasting during training and testing phases for the monthly inflow
985
in all input cases. This finding reveals the superiority of GWO meta-heuristic
986
algorithm in improving the accuracy of the standard AI in forecasting the monthly
987
inflow.
988
Keywords: Streamflow, Estimation, Evolutionary algorithms, Aswan High Dam
989 990 991
Monthly streamflow forecasting is required for short- and long-term water resources management especially in extreme events such as flood and drought.
992 993 994 995
Grey Wolf Optimization (GWO) algorithm has been integrated to Support Vector Machine (SVM), Artificial Neural Network (ANN) and Multiple
996
Linear Regression (MLR) in order to compare the accuracy of the proposed
997
model.
998 999
Integrated AI with GWO outperform the standard AI methods and can make
1000
better forecasting during training and testing phases for the monthly inflow in
1001
all input cases.
1002 1003
The findings reveal the superiority of GWO meta-heuristic algorithm in
1004
optimizing the parameter of the standard SVR to improve its accuracy in
1005
forecasting the streamflow.
1006 1007
Yazid Tikhamarine : Methodology, Software, Visualization.
1008
Doudja Souag-Gamane: Conceptualization, Supervision, Writing- Reviewing and Editing
1009
Ali Najah Ahmed: Data curation, Writing- Original draft preparation, Validation.
1010
Ozgur Kisi: Writing- Original draft preparation, Writing- Reviewing and Editing,
1011
Supervision.
1012
Ahmed El-Shafie: Writing- Original draft preparation, Writing- Reviewing and Editing,
1013
Supervision.
1014 1015 1016
Declaration of interests
1017 1018 1019
☒ The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
1020 1021 1022 1023
☒The authors declare the following financial interests/personal relationships which may be considered as potential competing interests:
The authors declare no conflict of interest.
1024 1025 1026 1027 1028