Accepted Manuscript
Optimal solution for novel grey polynomial prediction model Baolei Wei, Naiming Xie, Aqin Hu PII: DOI: Reference:
S0307-904X(18)30288-9 10.1016/j.apm.2018.06.035 APM 12334
To appear in:
Applied Mathematical Modelling
Received date: Revised date: Accepted date:
1 March 2018 12 June 2018 18 June 2018
Please cite this article as: Baolei Wei, Naiming Xie, Aqin Hu, Optimal solution for novel grey polynomial prediction model, Applied Mathematical Modelling (2018), doi: 10.1016/j.apm.2018.06.035
This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT
Highlights • The novel grey polynomial prediction model with nonlinear parameters is proposed. • The parameter optimization criteria in the existing grey prediction models are unified by weighted least square method.
• The algorithmic framework for selecting the polynomial order and estimating the nonlinear parameters
CR IP T
is proposed.
• The affine transformation of the accumulating sequence proves to be independent of modeling perfor-
AC
CE
PT
ED
M
AN US
mance.
1
ACCEPTED MANUSCRIPT
Optimal solution for novel grey polynomial prediction model Baolei Weia,∗, Naiming Xiea,b , Aqin Hua a College
of Economics and Management, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, PR China of Grey System Studies, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, PR China
CR IP T
b Institute
Abstract
The grey prediction model, as a time-series analysis tool, has been used in various fields only with partly known distribution information. The grey polynomial model is a novel method to solve the problem that the original sequence is in accord with a more general trend rather than the special homogeneous or nonhomogeneous trend, but how to select the polynomial order still needs further study. In this paper the tuned background coefficient is introduced into the grey polynomial model and then the algorithmic frame-
AN US
work for polynomial order selection, background coefficient search and parameter estimation is proposed. The quantitative relations between the affine transformation of accumulating sequence and the parameter estimates are deduced. The modeling performance proves to be independent of the affine transformation. The numerical example and application are carried out to assess the modeling efficiency in comparison with other conventional models.
1. Introduction
ED
1
M
Keywords: grey prediction model, background coefficient, weighted least square method, line search
Time series prediction has always been an important issue in economic, finance, marketing, as well as
3
social problems. Nowadays, there exist hundreds of sequential prediction techniques, such as LR (Linear
4
Regression), ES (Exponential Smoothing), ARIMA (Auto Regressive Integrated Moving Average), SVR
5
(Support Vector Regression), HMM (Hidden Markov Model) and ANN (Artificial Neural Network). The
6
results of these methods are always ensured just under a basic assumption of known distribution or large
7
sample. However, it is sometimes impossible to collect time series with large sample size especially for a
8
new system, such as the cold-start problem in recommendation system. In Deng’s seminal study [1], grey
9
system theory focusing on the small sample problems was first proposed. Grey prediction is an important
10
branch of grey system theory and generally written as GM(p, q) which is a p-th order model with q variables.
12 13
CE
AC
11
PT
2
Over the last 35 years, it has been widely utilized in various fields such as energy, economics, environment, transportation and so on [2–4]. The GM(1,1) is the foundation of grey prediction models and has been substantiated in real world
14
applications. In the meantime, much importance has been consistently placed on improving the modeling
15
accuracy, and most of the effort is invested in the following aspects: background coefficient optimization [5–8],
16
initial condition selection [9–12], hybrid approach combined with background coefficient and initial condition ∗ Corresponding
author. Email addresses:
[email protected] (Baolei Wei),
[email protected] (Naiming Xie)
Preprint submitted to Applied Mathematical Modelling
June 22, 2018
ACCEPTED MANUSCRIPT
17
optimization [13], fractional order accumulation [14], time-varying weighted accumulating generation [15],
18
properties analysis including the necessary and sufficient condition for modeling [16], the relative error upper
19
bound estimation [17], the independence of model’s accuracy and affine transformation [18], the influence of
20
sample size on modeling performance [19–21]. The modeling procedures of univariate grey prediction models are represented by GM(1,1) model, and
22
many extension models have already been reported in recent years. Guo and Guo [22] probed into the nature
23
of grey prediction model and looked ahead to the extension principle but short of details. Subsequently,
24
Cui et al. [23] put forward NGM(1,1) model by replacing the grey action constant with time linear function
25
so that it is applicable to the sequence in accord with quasi non-homogeneous exponential law. Similarly,
26
Qian, Dang and Liu [24] developed GM(1,1,tα ) model by adding a time power function to the grey action
CR IP T
21
term, and NGM(1,1) model proves to be a special form of this model with α = 1. In a general way, Ma, Hu and Liu [25] constructed a kernel regularized grey model KGM(1,1) by using a function of time as the
29
grey action term, and Luo and Wei [26] proposed a grey polynomial model GPM(1,1,N ) by using a time
30
polynomial function as the grey action term. These two models have larger application ranges, but each of
31
them suffers from some limitations, respectively, such as the selection of regularized parameter and kernel
32
function in KGM(1,1) model, and the selection of polynomial order in GPM(1,1,N ) model.
AN US
28
27
The discrete grey prediction models have long been considered novel methodologies, and they are essen-
34
tially the extensions of the existing models, such as DGM(1,1) model [27] corresponding to the GM(1,1)
35
model, and NDGM(1,1) model [28] corresponding to the NGM(1,1) model, etc. The multivariate grey predic-
36
tion models can be also regarded as extensions, such as GM(1,N ) model [29] corresponding to the GM(1,1)
37
model, NGM(1,N ) model [30] corresponding to the NGM(1,1) model, and KGM(1,N ) model corresponding
38
to the KGM(1,1) model [31], etc. In addition, the optimization strategies in the GM(1,1) model have been
39
gradually employed to improve the accuracies of the discrete and multivariate grey models. For example,
40
Zeng and Li [32] improved the NGM(1,N ) model by using the particle swarm optimization algorithm to
41
search the optimal background coefficient.
ED
M
33
In summary, the univariate model is the basis of the discrete and multivariate model construction. Luo
43
and Wei [26] proved that the above GM(1,1), NGM(1,1) and GM(1,1,tα ) model are all special forms of the
44
GPM(1,1,N ) model with deterministic polynomial order, and this model is applicable to including but not
45
limited to the quasi homogeneous and non-homogeneous exponential sequences, or even some fluctuating
46
sequences. However, there are still some drawbacks that are worth to be further investigated. In this paper,
47
the major points are summarized as follows: (1) the polynomial order is usually unknown in practice, and
48
the empirical criterion [26] is given under the assumption that the original sequence is composed of the
49
exponential and polynomial functions without noise and thus is likely to be inaccurate; (2) the background
51 52 53 54 55
56 57
CE
AC
50
PT
42
coefficient is fixed at 0.5 in the existing model, which may result in poor predictions sometimes; (3) the least squared error, least squared percentage error and least absolute percentage error are three common criteria in model parameter estimation, nevertheless the relationships among them and their corresponding properties have not yet been researched. In order to address these drawbacks in the existing approaches, we focus on the polynomial order selection and parameter estimation. In detail, the main work lies in the following aspects: • The grey polynomial model with tuned background coefficient is derived from the fundamental form,
and the algorithmic framework for nonlinear parameter (polynomial order, background coefficient and 3
ACCEPTED MANUSCRIPT
58
59
model parameters) estimation is presented. • The background coefficient and model parameters are estimated by line-search based weighted least
60
square method in which the weighted objective function is turned out to be the unified representation
61
of the three above optimization criteria in grey prediction models.
63
• The properties including the comparison between the proposed model and other grey prediction models as well as the influence of affine transformation on modeling performance are deeply analyzed.
CR IP T
62
64
The remainder of this paper has the following structure: the proposed model and its parameter estimation
65
algorithm are elaborated in Section 2; the properties are researched in Section 3; the numerical example and
66
application are respectively presented in Sections 4 and 5; the conclusions are drawn in Section 6.
67
2. Construction of GPMB(1,1,N ) model
68
2.1. Representation of GPMB(1,1,N ) model
70
71
where
k X i=1
(1)
x(0) (k), k = 1, 2, · · · , n.
(2)
M
The grey polynomial model is suggested to be the following difference equation x(1) (k − 1) + x(1) (k) 2k − 1 k N +1 − (k − 1) + β0 + β1 + · · · + 2 2 N +1
N +1
βN ,
ED
x(0) (k) = α 73
x(0) (1), x(0) (2), · · · , x(0) (n) , n ≥ 4, its first order
accumulating generation sequence is defined as n o X (1) = x(1) (1), x(1) (2), · · · , x(1) (n) , x(1) (k) =
72
Assuming that the original sequence is X (0) =
AN US
69
and its whitening equation is expressed as
d (1) x (t) = αx(1) (t) + β0 + β1 t + · · · + βN tN , dt
(4)
PT
74
(3)
where the order of polynomial satisfies N ≤ n − 4 and α, β0 , β1 , · · · , βN are model parameters.
The difference expression Eq. (3) is a roughly discrete approximation of Eq. (4), and the accuracy of
76
this approximate formula can be further improved. Without loss of generality, consider the integration on the both side of Eq. (4) in the interval [k − 1, k], then Z k Z k Z k Z (1) (1) x (t)dt = α x (t)dt + β0 dt + β1
AC
77
CE
75
k−1
78 79
k−1
k−1
k
k−1
tdt + · · · + βN
Z
k
tN dt.
(5)
k−1
According to the Newton-Leibniz formula and the definition of accumulating generation in Eq. (2), the
left side of Eq. (5) can be expressed as Z k x(1) (t)dt = x(1) (k) − x(1) (k − 1) = x(0) (k)
(6)
k−1
80
and the right side of Eq. (5) is equal to Z k N +1 k N +1 − (k − 1) 2k − 1 α x(1) (t)dt + β0 + β1 + · · · + βN . 2 N +1 k−1 4
(7)
ACCEPTED MANUSCRIPT
81 82
Rk
x(1) (t)dt denotes the area between t-axis and the curve x(1) (t) in the interval [k − 1, k]. It was calculated using the trapezoid formula in the traditional GPM(1,1,N ) model It is notable that the integration term
k−1
83
[26] and many other recent researches [14–17, 19–25, 29–31]. But the trapezoid formula is not accurate
84
enough regardless of that x(1) (t) is convex, concave or non-convex in each subinterval as shown in Figure 1.
85
y
y x (1) (t )
x (1 ) (k )
x (1 ) (k )
x (1) (k 1)
x (1) (k 1) k 1 (a) convex
k
y
y x (1) (t )
x (1) (k 1)
k 1 (b) concave
t
k
t
Figure 1: Geometric schematic representation of the integration term
87
Rk
k−1
theorem for definite integrals, there always exists a real number λ ∈ [0, 1] such that Z k x(1) (t)dt = λx(1) (k − 1) + (1 − λ)x(1) (k).
(8)
Substituting Eqs. (6), (7) and (8) into Eq. (5) gives the more accurate approximate formula, termed as GPMB(1,1,N ) model, expressed as
N +1 h i 2k − 1 k N +1 − (k − 1) x(0) (k) = α λx(1) (k − 1) + (1 − λ)x(1) (k) + β0 + β1 + · · · + βN , 2 N +1
where the unknown parameter λ ∈ [0, 1] is called background coefficient.
91
2.2. Framework for polynomial order selection and parameter estimation
(9)
ED
90
According to Eq. (9), it can be seen that the model parameters α, β0 , · · · , βN , background coefficient
PT
92
t
x(1) (t)dt.
M
89
k
(c) nonconvex
Assume that x(1) (t) is a monotone function in each subinterval, then according to the first mean value
k−1
88
k 1
AN US
86
y x (1) (t )
x (1 ) (k )
CR IP T
y
93
λ and polynomial order N have nonlinear relationship, which makes the ordinary least square method
94
inapplicable. In practice, the alternative polynomial order is always suggested to be N ∈ N = {0, 1, 2, 3} to
avoid overfitting and ill-condition problem [26, 33, 34], and therefore the enumeration strategy is acceptable.
96
On the basis of λ ∈ [0, 1], the line-search based weighted least square method is employed to search the
98 99
100 101 102 103
background coefficient and estimate the model parameters.
Line search algorithm is a numerical optimization method that can be applied to discontinuous or even
AC
97
CE
95
indifferentiable functions [34]. Each iteration is given by λi+1 = λi + ηpi
(10)
where η > 0 is the step length and pj = ±1 is the descent direction to guarantee the reduction of objective function.
Under each background coefficient, the model parameters α, β0 , β1 , · · · , βN are estimated from Eq. (9)
using the weighted least square method n o T −1 T 2 2 κ ˆ= α D W Y, ˆ βˆ0 βˆ1 · · · βˆN = arg min φ(κ) = kW (Y − Dκ)k2 = D T W 2 D κ
5
(11)
ACCEPTED MANUSCRIPT
105
106 107
where W = diag (w2 , w3 , · · · , wn ) is a positive definite matrix in which wk > 0 (k = 2, 3, · · · , n) is the weight of the k-th sample, and λi x(1) (1) + (1 − λi )x(1) (2) 1 (1) (1) 1 λi x (2) + (1 − λi )x (3) D= .. .. . . (1) (1) λi x (n − 1) + (1 − λi )x (n) 1
3 2 5 2
.. .
2n−1 2
··· ··· .. . ···
(0) x (2) x(0) (3) , Y = .. . .. . . nN +1 −(n−1)N +1 x(0) (n) 2N +1 −1 N +1 3N +1 −2N +1 N +1
Substituting the estimates α ˆ , βˆ0 , βˆ1 , · · · , βˆN into the general solution to the whitening equation (4), the
time response function (the details can be seen in reference [26]) is obtained as ˆ x ˆ(1) (t) = ceαt + γ0 + γ1 t + · · · + γN t N ,
110
(14)
constant c is optimized by minimizing the sum of the squared errors,
c
n X
k=1
x
(1)
(k) − x ˆ
(1)
(k)
2
n N X e2αˆ − 1 X (1) = 2αn x (k) − γi k i e ˆ −1 i=0 k=1
!
ˆ eα(k−2) .
(15)
Substituting the polynomial coefficients in Eq. (14) and the estimated integration constant in Eq. (15) into the time response function Eq. (13), the time response sequence is obtained as ! # " N n X e2αˆ − 1 X i α(k−2) ˆ ˆ (1) (1) γi k e eαk + γ0 + γ1 k + · · · + γN k N , x (k) − x ˆ (k) = 2αn e ˆ −1 i=0
ED
112
−1 0 βˆ0 .. ˆ β1 . . . .. N βˆN −ˆ α
(13)
Different form the traditional method taking x ˆ(1) (1) = x(1) (1) as the initial condition, the integration
cˆ = arg min 111
··· .. . .. . ···
AN US
109
where the polynomial coefficients γ0 , γ1 , · · · , γN are −ˆ α 1 γ0 γ1 0 −ˆ α γ= . = .. .. .. . . γN 0 0
M
108
(12)
N +1
CR IP T
104
(16)
k=1
116
PT
114 115
where k = 1, 2, · · · , n + p and p ∈ N+ is the number of samples to be predicted.
By using the first order inverse accumulating generation operator, the fitted and predicted values corre-
sponding to the original sequence can be obtained as n o ˆ (0) = x X ˆ(0) (1), x ˆ(0) (2), · · · , x ˆ(0) (n), x ˆ(0) (n + 1), · · · , x ˆ(0) (n + p) , where
CE
113
117 118
AC
x ˆ
(0)
(k) =
(
x ˆ(1) (1), x ˆ(1) (k) − x ˆ(1) (k − 1),
k=1 . k = 2, 3, · · · , n + p
(17)
(18)
The mean absolute percentage error (MAPE) is used to evaluate the fitting and predicting performance,
which is defined as MAPE =
T2 (0) X x (k) − x 1 ˆ(0) (k) × 100%, T2 − T1 + 1 x(0) (k)
(19)
k=T1
119
where MAPE is the fitting performance measure with T1 = 1 and T2 = n and the predicting performance
120
measure with T1 = n + 1 and T2 = n + p.
121 122
To sum up, the algorithmic framework for polynomial order selection, background coefficient search and parameter estimation is presented in Algorithm 1. 6
ACCEPTED MANUSCRIPT
Algorithm 1 Input: Original sequence X (0) , weighted matrix W . 1: Divide the original sequence into two parts: n o n o (0) (0) Xtrain = x(0) (1), x(0) (2), · · · , x(0) (n1 ) and Xtest = x(0) (n1 + 1), x(0) (n1 + 2), · · · , x(0) (n) ,
124
2.3. Computational steps of GPMB(1,1,N ) model
According to the above modeling processes, the computational steps of GPMB(1,1,N ) model can be
ED
123
M
AN US
CR IP T
where n denotes the number of samples in X (0) and traditionally n1 is the largest integer less than or equal to 54 n. 2: for all N ∈ N = {0, 1, 2, 3} do 3: Initialize the count i = 0 and the background coefficient λi = 0. 4: while λi ≤ 1 do 5: Calculate the estimates of model parameters by Eq. (11), the polynomial coefficients by Eq. (14) and the initial condition by Eq. (15); 6: Calculate the time response sequence by Eq. (16) and the fitted and predicted values corresponding to X (0) by Eq. (17); 7: Update the background coefficient by Eq. (10) in which the step length and descent direction can be set as η = 0.01 and pi = 1. 8: end while 9: Pick the best background coefficient and denote it as λopt . Here best is defined as having the smallest (0) MAPE on Xtrain , abbreviated as MAPEtrain . 10: end for 11: Select the best polynomial order and denote it as Nopt . Here best is defined as having the smallest (0) MAPE on Xtest , abbreviated as MAPEtest . Output: Nopt , λopt .
125
summarized as follows.
126
Step 1 Given the optimization criterion (i.e., the weighted matrix W ), compute the best polynomial order
130 131
132
133 134 135
PT
129
Step 2 Based on the whole sequence X (0) including all samples, calculate the time response sequence by substituting Nopt and λopt into Eq. (16); Step 3 Compute the fitted and predicted values corresponding to the original sequence by Eqs. (17) and
CE
128
and background coefficient according to Algorithm 1;
(18), as well as the performance measures by Eq. (19). 3. Affine property of GPMB(1,1,N ) model
AC
127
The weighted least square method is a unified representation of the common optimization criteria includ-
ing but not limited to the least squared error, least squared percentage error and least absolute percentage error. When wk = 1, the model parameters in Eq. (11) are estimated under the least squared error criterion.
136
ˆ = arg min κ κ
(
φ(κ) = kW (Y −
2 Dκ)k2
=
n X
k=2
7
x
(0)
(k) − x ˆ
(0)
2 (k)
)
.
(20)
ACCEPTED MANUSCRIPT
137
When wk = 1/x(0) (k), the model parameters in Eq. (11) are estimated under the least squared percentage
138
error criterion. ˆ = arg min κ κ
140
φ(κ) = kW (Y −
2 Dκ)k2
=
2 n (0) X x (k) − x ˆ(0) (k) x(0) (k)
k=2
)
.
(21)
When wk = 1/ x(0) (k) x(0) (k) − x ˆ(0) (k) , the model parameters in Eq. (11) are estimated under the least
absolute percentage error criterion. ( ˆ = arg min κ κ
φ(κ) = kW (Y −
2 Dκ)k2
) n (0) (0) X x (k) − x ˆ (k) . = x(0) (k)
CR IP T
139
(
k=2
(22)
Due to that the sample weights is not independent of the fitted values, the objective function in Eq. (22) is non-differentiable and the analytic solution in Eq. (11) is not applicable any more. The objective function
143
in Eq. (21) is the approximation of that in Eq. (22) which is always solved by the intelligent algorithms
144
such as particle swarm optimization and genetic algorithm [4].
145
Remark 1. Under the least squared error criterion, i.e., the sample weights wk = 1, let the background
146
coefficient be 0.5 and el be the (n − 1) × 1 vector having 1 in the l-th element and 0’s in others, then in
150 151
M
149
(1) When N = 0, GPMB(1,1,N ) model yields the GM(1,1) model [4] defined as i α h (1) x(0) (k) = x (k − 1) + x(1) (k) + β0 , 2 and the estimates of model parameters are
where D1 = D e1
ˆ= α κ ˆ
ED
148
GPMB(1,1,N ) model
e2 .
x(0) (k) =
153 154
155
156
T
= D1T D1
−1
D1T Y
i 2k − 1 α h (1) x (k − 1) + x(1) (k) + β0 + β1 , 2 2
CE
and the estimates of model parameters are
where D2 = D e1
AC
152
βˆ0
(2) When N = 1, GPMB(1,1,N ) model yields the NGM(1,1) model [23] defined as
PT
147
AN US
141 142
e2
e3 .
ˆ= α κ ˆ
βˆ0
βˆ1
T
= D2T D2
−1
D2T Y
(3) When β1 = β2 = · · · = βN −1 = 0, GPMB(1,1,N ) model yields the GM(1,1,tα ) model [24] defined as x(0) (k) =
i α h (1) k N +1 − (k − 1)N +1 x (k − 1) + x(1) (k) + β0 + βN , 2 N +1
and the estimates of model parameters are
where D3 = D e1
e2
ˆ= α κ ˆ
eN +2 .
βˆ0
βˆN
T
8
= D3T D3
−1
D3T Y
ACCEPTED MANUSCRIPT
157
Remark 1 indicates that the existing GM(1,1), NGM(1,1) and GM(1,1,tα ) model are all special cases
158
of GPMB(1,1,N ) model with fixed background coefficient. It is convenient to obtain GM(1,1), NGM(1,1),
159
160
GM(1,1,tα ) and other novel models by combining polynomial order and optimization criterion. n o (1) (1) (1) (1) (1) Theorem 1. Assume XAF = xAF (1), xAF (2), · · · , xAF (n) , where xAF (k) = ρx(1) (k) + ξ, ρ 6= 0, to be
161
the affine transformation of the accumulating generation sequence X (1) , then the fitted and predicted values
162
of GPMB(1,1,N ) models respectively constructed by X (1) and XAF satisfy x ˆAF (k) = ρˆ x(1) (k) + ξ, k =
163
1, 2, · · · , n + p.
Proof. From the definitions of accumulating generation and affine transformation, it follows that (0) (1) (1) xAF (2) − xAF (1) xAF (2) ρx(0) (2) (0) (1) (1) (0) xAF (3) xAF (3) − xAF (2) ρx (3) = = YAF = = ρY .. .. .. . . . (0) (1) (1) (0) ρx (n) xAF (n) − xAF (n − 1) xAF (n) and
DAF
ρ λx(1) (1) + (1 − λ)x(1) (2) + ξ ρ λx(1) (2) + (1 − λ)x(1) (3) + ξ = .. . (1) ρ λx (n − 1) + (1 − λ)x(1) (n) + ξ
1 1 .. . 1
3 2 5 2
.. .
2n−1 2
··· ···
2N +1 −1 N +1 3N +1 −2N +1 N +1
··· ···
.. .
nN +1 −(n−1)N +1 N +1
where P and Q are two (N + 2)-order nonsingular matrices respectively defined as ρ 0 ··· 0 1 0 ··· 0 0 1 · · · 0 ξ 1 · · · 0 . P = . . . and Q = . . . . . . ... . . .. .. .. .. .. 0 0 ··· 1 0 0 ··· 1
= DP Q
(23)
(24)
169 170 171
172 173
PT
0 0 .. . .
(25)
1
Under the same optimization criterion, the weighted matrices always satisfy WAF = $W , where $ is a
CE
168
Because the coefficient ρ 6= 0, P is invertible and the inverses of P and Q are 1 0 ··· 0 1 0 ··· −ξ 1 ··· 0 ρ · · · 0 1 and Q−1 = . P −1 = . . . .. . . . . . .. ρ .. .. .. . . 0 0 ··· 0 0 ··· ρ constant only depending on ρ, for instance, $ equals
1 ρ
under the three above optimization criteria. Denoting 0 (1) 0 T , it ˆ AF = α the model parameters of GPMB(1,1,N ) model constructed by XAF as κ ˆ βˆ00 βˆ10 · · · βˆN follows that h −1 T −1 T 2 i T 2 2 ˆ AF = DAF ˆ . (26) κ WAF DAF DAF WAF YAF = ρQ−1 P −1 D T W 2 D D W Y = ρQ−1 P −1 κ
AC
167
ED
M
166
CR IP T
165
(1)
AN US
164
(1)
Substituting Eq. (25) into Eq. (26) gives the model parameters satisfying α ˆ0 = α ˆ , βˆ00 = ρβˆ0 − ξ α ˆ , βˆj0 =
ρβˆj , j = 2, · · · , N , and the polynomial coefficients satisfying −1 0 −ˆ α 1 ··· 0 −ξ α ˆ + ρβˆ0 γ0 ργ0 + ξ . γ10 ργ1 ρβˆ1 0 −ˆ α .. 0 γAF = . = = .. , . . .. .. .. .. . .. . N . 0 ˆ γN ργ N ρβN 0 0 · · · −ˆ α 9
(27)
ACCEPTED MANUSCRIPT
174
as well as the integration constant satisfying # ! " n N X e2αˆ − 1 X (1) α(k−2) ˆ i x (k) − = ρˆ c. cˆAF = ρ 2αn γi k e e ˆ −1 i=0
(28)
k=1
175
Substituting Eqs. (26), (27) and (28) into the time response sequence gives that " ! # n N N X X e2αˆ − 1 X (1) (1) i α(k−2) ˆ αk ˆ x ˆAF (k) = ρ 2αn x (k) − γ k e e + ρ γi k i + ξ = ρˆ x(1) (k) + ξ. i e ˆ −1 i=0 i=0
176
By the inverse affine transformation,
1 ρ
(1)
x ˆAF − ξ
(29)
CR IP T
k=1
= x ˆ(1) (k), that is to say, the fitted and predicted
177
values are invariant whether the affine transformation is applied to the accumulating generation sequence
178
or not.
179
Furthermore, if the translation coefficient ξ = 0, the affine transformation of the accumulating generation sequence is equivalent to the multiple transformation of the original sequence. From Theorem 1, the multiple transformation of the original sequence has no influence on modeling performance. Therefore, a suitable ρ
182
can be selected to reduce the condition number in Eq. (11) without change of accuracy [35].
183
4. Numerical example
AN US
180 181
In this section, the GPMB(1,1,N ) model is build on the original sequence shown in Table 1, and the
185
conventional GPM(1,1,N ) model [26] is also constructed to illustrate the improvements. The modeling
186
results are also compared with the commonly used prediction models: NDGM(1,1) [28], Verhulst [36] and
187
CPR (Cubic Polynomial Regression) [33].
ED
M
184
Table 1: The actual values of original sequence.
189
2 5.39 13 4.35
3 4.61 14 5.41
4 4.36 15 7.05
5 3.92 16 9.10
6 3.36 17 11.64
7 2.87 18 14.94
8 2.84 19 19.10
9 2.97 20 24.29
10 2.95 21 30.71
The original sequence is divided into two parts: the first part accounting for almost eighty percent of the original sequence (the first 16 elements) is used to build model and denoted as X , and the rest is used
to validate extrapolation ability and denoted as Y.
191
4.1. Polynomial order selection, background coefficient search and parameter estimation
193 194
AC
190
192
11 3.03
CE
188
1 6.03 12 3.72
PT
Index Actual value Index Actual value
(0)
According to Algorithm 1, the modeling sequence X is divided into the training sequence Xtrain = (0) {6.03, 5.39, 4.61, 4.36, 3.92, 3.36, 2.87, 2.84, 2.97, 2.95, 3.03, 3.72}, and the testing sequence Xtest = {4.35, 5.41, 7.05, 9.10}.
195
Under the least squared error criterion (i.e., wk = 1) and least square percentage error criterion (i.e., wk =
196
1/x(0) (k)), the fitting error MAPEtrain and predicting error MAPEtest versus the increasing of background
197
coefficient for GPMB(1,1,N ) model with N = 0, 1, 2, 3 are displayed in Figure 2.
198 199
It can be seen in Figure 2 that the GPMB(1,1,N ) models under the least squared error criterion and the least squared percentage error criterion has the similar trend. 10
ACCEPTED MANUSCRIPT
9
(b)
0
0.6 0.2 0.4 0.8 background coefficient N=0
13
51.5
8
51
6
50.5
60 1
4
62
14
0
0.2 0.4 0.6 0.8 background coefficient N=1
MAPE
train
(%)
12 61
12
60
11
0
0.2 0.4 0.6 0.8 background coefficient
59 1
6
8
4
4
2
52
10
0
6
50.5
4
4
0
0.2 0.4 0.6 0.8 background coefficient
50 1
2
0 1
16 12
8
51
0.2 0.4 0.6 0.8 background coefficient N=2
8
8 6
10
12
50 1
51.5
10
8
4
0
0.4 0.2 0.6 0.8 background coefficient
0 1
150
10
100
5
50
(%)
10
N=3
15
test
52
16
0
0
0.2 0.4 0.6 0.8 background coefficient N=3
12
0 1
200
10
160
8
120
6
80
4
40
2
0
0.8 0.2 0.4 0.6 background coefficient
MAPE
61
12
N=2
10
(%)
MAPE
10
52.5
test
62
N=1
14
MAPE
11
train
(%)
63
CR IP T
N=0
12
(a)
0 1
AN US
Figure 2: Trend of the MAPEtrain and MAPEtest with the increasing of background coefficient for GPMB(1,1,N ) model under (a) the least squared error criterion and (b) the least squared percentage error criterion.
Under the least squared error criterion, the MAPEtrain of GPMB(1,1,0) model reaches the minimum
201
9.63% with λopt = 0.45 while its MAPEtest keeps decreasing; the MAPEtrain of GPMB(1,1,1) model reaches
202
the minimum 5.78% with λopt = 0.51 while its MAPEtest keeps increasing; the MAPEtrain and MAPEtest of
203
the GPMB(1,1,2) and GPMB(1,1,3) model both increase after their decreasing, and their MAPEtrain s reach
204
the minimums 2.79% and 2.70% respective with λopt = 0.51 and λopt = 0.49, meanwhile their corresponding
205
MAPEtest s are 1.25% and 4.89%.
M
200
Under the least squared percentage error criterion, both MAPEtrain and MAPEtest of GPMB(1,1,0) model
207
keep decreasing with the increasing of background coefficient; the MAPEtrain of GPMB(1,1,1) model reaches
208
the minimum 5.71% with λopt = 0.52 while its MAPEtest keeps increasing; the MAPEtrain and MAPEtest
209
of the GPMB(1,1,2) and GPMB(1,1,3) model both increase after their decreasing, and their MAPEtrain s
210
reach the minimums 2.70% and 2.64% with λopt = 0.50 and λopt = 0.53, respectively, meanwhile their
211
corresponding MAPEtest s are 3.35% and 7.58%.
PT
ED
206
To sum up, whichever optimization criterion is chosen, the GPMB(1,1,0) and GPMB(1,1,1) model are
213
underfitted while the GPMB(1,1,3) model is overfitted, and only the GPMB(1,1,2) model having minimal
215
216 217 218 219
predicting error MAPEtest is the best. 4.2. Numerical results and comparisons with other models According to the computational steps in Section 2.3, GPMB(1,1,2) models under the two above criteria
AC
214
CE
212
are both built on the sequence X , and the results are summarized in Table 2. In addition, several affine
transformations with different coefficients are applied to the accumulating sequence to further verify the conclusions in Theorem 1.
220
It can be seen in Table 2 that the GPMB(1,1,2) model under the least squared percentage error criterion
221
outperforms that under the least squared error criterion due to that the evaluation measure MAPE is in
222 223 224
step with the optimization criterion. For the sake of comparisons, the models below are all build on the sequence X under the least squared error criterion. The difference stepwise ratio criterion in reference [26] indicates that GPM(1,1,0) is the best 11
ACCEPTED MANUSCRIPT
Table 2: Parameter values and errors of GPMB(1,1,2) model under different optimization criteria.
Least squared error Least squared percentage error
ρ 1.0 0.1 0.1 1.0 0.1 0.1
ξ 0.0 0.0 -0.6 0.0 0.0 -0.6
λopt 0.51 0.51 0.51 0.50 0.50 0.50
α ˆ 0.2020 0.2020 0.2020 0.1896 0.1896 0.1896
βˆ0 6.2129 0.62129 0.74248 6.2795 0.62795 0.74173
βˆ2 0.0769 0.00769 0.00769 0.0772 0.00772 0.00772
βˆ1 -1.8868 -0.18868 -0.18868 -1.8458 -0.18458 -0.18458
γ0 -3.1708 -0.31708 -0.91708 -4.4153 -0.44153 -1.04153
γ1 5.5725 0.55725 0.55725 5.4422 0.54422 0.54422
γ2 -0.3807 -0.03807 -0.03807 -0.4069 -0.04069 -0.04069
cˆ 3.2902 0.32902 0.32902 4.4918 0.44918 0.44918
Errors (%) MAPEX =2.45 MAPEY =1.75 MAPEX =2.46 MAPEY =1.19
CR IP T
Criteria
225
model, and the fitted and predicted values are listed in Table 3. The results of NDGM(1,1), Verhulst and
226
CPR are also obtained in Table 3.
Table 3: Fitted and predicted values of GPMB(1,1,2) and other models. GPM(1,1,0)
6.55 6.92 7.30 7.71 8.14
NDGM(1,1) Values 6.03 5.11 4.63 4.29 4.04 3.86 3.73 3.63 3.56 3.51 3.48 3.45 3.43 3.42 3.41 3.40
APE(%) 0.00 5.24 0.49 1.64 3.03 14.82 29.85 27.87 19.96 19.07 14.73 7.25 21.12 36.83 51.67 62.64 19.76 70.84 77.30 82.26 86.06 88.98 81.09
Verhulst
Values 6.03 0.77 0.87 0.97 1.10 1.23 1.39 1.56 1.75 1.96 2.20 2.46 2.75 3.07 3.43 3.82
APE(%) 0.00 85.73 81.21 77.64 72.02 63.30 51.71 45.18 41.16 33.55 27.49 33.87 36.74 43.17 51.34 58.00 50.13 63.47 68.40 72.60 76.18 79.21 71.97
AN US
APE(%) 12.22 46.17 33.56 25.83 12.90 7.29 32.62 41.50 42.85 51.84 56.08 34.23 21.19 2.88 16.64 31.82 29.35 43.72 53.71 61.77 68.26 73.49 60.19
M
Values 6.77 2.90 3.06 3.23 3.41 3.60 3.81 4.02 4.24 4.48 4.73 4.99 5.27 5.57 5.88 6.20
ED
11.66 15.07 19.42 24.91 31.80
APE(%) 0.29 1.08 3.52 2.35 3.10 1.40 7.95 1.78 5.49 2.54 3.26 2.97 0.43 1.07 0.94 1.00 2.45 0.15 0.87 1.66 2.54 3.53 1.75
3.39 3.39 3.39 3.39 3.38
4.25 4.72 5.23 5.79 6.38
CPR
Values 5.89 5.40 4.87 4.34 3.84 3.40 3.05 2.81 2.74 2.84 3.16 3.73 4.57 5.72 7.21 9.08 11.34 14.04 17.20 20.86 25.05
APE(%) 2.36 0.17 5.40 0.39 2.05 1.12 5.76 0.93 8.59 3.87 4.12 0.17 4.81 5.45 2.27 0.25 2.98 2.62 6.40 11.02 16.43 22.60 11.81
It can be seen in Table 3 that on the modeling sequence X , the MAPE of GPMB(1,1,2) model is close
CE
227
Values 6.05 5.33 4.77 4.26 3.80 3.41 3.10 2.89 2.81 2.88 3.13 3.61 4.37 5.47 6.98 9.01
PT
Actual values 1 6.03 2 5.39 3 4.61 4 4.36 5 3.92 6 3.36 7 2.87 8 2.84 9 2.97 10 2.95 11 3.03 12 3.72 13 4.35 14 5.41 15 7.05 16 9.10 MAPE (%) 17 11.64 18 14.94 19 19.10 20 24.29 21 30.71 MAPE (%)
GPMB(1,1,2)
Index
228
to that of CPR model, and much smaller than that of GPM(1,1,0), NDGM(1,1) and Verhulst model; on
229
the validating sequence Y, the MAPE of CPR model is more than 6 times larger than that of GPMB(1,1,2)
231
232
model which is also the smallest among these models. So GPMB(1,1,2) model can catch the overall trend
AC
230
accurately and has the best performance in this example. 5. Application
233
The annual per capita electricity consumption (ApCEC) of China is collected from the official website
234
National Bureau of Statistics of China (http://data.stats.gov.cn/english/easyquery.htm?cn=C01), as shown
235
in Table 4. The sequence from 2000 to 2010 is used to build the models, and that from 2011 to 2015 is used
236
to validate accuracy.
12
ACCEPTED MANUSCRIPT
Table 4: Annual per capita electricity consumption (kilowatt hour) of China form 2000 to 2015. Year ApCEC Year ApCEC
2000 1066.9 2008 2607.6
2001 1157.6 2009 2781.7
2002 1286.0 2010 3134.8
2003 1477.1 2011 3497.0
2004 1695.2 2012 3684.2
2005 1913.0 2013 3993.0
2006 2180.6 2014 4132.9
2007 2482.2 2015 4231.0
Under the least squared percentage error criterion, the polynomial order and background coefficient are
238
respectively selected as Nopt = 2 and λopt = 0.49 according to Algorithm 1. In particular, there only exists
239
slight difference between GPMB(1,1,2) model (MAPEtrain = 0.39%, MAPEtest = 9.13%) and GPMB(1,1,3)
240
model (MAPEtrain = 0.35%, MAPEtest = 9.83%). But GPMB(1,1,3) is an overfitted model with good
241
fitting but bad predicting accuracy. In addition, the results of GPMB(1,1,2) model are also compared to the
242
GPM(1,1,N ), NDGM(1,1), Verhulst and CPR model. The fitted and predicted values are listed in Table 5
243
and plotted in Figure 3.
CR IP T
237
GPMB(1,1,2)
GPM(1,1,0)
NDGM(1,1)
Verhulst
CPR
ED
M
Values APE(%) Values APE(%) Values APE(%) Values APE(%) Values APE(%) 1068.3 0.13 1040.2 2.50 1066.9 0.00 1066.9 0.00 1055.6 1.06 1149.7 0.69 1228.3 6.11 1147.2 0.90 1195.2 3.25 1158.7 0.10 1292.6 0.51 1365.7 6.20 1315.7 2.31 1357.8 5.58 1304.8 1.46 1484.7 0.51 1518.4 2.80 1496.4 1.31 1525.2 3.25 1485.8 0.59 1700.4 0.31 1688.2 0.41 1690.1 0.30 1703.9 0.51 1694.2 0.06 1927.6 0.76 1877.1 1.88 1897.8 0.79 1897.3 0.82 1922.0 0.47 2160.2 0.94 2087.0 4.29 2120.5 2.75 2108.2 3.32 2161.5 0.88 2395.4 3.50 2320.4 6.52 2359.4 4.95 2339.0 5.77 2404.8 3.12 2631.9 0.93 2580.0 1.06 2615.4 0.30 2592.1 0.59 2644.2 1.40 2869.0 3.14 2868.5 3.12 2890.0 3.89 2870.1 3.18 2871.8 3.24 3106.4 0.91 3189.3 1.74 3184.4 1.58 3175.7 1.31 3079.9 1.75 1.12 3.33 1.73 2.51 1.28 3343.9 4.38 3546.0 1.40 3500.0 0.09 3512.0 0.43 3260.6 6.76 3581.5 2.79 3942.7 7.02 3838.5 4.19 3882.0 5.37 3406.1 7.55 3819.1 4.36 4383.6 9.78 4201.4 5.22 4289.4 7.42 3508.7 12.13 4056.7 1.84 4873.9 17.93 4590.5 11.07 4738.1 14.64 3560.5 13.85 4294.4 1.50 5419.0 28.08 5007.7 18.36 5232.2 23.66 3553.7 16.01 2.97 12.84 7.78 10.31 11.26
PT
Year Actual values 2000 1066.9 2001 1157.6 2002 1286.0 2003 1477.1 2004 1695.2 2005 1913.0 2006 2180.6 2007 2482.2 2008 2607.6 2009 2781.7 2010 3134.8 MAPE(%) 2011 3497.0 2012 3684.2 2013 3993.0 2014 4132.9 2015 4231.0 MAPE(%)
AN US
Table 5: Fitted and predicted values of annual per capita electricity consumption of China form 2000 to 2015.
It can be seen in Table 5 that the MAPEs on the fitting sequence (from 2000 to 2010) of these models are all less than 5%, which indicates that they have almost the same fitting performance. But these models
246
differ form each other in the MAPEs on the predicting sequence (from 2011 to 2015). It is shown in Figure
247
3 that the annual per capita electricity consumption from 2011 to 2015 are overestimated by GPM(1,1,0),
248
NDGM(1,1) and Verhulst model, and on the contrary they are underestimated by CPR model. Only
250
251
AC
249
CE
244 245
GPMB(1,1,2) model presents an accurate estimation of the annual per capita electricity consumption and thus outperforms the other four models in this case study. 6. Conclusion
252
Different form GM(1,1) model (or NGM(1,1) model) based on the hypothesis that the original sequence
253
is in accord with homogeneous (or non-homogeneous) exponential trend, GPMB(1,1,N ) model is proposed
254
for a more general sequence with including but not limited to homogeneous and non-homogeneous trend.
255
In this paper, the main contributions are briefly summarized as follows: 13
6000 5500 5000 4500 4000
Actual values GMPB(1,1,2) GMP(1,1,0) NDGM(1,1) Verhulst CPR
Sequence for buliding models ←
3500 3000 2500 2000 1500 1000 2000 2001 2002
2003 2004
2005 2006
2007 2008 Year
2009 2010
CR IP T
Fitted and predicted values of ApCEC (kilowatt hour)
ACCEPTED MANUSCRIPT
2011 2012
2013 2014
2015
AN US
Figure 3: Fitted and predicted values from various models.
256
1. The most commonly used three optimization criteria, i.e., the least squared error criterion, the least
257
squared percentage error criterion and the least absolute percentage error criterion in grey prediction
258
model, are uniformly represented by weighted least square method.
2. The algorithm for polynomial order selection, background search and nonlinear parameter estimation of GPMB(1,1,N ) model is proposed, and the numerical example and application indicate that it is
261
more robust than the conventional difference stepwise ratio criterion in reference [26].
M
259 260
3. The mathematical analysis indicates that GM(1,1), NGM(1,1) and GM(1,1,tα ) model are all special
263
forms of GPMB(1,1,N ) model, and that the multiple transformation for original sequence and the affine
264
transformation for accumulating sequence are proved to be independent of modeling performance.
ED
262
Although the numerical example and application have verified the effectiveness of GPMB(1,1,N ) model,
266
large-scale simulation experiments under different signal-to-noise ratios still need to be designed to quantify
267
the influence of noise level and sample size on the modeling performance in the future work.
268
other hand, it should be also noticed that the model parameters in Eq. (11) are estimated based on the
269
difference equation (9) while the fitted and predicted values are obtained according to the analytic solution
270
to whitening equation (4). The skip between these two equations may introduce new errors [27]. Therefore,
271
the feature work can be focused on constructing a model based on Eq. (3) directly
273
x(0) (k) = ax(1) (k − 1) + b0 + b1 k + · · · + bN k N ,
where the polynomial feature and the model parameters can be together obtained by using the `1 -regularized least square method [37], arg min a,b0 ,··· ,bN
274
On the
CE
AC
272
PT
265
n X
k=2
x
(0)
(k) − x ˆ
(0)
! N 2 X (k) + ϑ |a| + |bi | i=0
and ϑ > 0 is the regularized parameter. 14
(30)
ACCEPTED MANUSCRIPT
275
Acknowledgement Thank the editors and the anonymous referees for their insightful comments to improve the paper. This
277
work is supported by National Natural Science Foundation of China under grant 71671090, Aeronautical
278
Science Foundation of China under grant 2016ZG52068 and Qinglan Project for excellent youth or middle-
279
aged academic leaders in Jiangsu Province, China.
280
References
281
References
289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331
AN US
288
M
287
ED
285 286
PT
284
[1] J. L. Deng, Control problems of grey systems, Systems & Control Letters 1 (1982) 288–294. [2] M. S. Yin, Fifteen years of grey system theory research: A historical review and bibliometric analysis, Expert Systems with Applications 40 (2013) 2767–2775. [3] S. F. Liu, Y. J. Yang, N. M. Xie, J. Forrest, New progress of grey system theory in the new millennium, Grey Systems: Theory & Application 6 (2016) 2–31. [4] N. M. Xie, R. Z. Wang, A historic review of grey forecasting models, Journal of Grey System 29 (2017) 1–29. [5] X. Li, Y. Tao, Y. Zheng, Model GM(1,1,β) and its applicable region, Grey Systems: Theory & Application 3 (2013) 266–275. [6] D. C. Li, C. Yeh, C. J. Chang, An improved grey-based approach for early manufacturing data forecasting, Computers & Industrial Engineering 57 (2009) 1161–1167. [7] D. C. Li, C. J. Chang, C. C. Chen, W. C. Chen, Forecasting short-term electricity consumption using the adaptive grey-based approachan asian case, Omega 40 (2012) 767–773. [8] Z. Zhao, J. Wang, J. Zhao, Z. Su, Using a grey model optimized by differential evolution algorithm to forecast the per capita annual net income of rural households in china, Omega 40 (2012) 525–532. [9] Y. G. Dang, S. F. Liu, K. J. Chen, The GM models that x(n) be taken as initial value, Kybernetes 33 (2004) 247–254. [10] T. L. Tien, A new grey prediction model FGM(1, 1), Mathematical & Computer Modelling 49 (2009) 1416–1426. [11] Y. H. Wang, Y. G. Dang, Y. Q. Li, S. F. Liu, An approach to increase prediction precision of GM(1,1) model based on optimization of the initial condition, Expert Systems with Applications 37 (2010) 5640–5644. [12] J. Xu, T. Tan, M. Tu, L. Qi, Improvement of grey models by least squares, Expert Systems with Applications 38 (2011) 13961–13966. [13] Y. H. Wang, Q. Liu, J. R. Tang, C. W. Bin, L. X. Zhong, Optimization approach of background value and initial item for improving prediction precision of GM(1,1) model, Journal of Systems Engineering and Electronics 25 (2014) 77–82. [14] L. F. Wu, S. F. Liu, L. G. Yao, S. L. Yan, D. L. Liu, Grey system model with the fractional order accumulation, Communications in Nonlinear Science & Numerical Simulation 18 (2013) 1775–1785. [15] L. F. Wu, S. F. Liu, Y. J. Yang, A gray model with a time varying weighted generating operator, IEEE Transactions on Systems, Man & Cybernetics: Systems 46 (2016) 427–433. [16] C. I. Chen, S. J. Huang, The necessary and sufficient condition for GM(1,1) grey prediction model, Applied Mathematics & Computation 219 (2013) 6152–6162. [17] J. Liu, X. P. Xiao, J. H. Guo, S. H. Mao, Error and its upper bound estimation between the solutions of GM(1,1) grey forecasting models, Applied Mathematics & Computation 246 (2014) 648–660. [18] X. P. Xiao, H. Guo, S. H. Mao, The modeling mechanism, extension and optimization of grey GM(1,1) model, Applied Mathematical Modelling 38 (2014) 1896–1910. [19] T. X. Yao, S. F. Liu, N. M. Xie, On the properties of small sample of GM(1,1) model, Applied Mathematical Modelling 33 (2009) 1894–1903. [20] L. F. Wu, S. F. Liu, L. G. Yao, S. L. Yan, The effect of sample size on the grey system model, Applied Mathematical Modelling 37 (2013) 6577–6583. [21] M. Tabaszewski, C. Cempel, Using a set of GM(1,1) models to predict values of diagnostic symptoms, Mechanical Systems & Signal Processing s5253 (2015) 416–425. [22] R. Guo, D. Guo, Random fuzzy variable foundation for grey differential equation modeling, Soft Computing 13 (2009) 185–201. [23] J. Cui, S. F. Liu, B. Zeng, N. M. Xie, A novel grey forecasting model and its optimization, Applied Mathematical Modelling 37 (2013) 4399–4406. [24] W. Y. Qian, Y. G. Dang, S. F. Liu, Grey GM(1,1,tα ) model with time power and its application, Systems Engineering — Theory & Practice 32 (2012) 2247–2252. [25] X. Ma, Y. S. Hu, Z. B. Liu, A novel kernel regularized nonhomogeneous grey model and its applications, Communications in Nonlinear Science & Numerical Simulation 48 (2016) 51–62. [26] D. Luo, B. L. Wei, Grey forecasting model with polynomial term and its optimization, Journal of Grey System 29 (2017) 58–69. [27] N. M. Xie, S. F. Liu, Discrete grey forecasting model and its optimization, Applied Mathematical Modelling 33 (2009) 1173–1186.
CE
283
AC
282
CR IP T
276
15
ACCEPTED MANUSCRIPT
339 340 341 342 343 344 345 346 347 348 349 350
CR IP T
338
AN US
337
M
336
ED
335
PT
334
[28] N. M. Xie, S. F. Liu, Y. J. Yang, C. Q. Yuan, On novel grey forecasting model based on non-homogeneous index sequence, Applied Mathematical Modelling 37 (2013) 5059–5068. [29] B. Zeng, C. M. Luo, S. F. Liu, Y. Bai, C. Li, Development of an optimization method for the GM(1,N ) model, Engineering Applications of Artificial Intelligence 55 (2016) 353–362. [30] B. Zeng, C. M. Luo, S. F. Liu, C. Li, A novel multi-variable grey forecasting model and its application in forecasting the amount of motor vehicles in Beijing, Computers & Industrial Engineering 101 (2016) 479–489. [31] X. Ma, Z. B. Liu, The kernel-based nonlinear multivariate grey model, Applied Mathematical Modelling 56 (2018) 217–238. [32] B. Zeng, C. Li, Improved multi-variable grey forecasting model with a dynamic background-value coefficient and its application, Computers & Industrial Engineering 118 (2018) 278–290. [33] C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics), Springer-Verlag New York, Inc., 2006. [34] N. Jorge, J. W. Stephen, Numerical Optimization, Springer-Verlag New York, Inc., 2006. [35] K. Jing, Y. Z. Liu, Morbidity problem of GM(1,1) model, Control & Decision 31 (2016) 869–874. [36] C. I. Chen, H. L. Chen, S. P. Chen, Forecasting of foreign exchange rates of taiwan’s major trading partners by novel nonlinear Grey Bernoulli model NGBM(1,1), Communications in Nonlinear Science & Numerical Simulation 13 (2008) 1194–1204. [37] K. Fountoulakis, J. Gondzio, Performance of first- and second-order methods for `1 -regularized least squares problems, Computational Optimization & Applications 65 (2016) 605–635.
CE
333
AC
332
16