Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange

Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange

INS 10870 No. of Pages 17, Model 3G 19 May 2014 Information Sciences xxx (2014) xxx–xxx 1 Contents lists available at ScienceDirect Information Sc...

4MB Sizes 4 Downloads 101 Views

INS 10870

No. of Pages 17, Model 3G

19 May 2014 Information Sciences xxx (2014) xxx–xxx 1

Contents lists available at ScienceDirect

Information Sciences journal homepage: www.elsevier.com/locate/ins 6 7

Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange

3 4 5 8

Q1

9

Martha Pulido, Patricia Melin ⇑, Oscar Castillo Tijuana Institute of Technology, Division of Graduate Stusies and Research, Calzada Tecnologico s/n, Fracc. Tomas Aquino, 22379 Tijuana, BC, Mexico

10

a r t i c l e

1 3 2 2 13 14 15 16 17

i n f o

Article history: Received 23 October 2013 Received in revised form 27 March 2014 Accepted 3 May 2014 Available online xxxx

18 19 20 21 22

Keywords: Ensemble neural network Particle swarm optimization Time series prediction

a b s t r a c t This paper describes a hybrid method based on particle swarm optimization for designing ensemble neural networks with fuzzy aggregation of responses to forecast complex time series. The time series that was considered in this paper, to compare the hybrid approach with traditional methods, is the Mexican Stock Exchange, and the results shown are for the optimization of the structure of the ensemble neural network with type-1 and type-2 fuzzy logic integration. Simulation results show that the optimized ensemble neural network approach produces good prediction of the Mexican Stock Exchange. Ó 2014 Published by Elsevier Inc.

24 25 26 27 28 29 30 31 32 33

34 35

1. Introduction

36

A time series is a sequence of data points, measured typically at successive points in time and spaced at uniform time intervals [2,11]. Examples of time series are the daily closing value of the Mexican Stock Exchange. Time series are very frequently plotted via line charts. Time series are used in statistics, signal processing, pattern recognition, econometrics, mathematical finance, weather forecasting, earthquake prediction, electroencephalography, control engineering, astronomy and communications engineering. Stock fund managers and financial analysts attempt to predict price activity in the stock market on the basis of either their professional knowledge or stock analyzing tools, in order to obtain gains in prices in the stock market. For more than one decade, time series models have also been applied to solve various domain problems, such as financial forecasting. In this paper the design of a neural network ensemble with the Particle Swarm Optimization algorithm (PSO) and its integration with type-1 and type-2 fuzzy systems, along with simulation results for the prediction of the Mexican Stock Exchange are presented. In the literature there has been a lot of recent work in time series prediction [3,4,7–10,20,22,26–28,37,42,46], which indicate the recent importance of the topic. Time series prediction with hybrid models has also been a hot topic in recent years [40,41,43]. This paper shows the results of an optimized ensemble neural network and its fuzzy response aggregation for predicting the time series of the Mexican Stock Exchange [1]. The PSO technique is used to optimize the neural network ensemble, as this can help in finding a good solution to a complex problem, which in this case is to find the best ensemble neural network architecture to obtain

37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52

Q2

⇑ Corresponding author. Tel.: +52 6646236318. E-mail address: [email protected] (P. Melin). http://dx.doi.org/10.1016/j.ins.2014.05.006 0020-0255/Ó 2014 Published by Elsevier Inc.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 2

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

81

the minimum forecast error for the time series mentioned above and to integrate the results with type-1 and type-2 fuzzy systems. The main contribution of the paper is the proposed model of a neural network ensemble with type-2 fuzzy logic for response integration. In addition the particle swarm optimization method determines the number of modules of the neural network ensemble, number of layers and number of neurons per layer, and thus obtains the best architecture of the ensemble neural network. After obtaining this architecture the results from the modules are aggregated with type-1 and type-2 fuzzy systems, the inputs to the fuzzy system are the responses according to the number of network modules and this are the number of inputs of the fuzzy system. In this case, the maximum number of inputs is of 5 inputs and one output with two Gaussian membership functions and these are be granulated in two linguistic variables that are low and high forecast, and the forecast output will also be low or high and thereby obtain the forecast error for this series of the Mexican Stock Exchange. The proposed hybrid ensemble of neural networks with fuzzy response aggregation and its optimization with PSO is main contribution of the paper, as this hybrid approach has not been proposed previously in the literature for this kind of time series prediction problems. This paper considers the times series of the Mexican Stock Exchange because the problem is quite complex and the time series is chaotic in some periods of time and for this reason an ensemble model is justified. To consider different possible behaviors we have also previously worked with other time series, such as the Mackey–Glass s ¼ 17; 34 and 68 and this series has chaotic behavior and as the value of s increases it will be more chaotic. Also the US/Dollar/MX Peso and the Dow Jones time series, which are quite complex, have been considered with ensembles and have produced good results in these cases [30]. We can find that there are similar ensemble approaches in the literature, but the main difference is that in the proposed approach in this paper we use type-2 fuzzy aggregation of responses in the ensemble, which is not found into other similar works. The idea of using type-2 fuzzy logic is the better manage the uncertainty in prediction. In addition, we also use PSO to optimize ensemble neural network, which is not so commonly used in ensembles. We consider this as the new contribution, since the results obtained are better than other methods proposed in the literature. The rest of the paper is organized as follows: Section 2 describes the architecture of the ensemble neural network, basic concepts of the proposed method, Section 3 describes the concepts particle swarm optimization, In Section 4 Optimization Ensemble Neural Network with Particle Swarm Optimization. Section 5 describes the concepts of fuzzy systems as methods of integration. In Section 6 the methodology for solving it are represented, Section 7 describes the simulation results of the proposed method, in Section 8 the comparison and the t student test are presented, and Section 9 shows the conclusions.

82

2. Time series and prediction

83

103

Time series are analyzed to understand the past and to predict the future, enabling managers or policy makers to make properly informed decisions. A time series analysis quantifies the main features in data and the random variation. These facts, combined with improved computing power, have made time series methods widely applicable in government, industry, and commerce. Time series models often form the basis of computer simulations. Some examples are assessing different strategies for control of inventory using a simulated time series of demand; comparing designs of wave power devices using a simulated series of sea states; and simulating daily rainfall to investigate the long-term environmental effects of proposed water management policies A time-series is defined as a sequence of observations on a set of values that takes a variable (quantitative) at different points in time. Time series are widely used today because organizations need to know the future behavior of certain relevant phenomena in order to plan, prevent, and so on, their actions. That is, to predict what will happen with a variable in the future from the behavior of that variable in the past [1]. The data can behave in different ways over time, this may be a trend, which is the component that represents a long-term growth or decline over a period of time. A cycle is also possible, which refers to the wave motion that occurs around the trend, or may not have a defined or random manner; there are seasonal variations (annual, biannual, etc.), which are a behavior patterns that are repeated year after year at a particular time [2–4]. The word ‘‘prediction’’ comes from the Latin prognosticum, which means I know in advance. Prediction is to issue a statement about what is likely to happen in the future, based on analysis and considerations of experiments. Making a forecast is to obtain knowledge about uncertain events that are important in decision-making [5]. Time series prediction tries to predict the future based on past data, it take a series of real data xt  n, . . . xt  2.0 xt  1, xt and then obtains the prediction of the data xt + 1, xt + 2 . . . xn + n. The goal of time series prediction or its model is to observe the series of real data, so that future data may be accurately predicted [29].

104

3. Ensemble neural network

105

Artificial neural networks are inspired by the biological nervous system architecture, which consists of a large number of relatively simple neurons that work in parallel to facilitate rapid decision making [32]. A neural network is a system of parallel processors interconnected as a directed graph. Schematically each processing element (neurons) in the network is represented as a node. Its most important advantage is to solve problems that are

53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80

84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102

106 107 108

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

3

115

too complex for conventional technologies, problems that have no solution algorithm and its solution algorithm is very difficult finding. Artificial neural networks are formed by a large number of neurons; these are not called artificial neurons but output nodes or units. Ensemble models are formed by a collection of neural networks with the idea of combining multiple decisions and achieving a better result. These ensemble models behave remarkably well, recently it has become a very hot topic in both the neural networks and machine learning communities [21], and has already been successfully applied to diverse areas such as face recognition [35,19], optical character recognition [13,18,28], scientific image analysis [40], medical diagnosis [13,46], seismic signals classification [36], etc.

116

3.1. Training of the neural networks

117

In the literature there are many methods for training neural networks like, the gradient descendent method, Levenberg– Marquardt and others. The Levenberg–Marquardt method can handle well ill conditioned matrices JTJ by using the following equation:

109 110 111 112 113 114

118 119

120

122 123 124 125

126 128

 1 hnext ¼ hnow  J T J þ kI gh

ð1Þ

where k is some nonnegative value and g h  12 g for simplicity.For instance, we consider a trivial nonlinear model f(xk  h) with two adjustable parameters p and q; i.e., h = (p, q)T. The objective is to find the optimal h⁄ = (1, 10) that minimizes the following sum of squared errors.

EðhÞ ¼

10 X 2 ðt k f ðxk ; hÞÞ

ð2Þ

k¼1

131

where tk is the desired output E is the objective function, and the Levenberg–Marquardt direction that is determined by using Eq. (1) when k ¼ 0:07 or any other particular value. Clearly, the Levenberg–Marquardt direction (in highlighted arrow) is an intermediate between the Gauss Newton direction ðk ! 0Þ and the steepest descendent direction ðk ! 1Þ.

132

4. Particle swarm optimization

133

Particle Swarm Optimization (PSO) is a bio-inspired optimization method proposed by Eberhart and Kennedy [14,25] in 1995. PSO is a search algorithm based on the behavior of biological communities that exhibits individual and social behavior [15,16], and examples of these communities are groups of birds, schools of fish and swarms of bees [23,24,44]. A PSO algorithm maintains a swarm of particles, where each particle represents a potential solution. In analogy with the paradigms of evolutionary computation, a swarm is similar to a population, while a particle is similar to an individual. In simple terms, the particles are ‘‘flown’’ through a multidimensional search space, where the position of each particle is adjusted according to its own experience and that of its neighbors. Let xi denote the position i in the search space at time step t; unless otherwise stated, t denotes discrete time steps. The position of the particle is changed by adding a velocity, mt(t), to the current position, i.e.

129 130

134 135 136 137 138 139 140

141

142 144 145 146 147 148

149 151

152 153 154 155

156 158

159 160 161

162 164

165 166 167

168 170

xi ðt þ 1Þ ¼ xi ðtÞ þ v i ðt þ 1Þ

ð3Þ

with xi(0)  U(Xmin, Xmax). It is the velocity vector the one that drives the optimization process, and reflects both the experimental knowledge of the particles and the information exchanged in the vicinity of particles. For gbest PSO, the particle velocity is calculated as:

v ij ðt þ 1Þ ¼ v ij ðtÞ þ c1 r1 ½yij ðtÞ  xij ðtÞ; þc2 r2 ðtÞ½y^j ðtÞ  xij ðtÞ

ð4Þ

where vij(t) is the velocity of the particle i in dimension j at time step t; c1 and c2 are the positive acceleration constants used to scale the contribution of cognitive and social skills, respectively, yr1j(t), yr2j(t)  U(0, 1) are random values in the range [0, 1]. The best personal position in the next time step t + 1 is calculated as:

 yi ðt þ 1Þ ¼

yi ðtÞ

if f ðxi ðxi ðt þ 1ÞÞ P f yi ðtÞÞ

xi ðt þ 1Þ if f ðxi ðxi ðt þ 1ÞÞ > f yi ðtÞÞ

ð5Þ

where f : Rnx ! R is the fitness function, as with EAs, the goal is measuring the fitness function closely corresponding to the optimal solution, for example the objective function quantifies the performance, or the quality of a particle (or solution). The overall best position, y_ (i) (t) at time step t, is defined as:

^ðtÞfyo ðtÞ; . . . ; ys ðtÞgf ðyðtÞÞ ¼ min ff ðyo ðtÞÞ; . . . f ðys ðtÞÞ;g y

ð6Þ

^ the where ns is the total number of particles in the swarm, More importantly, the above equation defining and establishing y best position is uncovered by either of the particles so far as this is usually calculated as the best personal best position. The global best position can also be selected from the particles of the current swarm, in which case it expressed as:

^ðtÞ ¼ min ff ðxo ðtÞÞ; . . . f ðxns ðtÞÞ;g y

ð7Þ

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 4

172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192

193 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210

5. Fuzzy systems According to [23,24,44], fuzzy logic was conceived by Lotfi A. Zadeh in 1965, based on a fuzzy set that differs from the traditional sets considering now that the membership degree is define by a function that represents a membership, which evaluates the input, and according to some predefined rules, assigns the membership degree to a fuzzy set. Fuzzy Set: Let X be a space of objects and x be a generic element of X. A classical set A, A # X, is defined as a collection of elements or objects x e X, such that each x can either belong or not belong to the set A. By defining a characteristic function for each element x in X, we can represent a classical set A by a set of ordered pairs (x, 0) or (x, 1) which indicates x R A or xA, respectively. Unlike the aforementioned conventional set, a fuzzy set expresses the degree to which an element belongs to a set. Hence the characteristic function of a fuzzy set is allowed to have values between 0 and 1, which denotes the degree of membership of an element in a given set. The basic structure of a fuzzy inference system consists of three conceptual components: a rule base, which contains a selection of fuzzy rules; a dictionary, which defines the membership functions used in the fuzzy rules; and a reasoning mechanism, which performs the inference produce (usually the fuzzy reasoning) [9]. Q3 According to [36], there is another class of fuzzy systems called type-2 fuzzy systems, which were also proposed by Zadeh. The reason to augment the original fuzzy systems is to consider higher levels of uncertainty, broadening their scope. In type-2 fuzzy systems the membership functions can now return a range of values that varies depending on the uncertainty involved, not only in the input but also on the same membership function. The structure of the type-2 fuzzy rules is the same as for the type-1 case because the distinction between type-2 and type1 is associated with the nature of the membership functions. Hence, the only difference is that now some or all the sets involved in the rules are of type-2 [46]. As an example: Consider the case of a fuzzy set characterized by a Gaussian membership function with mean m and a standard deviation that can take values in [1r1, r2 ], i.e.,



lðxÞ ¼ exp 1=2

hx  mi2  ;

r½r1 ; r2 

r

ð8Þ

Corresponding to each value of r we will get a different membership curve (as shown in Fig. 1). So, the membership grade of any particular x(except x = m) can take any of a number of possible values depending upon the value of r i.e., the membership grade is not a crisp number, it is a fuzzy set. Fig. 3 shows the domain of the fuzzy set associated with x = 0.7; however, the membership function associated with this fuzzy set is not shown in the Fig. 2. Gaussian type-2 fuzzy set. A Gaussian type-2 fuzzy set is one in which the membership degree of every domain point is a Gaussian type-1 set contained in [0, 1]. Interval type-2 fuzzy set. An interval type-2 fuzzy set is one in which the membership degree of every domain point is a crisp set whose domain is some interval contained in [0, 1]. Footprint of uncertainty: The uncertainty in the primary memberships of a type-2 fuzzy set, Ã, consists of a bounded region that we call the ‘‘footprint of uncertainty’’ (FOU). Mathematically, it is the union of all primary membership functions [22]. Upper and lower membership functions. An ‘‘upper membership function’’ and a ‘‘lower membership functions’’ are two type-1 membership functions that are the bounds for the FOU of a type-2 fuzzy set Ã. The upper membership function is associated with the upper bound of the FOU(Ã). The lower membership function is associated with the lower bound of the FOU(Ã).

Pron1Low

1

Degree of membership

171

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

Pron1High

0.8

0.6

0.4

0.2

0 2.5

2.6

2.7

2.8

2.9

3

3.1

3.2

3.3

3.4

3.5

Prediction1 Fig. 1. Type-1 membership functions.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 5

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

1

Pron1Low

Pron2High

0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1 0 2.5

2.6

2.7

2.8

2.9

3

3.1

3.2

3.3

3.4

3.5

Fig. 2. Type-2 membership functions.

211 212 213

214

Operations of type-2 fuzzy sets f1 and A f2 is another type-2 fuzzy set, just as the union of type-1 fuzzy sets A1 and Union of type-2 fuzzy sets. The union of A A2 is another type-1 fuzzy set. More formally, we have the following expression:

f1 [ A f2 ¼ A

216 217 218

219

x2X

223

224

f1 \ A f2 ¼ A

228 229

230

1

ð9Þ

2

l Ae \ Ae ðxÞ=x 1

ð10Þ

2

Complement of a type-2 fuzzy set. The complement of a set is another type-2 fuzzy set, just as the complement of type-1 fuzzy set A is another type-1 fuzzy set. More formally we have:

f1 ¼ A

Z x

leA 0 1 ðxÞ=x

ð11Þ

where the prime denotes complement in the above equation. In this equation le0 is the secondary membership function. A1 Type-2 fuzzy rules. Consider a type-2 FLS having r inputs x1 2 X 1 ; . . . ; xr 2 X r and one output y e Y. As in the type-1 case, we can assume that there are M rules; but, in the type-2 case the ‘th rule has the form:

~ 1 and . . . xp is A e 1 ; THEN y is Y b1 R1: IF x1 is A 1 P

232

233

Z x2X

226 227

l Ae [ Ae ðxÞ=x

f1 and A f2 is another type-2 fuzzy set, just as the intersection of Intersection of type-2 fuzzy sets. The intersection of A f type-1 fuzzy sets A1 and A2 is another type-1 fuzzy set. More formally, we have the following expression:

221 222

Z

1 ¼ 1; . . . M:

6. Optimization of the ensemble neural network with particle swarm optimization

234

The operation of the optimization algorithm by the swarm of particles is given by a sequence of steps, which are:

235

1. 2. 3. 4.

236 237 238

ð12Þ

Generate the initial swarm. Evaluate the particles of the swarm. Updating the particle velocity. Calculating new positions of the particles.

239 240 241 242 243 244 245 246 247 248

From steps 2–4, begins an iterative process until a stopping criterion is satisfied. In Fig. 3, we can see the outline of the algorithm of particle swarm optimization. To optimize the Ensemble Neural Network following the outline of the algorithm of particle swarm optimization, the first step is to create a swarm of particles, and to create the swarm we have to choose the representation of candidate solutions (particles), for which we choose a real representation. Once we have the representation of candidate solutions (particles), set the number of variables to indicate the position of the particle or possible solution to our problem, number of the particle. Set the number of variables that will form the particle, the initial population is generated randomly within a set range for each variable according to upper and lower limits of the search space for each variable. Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 6

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

Fig. 3. Ensemble neural network model. 249

7. Model description

250

This section describes the proposed model for the neural network ensembles, the objective is to optimize the architecture of a neural network with the particle swarm optimization method. Fig. 3 represents the Ensemble Neural Network Model, where we have first the historical data, then the optimization algorithm based on swarms means of particles determines the number of modules of the ensemble neural network, which can be from 1 to 5 modules, as well as the number of layers that this module, that could of from 1 to 3 and the number of neurons by layer that could be of 1 to 3. Then after the responses of neural network ensemble are obtained the integration is performed with type-1 and type-2 fuzzy systems. The input and output linguistic variables of the fuzzy system have Gaussian membership functions. The performance of the type-2 fuzzy integrators is analyzed under different levels of uncertainty to find out the best design of the membership functions for the rules of the fuzzy system. For the type-2 fuzzy integrator using 2 membership functions, which are called ‘‘low prediction’’ and ‘‘high prediction’’ for each of the inputs and the output of the fuzzy system. Membership functions are of Gaussian type and we deal with 3 footprint uncertainty sizes of 0.3, 0.4 and 0.5 to obtain a better prediction of the time series. If we consider 5 modules in the ensemble then the fuzzy system will have 5 input linguistic variables and the rules used are determined according to the possible combinations according to the number of inputs and membership functions of the fuzzy system. Considering that have 5 inputs and two membership functions for each of the inputs then there are 32 possible combinations. Each of the modules in the ensemble is a neural network and the calculations are performed as follows: x of a node is defined by the weighted sum of the incoming signals plus a bias term. For instance, the net The net input  input of the output of the node j is given as follows.

251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269

xj ¼

X

xij þ xj ;

i

271 272 273 274 275

276 278

x ¼ f ðxj Þ ¼

1 ; 1 þ expðxj Þ

ð13Þ

where xi is the output of node i located in any of the previous layers, xij is the weight associated with the link connecting nodes i and j, and xj is the bias of node j. Since the weights xij are actually internal parameters associated with each node j changing the weights of the node will alter behavior of the node and in turn alter behavior of the whole backpropagation MLP. A squared error measure for the pth input–output pair is defined as:

EP ¼

X

ðdk  xk Þ

2

ð14Þ

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx 279 280 281 282 283 284

285 287

7

where dk is the desired output for the node xk is the actual output for the node when the input part of the pth data pair is presented. In addition a fuzzy system is used as a response integrator for the ensemble. This means that the outputs of the modules (neural network) are the inputs to the fuzzy system (integrator of the responses). The global or total results are obtained by using the inference process in the fuzzy system. The total result of the fuzzy integrator system is calculated with the following equation:

Pn i¼1 li ðxÞdx y¼ P ; n i¼1 li ðxÞ

ð15Þ

288

where x are the input data and l are the corresponding membership functions.

289

8. Methodology

290

This section describes the general architecture of the proposed ensemble neural network model. The aim of this work was to implement a PSO algorithm to optimize the neural network architectures, for each of the modules, number of neurons and number of layers and thus find an architecture for the neural network that produces the best results in each particular time series that is considered. Fig. 4 shows the overall architecture, where historical data for each time series prediction, in this case the Mexican Stock Exchange is used as input, and then the time series analyzed and in the next step the data provided to the modules, which are optimized with PSO in which the particle contains information of the ensemble neural network for the optimization. The particle gives us the number of modules, number of layers and number of neurons are then applied to the neural network ensemble and then these results of these modules are integrated with an integration method based on type-1 and type-2 fuzzy logic and finally obtain the prediction error.

291 292 293 294 295 296 297 298

Fig. 4. General architecture of the proposed ensemble neural network design.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 8 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

The Mexican Stock Exchange (MSE) is a financial institution that operates by a grant from the Department of Finance and Public Credit, with the goal of following closely the Securities Market of Values in Mexico. Derived from the monitoring of global trends and changes that have occurred in legislation, the MSE concluded with the conversion process, becoming a company whose shares are likely to be traded on the market stock exchange, with the initial public offering taking place on June 13 of 2008 with its shares representing its capital [31]. Data of the Mexican Stock Exchange time series: In this case 800 points are used, that correspond to a period from 11/09/ 05 to 01/15/09 where we plot in the x axis the historical data in this case the day and in the y axis the closing of the Mexican Stock Exchange (as shown in Fig. 5). In the experiment 70% of the data were used for the ensemble neural network training and 30% to test the network [31]. After some tests were performed, another period of data from 04/08/08 to 05/09/11 for this series was also selected, where we plot in the x axis the historical data in this case day, and the y axis is the closing of the Mexican Stock Exchange (as shown in Fig. 6). For the optimization of the structure of the ensemble neural network a particle swarm optimization algorithm was used. The structure of the swarm particle represents the structure of the ensemble. The objective function is defined to minimize the prediction error as follows:

314

EM ¼

318

D

ð16Þ

i¼1

where a, corresponds to the predicted data depending on the output of the network modules, X represents real data, D the Number of Data points and EM is the total average prediction error.

3.4

Mexican Stock Exchange

x 104

3.2 3 2.8 2.6

Closing

317

!,

2.4 2.2 2 1.8 1.6 1.4 0

100

200

300

400

500

600

700

800

700

800

Days Fig. 5. Mexican Stock Exchange of time series 1.

4

Mexican Stock Exchange

x 104

3.5

Closing

316

D X jai  xi j

3

2.5

2

1.5 0

100

200

300

400

500

600

Days Fig. 6. Mexican Stock Exchange of time series 2.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 9

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346

The corresponding particle structure is shown in Fig. 7. Fig. 7 represents the particle structure to optimize the ensemble neural network, where the parameters that are optimized are the number of modules, number of layers, and number of neurons of the ensemble neural network. PSO determines the number of modules, number of layers and number of neurons per layer, that the neural network ensemble should have and that meets the objective to achieve a better Prediction error. The parameters for the particle swarm optimization algorithm are: 100 Particles, 100 iterations, Cognitive Component (C1) = 2, Social Component (C2) = 2, Constriction coefficient of linear increase (C) = (0–0.9) and Inertia weight with linear decrease (W) = (0.9–0). We consider a number of 1–5 modules, number of layers of 1–3, number neurons of 1–30. The network parameters used to train the neural network ensemble for each of the ensembles are 100 epochs, learning rate of = 0.01, the error goal 0.01, training method of Levenberg–Marquardt (trainlm), these parameters were used in this work because previous tests were performed and we managed to obtain a good prediction error with these parameters. The parameters for the particle swarm optimization algorithm are: 100 particles, 100 iterations and these parameters were used because previously an optimization with a genetic algorithm was considered and to compare it with this method we used similar parameters, Cognitive Component (C1) = 2, Social Component (C2) = 2, Constriction coefficient of linear increase (C) = (0–0.9) and Inertia weight with linear decrease (W) = (0.9–0), these parameters were used since we did perform a manual testing and by changing the values and with these parameters we obtain a better prediction error also. We consider a number of 1–5 modules, number of layers of 1–3, neurons number of 1–30, as maximum number of modules, layers and neurons that had already previously been tested with 3 modules and increasing error gives us better prediction. The aggregation of the responses of the optimized ensemble neural network is performed with type-1 and type-2 fuzzy systems. In this work a fuzzy system consisting of 5 inputs depending on the number of modules of the neural network ensemble and one output are used. Each input and output linguistic variable of the fuzzy system uses 2 Gaussian membership functions because the performance of the type-2 fuzzy aggregators is analyzed under different levels of uncertainty to find out the best design of the membership functions for the 32 rules of the fuzzy system with good results. Previous tests have been performed only with a three input fuzzy system, however the fuzzy system changes according to the responses of the neural network to the fuzzy system to give us better prediction error. In the type-2 fuzzy system we also change the levels of uncertainty to obtain the best prediction error. Fig. 8 shows a fuzzy system consisting of 5 inputs depending on the number of modules of the neural network ensemble and one output. Each input and output linguistic variable of the fuzzy system uses 2 Gaussian membership functions. The

Number Of Module

Number Of Layers

Neurons 1

...

Neurons n

Fig. 7. Particle structure to optimize the ensemble neural network.

Prediction1 (2)

MexicanStock Prediction2 (2)

(mamdani) Prediction3 (2) 32 rules

Prediction (2)

Prediction4 (2)

Prediction5 (2) System MexicanStock: 5 inputs, 1 outputs, 32 rules

Fig. 8. Type-2 fuzzy system for the Mexican Stock Exchange.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 10

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

Fig. 9. Type-2 fuzzy rules for predicting the Mexican Stock Exchange.

347 348

performance of the type-2 fuzzy aggregators is analyzed under different levels of uncertainty to find out the best design of the membership functions for the 32 rules of the fuzzy system. Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

11

350

Fig. 9 represents the possible 32 rules of the fuzzy system; we have 5 inputs in the fuzzy system with 2 membership functions, and outputs with 2 membership functions. These fuzzy rules are used for the type-1 and type-2 fuzzy systems.

351

9. Simulation results

352

In this section we present the simulation and test results obtained with the Particle swarm optimization algorithm applied in the design of the ensemble neural network for the Mexican Stock Exchange. A particle swarm algorithm was used to optimize the structure of an ensemble neural network and the best obtained architecture is represented in Fig. 10. Fig 10 shows the architecture of the neural network obtained by particle swarm optimization. We used two layers in each module. In module 1, in the first layer we used 21 neurons and 16 neurons in the second layer and in module 2 we used 21 neurons in the first layer and 28 neurons in the second. The optimization method using swarms of particles determines the number of neurons in the neural network ensemble. The Levenberg–Marquardt (LM) training method was used; we also applied 3 delays to the network.

349

353 354 355 356 357 358 359

Fig. 10. Ensemble neural network architecture for the MSE series 1.

Table 1 Particle swarm results for the ensemble neural network.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 12

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

Mexican Stock Exchange

x 10 4 3.4 3.2 3

Close

2.8 2.6 2.4 2.2 2 1.8 1.6 500

550

600

650

700

750

800

Days Fig. 11. Prediction with the optimized ensemble neural network of the MSE for time series 1.

Table 2 Results type-1 fuzzy aggregation of MSE for time series 1.

360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381

Table 1 shows the results of the optimization algorithm using particles for which 30 experiments were performed, but this table only shows the best 10 results, where the best obtained result was in the 6 wherein the prediction error was: 0.0089454. Fig. 11 shows the plot of real data against the predicted data generated by the ensemble neural network for time series Q4 number 1. Pink color1 represents the real data and the green color represents prediction data. Fuzzy integration is performed by implementing a type-1 fuzzy system in which the best result was in the experiment of row number 7 of Table 2 with an error of: 0.1458. Fuzzy aggregation is also performed by implementing a type-2 fuzzy system in which the results are as follows: for the best experiment with a degree of uncertainty of 0.3 a forecast error of 0.3685 was obtained, and with a degree of uncertainty of 0.4 a forecast error of 0.3960 was obtained and with a degree of uncertainty of 0.5 a forecast error of 0.3761 was obtained, as shown in Table 3. Fig. 12 shows the architecture of the neural network obtained by the particle swarm optimization method. We used two layers in each module. In module 1, in the first layer we used 12 neurons and 11 neurons, in the second layer; in module 2 we used 6 neurons in the first layer and 18 neurons in the second. The optimization method using swarms of particles determines the number of neurons in the neural network ensemble. The Levenberg–Marquardt (LM) training method was used; we also applied 3 delays to the network. Table 4 shows the particle swarm results (as shown in Fig. 12), where the best prediction error is of 0.0079875, which is shown in row number two. Fig. 13 shows the plot of real data against the predicted data generated by the ensemble neural network for time series number 2. The Pink color represents the real data and the green color represents the prediction data. Fuzzy aggregation is performed by implementing a type-1 fuzzy system in which the best result was in the experiment of row number 2 of Table 5 with an error of: 0.0952.

1

For interpretation of color in Figs. 11 and 13, the reader is referred to the web version of this article.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

13

Table 3 Results of type-2 fuzzy aggregation of MSE for time series 1.

384

Fuzzy aggregation is also performed by implementing a type-2 fuzzy system in which the results for the best experiment were follows: with a degree of uncertainty of 0.3 yielded a forecast error of 0.1780, with a degree of uncertainty of 0.4 an error of 0.1750 and with a degree of uncertainty of 0.5 an error of 0.1829, as shown in Table 6.

385

10. Comparison of results

386

A comparison of results for this time series of the Mexican Stock Exchange for the period of 11/09/05 to 01/15/09 with the paper entitled: ‘‘An Ensemble Neural Network Architecture with Fuzzy Response Integration for Complex Time Series Prediction’’ [34], shows that the best result was using an ensemble neural network architecture with 1 layer using 3 delays. The error obtained by the average integration was 0.0479, by the average weighted integration was 0.0519 and by fuzzy integration was 0.1094, respectively. In the same work the authors decided to implement an evolutionary approach to optimize the membership functions and rules of this system and the obtained error was of 0.045848. The genetic algorithm was applied to optimize the monolithic neural network for the Mexican Stock Exchange time series: 0.1092, (as shown in Fig. 9 and Table 1). In this paper the best result when applying the genetic algorithm to optimize the ensemble neural network was: 0.00112, (as shown in Fig. 11 and Table 4). This shows that our hybrid ensemble neural approach produces better results for this time series. After a comparison of results for this time series of the Mexican Stock Exchange for the period of 11/09/05 to 01/15/09 with the paper entitled: ‘‘A new Approach for Time Series Prediction Using Ensembles of ANFIS Models’’ [30], where the best result was achieved when integration by weighted average was used (an error of 0.0123 was obtained). Also, in this paper with integration by average an error of 0.0127 was obtained, which shows a very similar result. In this case, no significant difference can be found. In the literature there are only two references to forecast the time series of the Mexican Stock Exchange, and this is why the comparison was made with only these two studies. After a comparison of results for this time series of the Mexican Stock Exchange for the period of 11/09/05 to 01/15/09 and period of data from 04/08/08 to 05/09/11 with the paper entitled: ‘‘ Genetic Optimization of Ensemble Neural Networks

382 383

387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403

Fig. 12. Ensemble neural network architecture for the MSE series 2.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 14

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

Table 4 Particle swarm results for the ensemble neural network of MSE for time series 2.

Mexican Stock Exchange

3.6 3.4 3.2 3

Close

2.8 2.6 2.4 2.2 2 1.8 1.6 500

550

600

650

700

750

800

Days Fig. 13. Prediction with the optimized ensemble neural network of the MSE for time series 2.

409

for Complex Time Series Prediction of the Mexican Exchange Prediction’’ [33], where the best result was obtained for the network optimization for series 1 was: 0.0112 (as shown in Fig. 9 and Table 1) and for the series 2 where the best result was obtained for the network optimization for series 1 was: 0.0132 (as shown in Fig. 11 and Table 4). When we apply the optimization by the particle swarm method to optimize the architecture the for the neural network ensemble, it improves the prediction error for this time series of the Mexican Stock Exchange in 95%, as can be seen when comparing with the Refs. [30,33].

410

10.1. Statistical t students test

404 405 406 407 408

411 412 413

Q5

In statistics, the student t-test or t-test is a test that uses the statistical Student’s t distribution to find out if a proposed null hypothesis is true. The test is applied when the population is normally distributed, but the sample size is too small for making the inference based on a normally distributed statistic, using an estimate of the standard deviation instead of the Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 15

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx Table 5 Results type-1 fuzzy aggregation of MSE for time series 2.

Table 6 Results type-2 fuzzy aggregation of MSE for time series 2.

Table 7 Results of t student test for the time series. Time series

N(type-1)

N(type-2)

t value

P value

Mexican Stock Exchange series 1 Mexican Stock Exchange series 2

30 30

90 90

5.96 4.56

0.00 0.00

Table 8 Results of MSE for time series 1. Experiment Experiment Experiment Experiment Experiment Experiment Experiment Experiment Experiment Experiment Experiment

414 415 416 417 418 419 420 421 422

1 2 3 4 5 6 7 8 9 10

Prediction error for ensemble neural network

Prediction error with type-1 fuzzy aggregation

Prediction error with type-2 fuzzy aggregation

0.405 0.5763 0.3333 1.8024 1.7938 1.7880 0.3337 0.2724 1.121 0.2916

0.1726 0.0952 0.1565 0.1554 0.3523 0.2608 0.1070 0.2497 0.1126 0.1400

0.1700 0.1750 0.2176 0.2152 0.1829 0.2129 0.2182 0.2258 0.2316 0.1319

actual value. In this section the results with the t student test are shown, and it can be noticed that there is statistical evidence of significant difference in the results between the type-1 and type-2 fuzzy systems. In other words, it can be said that for this time series prediction results are better with type-2 fuzzy aggregation than with type-1 fuzzy integration. The number of samples used for the type-1 were 10 and 30 samples for type-2, the t value obtained for type-1 in series number 1 is: 5.96 and 4.56 for series number 2, so that the results show that in our tests more than 99% confidence was obtained of significant improvement with type-2 fuzzy logic according to the achieved value of P, as shown in Table 7. Table 8 shows results of the prediction error for the Ensemble Neural Network (without fuzzy aggregation), the prediction error with type-1 fuzzy aggregation and the prediction error with type-2 fuzzy and a 0.3 level of uncertainty for the time series 1.

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 16

M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx

425

In Table 8 it is clearly noted that using the response aggregation (type-1 or type-2) produces results that are significantly improved with respect to only using the results of the modules in the ensemble (first column of results). Both type-2 and type-1 fuzzy aggregation in the ensemble clearly help outperform the ensemble without fuzzy logic.

426

11. Conclusions

427

433

In this paper we considered the PSO algorithm to optimize the architecture of the ensemble neural network to predict the time series of the Mexican Stock Exchange, where good results were obtained, and we can say that this algorithm is good in speed when compared with other optimization techniques. PSO is an effective and efficient meta-heuristic to find the solution of this kind of problems. In conclusion the use of ensemble neural networks with type-2 fuzzy integration could be a good choice in predicting complex time series. Future works would be to optimize with the particle swarm algorithm the type-1 and type-2 fuzzy system aggregators to obtain an even better prediction error. type-1 and type-2 fuzzy systems could be optimized in terms of the parameters of the membership functions, membership type and number of rules.

434

12. Uncited references

423 424

428 429 430 431 432

435

Q6

[6,12,17,38,39,45].

436

Acknowledgment

437 438

We would like to express our gratitude to the CONACYT, Tijuana Institute of Technology for the facilities and resources granted for the development of this research

439

References

440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485

[1] A. Andreas, W. Sheng, Introduction optimization, Practical Optimization Algorithms and Engineering Applications, Springer, 2007. pp. 1–4. [2] P.D. Brockwell, R.A. Davis, Introduction to Time Series and Forecasting, Springer-Verlag, New York, 2002, pp. 1–219. [3] O. Castillo, P. Melin, Hybrid intelligent systems for time series prediction using neural networks, fuzzy logic, and fractal theory, IEEE Trans. Neural Netw. 13 (6) (2002) 1395–1408. [4] O. Castillo, P. Melin, Simulation and forecasting complex economic time series using neural networks and fuzzy logic, Proc. Int. Neural Netw. Conf. 3 (2001) 1805–1810. [5] O. Castillo, P. Melin, Simulation and forecasting complex financial time series using neural networks and fuzzy logic, Proc. IEEE Int. Conf. Syst. Man Cybernet. 4 (2001) 2664–2669. [6] O. Castillo, P. Melin, Type-2 Fuzzy Logic: Theory and Applications, Springer-Verlag, New York, 2008, pp. 29–40. [7] J.R. Chang, L.Y. Wei, C.H. Cheng, A hybrid ANFIS model based on AR and volatility for TAIEX forecasting 11(1) (2011) 1388–1395. [8] J.R. Chang, L.Y. Wei, C.H. Cheng, OWA-based ANFIS model for TAIEX forecasting, Econ. Model. 30 (2013) 442–448. [9] C.H. Cheng, T.L. Chen, H.J. Theo, C.H. Chiang, Fuzzy time-series based on adaptive expectation model for TAIEX forecasting, Expert Syst. Appl. 34 (2) (2008) 1126–1132. [10] C.H. Cheng, G.W. Cheng, J.W. Wang, Multi-attribute fuzzy time series method based on fuzzy clustering 34(2) (2008) 1235–124. [11] P. Cowpertwait, A. Metcalfe, Time series, Introductory Time Series R, Springer, Dordrecht, Heidelberg, London, New York, 2009. pp. 2–5. [12] P. Cunningham, J. Carney, S. Jacob, Stability problems with artificial neural networks and the ensemble solution, Artif. Intell. Med. 20 (3) (2000) 217–225. [13] H. Drucker, R. Schapire, P. Simard, Improving performance in neural networks using a boosting algorithm, in: S.J. Hanson, J.D. Cowan, C.L. Giles (Eds.), Advances in Neural Information Processing Systems 5, Denver, CO, Morgan Kaufmann, San Mateo, CA, 1993, pp. 42–49. [14] R. Eberhart, J. Kennedy, A new optimizer using particle swarm theory, in: Proc. 6th Int. Symp. Micro Machine and Human Science (MHS), October 1995, pp. 39–43. [15] R. Eberhart, Y. Shi, J. Kennedy, Swam Intelligence, Morgan Kaufmann, San Mateo, California, 2001. [16] A.P. Engelbrecht, Fundamentals of Computational Swarm Intelligence, John Wiley & Sons, 2006. [17] H.J. Escalante, M. Montes, L.E. Sucar, Particle swarm model selection, J. Mach. Learn. Res. 10 (2009) 405–440. [18] J. Hampshire, A. Waibel, A novel objective function for improved phoneme recognition using time-delay neural networks, IEEE Trans. Neural Netw. 1 (2) (1990) 216–228. [19] L.K. Hansen, P. Salomon, Neural network ensembles, IEEE Trans. Patt. Anal. Mach. Intell. 12 (10) (1990) 993–1001. [20] K. Huarng, Effective lengths of intervals to improve forecasting in fuzzy time series 123(3) (2001) 387–394. [21] J.S.R. Jang, C.T. Sun, E. Mizutani, Neuro-Fuzzy and Soft Computing, Prentice Hall, 1996. [22] N. Karnik, J.M. Mendel, Applications of type-2 fuzzy logic systems to forecasting of time-series, Inform. Sci. 120 (1–4) (1999) 89–111. [23] N. Karnik, J.M. Mendel, Introduction to type-2 fuzzy logic systems, IEEE Trans. Signal Process. 2 (1998) 915–920. [24] N. Karnik, J.M. Mendel, Operations on type-2 fuzzy sets, Fuzzy Sets Syst. 122 (2001) 327–348. [25] J. Kennedy, R. Eberhart, Particle swam optimization, Proc. IEEE Int. Conf. Neural Network (ICNN) 4 (1995) 1942–1948. [26] R. Maclin, J.W. Shavlik, Combining the predictions of multiple classifiers: using competitive learning to initialize neural networks, in: Proc. IJCAI-95, Montreal, Canada, Morgan Kaufmann, San Mateo, CA, 1995, pp. 524–530. [27] L.P. Maguire, B. Roche, T.M. McGinnity, L.J. McDaid, Predicting a chaotic time series using a fuzzy neural network, Inform. Sci. 112 (1–4) (1998) 125–136. [28] J. Mao, A case study on bagging, boosting and basic ensembles of neural networks for O CR, Proc. IJCNN-98, vol. 3, IEEE Computer Society Press, Anchorage, AK, Los Alamitos, CA, 1998, pp. 1828–1833. [29] P. Melin, O. Castillo, S. Gonzalez, J. Cota, W. Trujillo, P. Osuna, Design of Modular Neural Networks with Fuzzy Integration Applied to Time Series Prediction, vol. 41/2007, Springer, Berlin/Heidelberg, 2007. [30] P. Melin, J. Soto, O. Castillo, J. Soria, A new approach for time series prediction using ensembles of ANFIS models, Expert System with Applications, 2011. [31] Mexico Bank Database. (30.08.11). [32] I. Multaba, A. Hussain, Application of Neural Networks and Other Learning Technologies in Process Engineering, Imperial Collage Press, 2001 (03.12.08).

Q7

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006

INS 10870

No. of Pages 17, Model 3G

19 May 2014 M. Pulido et al. / Information Sciences xxx (2014) xxx–xxx 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508

Q8

17

[33] M. Pulido, O. Castillo, P. Melin, Genetic optimization of ensemble neural networks for complex time series prediction of the mexican exchange prediction, Int. J. Innov. Comput. Inform. Control 9(10), 4151–4166. [34] M. Pulido, A. Mancilla, P. Melin, An ensemble neural network architecture with fuzzy response integration for complex time series prediction, Evolutionary Design of Intelligent Systems in Modeling, Simulation and Control. [35] A. Sharkey, Combining Artificial Neural Nets: Ensemble and Modular Multi-Net Systems, Springer-Verlag, London, 1999. [36] Y. Shimshoni, Intrator classification of seismic signal by integrating ensemble of neural networks, IEEE Trans. Signal Process. 461 (5) (1998) 1194–1201. [37] S.R. Singh, A simple method of forecasting based on fuzzy time series, Appl. Math. Comput. 186 (1) (2007) 330–339. [38] T. Takagi, M. Sugeno, Fuzzy systems theory and its applications to modeling and control, IEEE Trans. Syst. Man Cybernet. SMC-15 (1) (1991) 116–132. [39] Y. Valle, G.K. Venayagamoorthy, S. Mohagheghi, J.-C. Hernandez, R.G. Harley, Particle swarm optimization: basic concepts, variants and applications in power systems, IEEE Trans. Evol. Comput. (2008) 171–195. [40] L.-Y. Wei, C.-H. Cheng, A hybrid recurrent neural networks model based on synthesis features to forecast the taiwan stock market, Int. J. Innov. Comput. Inform. Control 8(8), 5559–5571. [41] J. Xue, Z. Xu, J. Watada, Building an integrated hybrid model for short-term and mid-term load forecasting with genetic optimization, Int. J. Innov. Comput. Inform. Control 8 (10) (2012) 7381–7391. [42] R.N. Yadav, P.K. Kalra, J. John, Time series prediction with single multiplicative neuron model, Soft Comput. Time Ser. Predict. Appl. Soft Comput. 7 (4) (2007) 1157–1163. [43] R. Yusof, U. Khairuddin, M. Khalid, A new mutation operation for faster convergence in genetic algorithm feature selection, Int. J. Innov. Comput. Inform. Control 8 (10) (2012) 7363–7379. [44] L.A. Zadeh, in: R.R. Yager et al. (Eds.), Fuzzy Sets and Applications: Selected Papers, John Wiley, New York, 1987. [45] L. Zhao, Y. Yang, PSO-based single multiplicative neuron model for time series prediction, Expert Syst. Appl. 36 (2) (2009) 2805–2812. Part 2. [46] Z.-H. Zhou, Y. Jiang, Y.-B. Yang, S.-F. Chen, Lung cancer cell identification based on artificial neural network ensembles, Artif. Intell. Med. 24 (1) (2002) 25–36.

509

Please cite this article in press as: M. Pulido et al., Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange, Inform. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.05.006