[ieee 2013 sixth international conference on advanced computational intelligence (icaci) - hangzhou,...

6
2013 Sixth Inteational Conference on Advanced Computational Intelligence October 19-21, 2013, Hangzhou, China Ensembles of Echo State Networks for Time Series Prediction Wei Yao, Zhigang Zeng, Cheng Lian, and Huiming Tang Abstct- In time series prediction tasks, dynamic models are less popular than static models, while they are more suitable for modeling the underlying dynamics of time series. In this paper, a novel architecture and supervised learning principle for recurrent neural networks, namely echo state networks, are adopted to build dynamic time series predictors. Ensemble techniques are employed to overcome the randomness and instability of echo state predictors, and a dynamic ensemble predictor is therefore established. The proposed predictor is tested in numerical experiments and different strategies for training the predictor are also comparatively studied. A case study is then conducted to test the predictor's performance in realistic prediction tasks. I. I NTRODUCTION T IME series analysis and prediction theory has many practical utilities, in tasks such as economic predictions, weather forecasts, early warnings for geological disasters and so on [1-4]. A time series can be considered as a sequential observation of a natural or an artificial system, and from such a perspective, time series prediction is in fact the prediction of the developing trend of a target system. When the system is completely known, it is not a demanding job to predict the developing trend. However, the dynamics of realistic systems are usually very complex and their internal mechanisms can hardly be described in a clear mathematical model. Therefore, time series prediction techniques, which are based on observations, are resorted to. The general idea of basically all the time series prediction techniques is to learn the developing patterns from previ- ously observed values and then give predictions about the future values. The most widely used conventional time series prediction methods are auto-regressive models [5] and gray models [6]. With the development of neural computing techniques, neural networks have also been more and more popular a choice for handling time series data. Neural networks are among the most important data driven modeling methods, and there are many different neural networks. The structures of neural networks can be divided into two main cate- gories, namely the feed-forward networks and the recurrent networks. In feed-forward networks, the computing units, neurons, are arranged in different layers, and there are no Wei Yao is with the School of Computer Science. South-Central Univer- sity for Nationalities.Wuhan. China (email: [email protected]). Zhigang Zeng (Corresponding Author) and Cheng Lian e with the School of Automation. Huazhong University of Science and Technology. Wuhan,China (email: [email protected].liancheng@hust.edu.cn). Huiming Tang is with the Faculty of Engineering, China University of Geosciences,Wuhan,China. The work is supported by the Natural Science Foundation of China under Grants 6 1203286 and 6 1125303, the 973 Program of China under Grant 2011 CB710606, the Specialized Resech Fund for the Doctoral Program of Higher Education of China under Grant 20I00 142110021. 978-1-4673-6343-3/13/$3\.00 ©2013 IEEE 299 close loop connections in the structure. In recurrent networks, neurons are more freely connected, and connections between neurons may form directed cycles. Difference in structures resulted in difference in function. Feed-forward networks are static models, while recurrent networks are dynamic. There are many successful applications of feed-forward neural networks in time series prediction tasks as well as other machine learning tasks [7-9]. However, the realistic systems behind the time series observations are in essence dynamic processes. When applying feed-forward neural net- works, the time series prediction problem has to be trans- formed into a static regression problem, and the sequential observations have to be rearranged as input-output couples, which will lose the natural temporal ordering property of a time series. Therefore, using recurrent neural networks to establish dynamic prediction models will be a more direct and proper solution. The main issue hindering a wider usage of recurrent networks in time series prediction tasks is the lack of efficient learning algorithms. In this paper, Echo State Network (ESN) [10] is employed to establish dynamic predictors. ESN pro- vide an efficient supervised learning principle for dynamic modeling. Details will be explained in follow-up sections. On the other hand, in order to obtain more stable and accu- rate predictions, ensemble approaches are usually adopted in time series tasks. In ensemble approaches, multiple predictors are established independently to predict the future values of the same time series. Multiple and usually various predictions for the same process are obtained. These predictions are then combined to generate a single prediction. It is not surprising that the ensemble can give more stable prediction than its component predictors, since the combina- tion can "neutralize" the performance between the good com- ponent predictors and the bad component predictors. And it is also widely recognized a fact that the combined prediction can sometimes be superior to the best prediction achieved by all the component predictors. Ensembles of feedforward neural networks and statistic models are common, while ensembles of recurrent neural networks are less frequently reported. In our study, ensembles of ESN are established and their performances are comparatively studied. This paper is organized as follows. In Section II, a brief introduction about the structure and the learning algorithm for ESN is provided. In Section III, ensemble prediction ap- proaches are discussed and the most proper ensemble strategy om some candidates for ESN predictors is proposed. In Section IV, the structure of the ESN ensemble predictor and the proposed ensemble prediction method are explained in detail. Numerical experiments on Mackey-Glass benchmark data and case study on realistic landslide data are described

Upload: huiming

Post on 11-Apr-2017

214 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

2013 Sixth International Conference on Advanced Computational Intelligence October 19-21, 2013, Hangzhou, China

Ensembles of Echo State Networks for Time Series Prediction

Wei Yao, Zhigang Zeng, Cheng Lian, and Huiming Tang

Abstract- In time series prediction tasks, dynamic models are less popular than static models, while they are more suitable for modeling the underlying dynamics of time series. In this paper, a novel architecture and supervised learning principle for recurrent neural networks, namely echo state networks, are adopted to build dynamic time series predictors. Ensemble techniques are employed to overcome the randomness and instability of echo state predictors, and a dynamic ensemble predictor is therefore established. The proposed predictor is tested in numerical experiments and different strategies for training the predictor are also comparatively studied. A case study is then conducted to test the predictor's performance in realistic prediction tasks.

I. INTRODUCTION

T IME series analysis and prediction theory has many practical utilities, in tasks such as economic predictions,

weather forecasts, early warnings for geological disasters and so on [1-4]. A time series can be considered as a sequential observation of a natural or an artificial system, and from such a perspective, time series prediction is in fact the prediction of the developing trend of a target system. When the system is completely known, it is not a demanding job to predict the developing trend. However, the dynamics of

realistic systems are usually very complex and their internal

mechanisms can hardly be described in a clear mathematical

model. Therefore, time series prediction techniques, which are based on observations, are resorted to.

The general idea of basically all the time series prediction techniques is to learn the developing patterns from previ­ously observed values and then give predictions about the future values. The most widely used conventional time series

prediction methods are auto-regressive models [5] and gray

models [6]. With the development of neural computing techniques,

neural networks have also been more and more popular a choice for handling time series data. Neural networks are among the most important data driven modeling methods, and there are many different neural networks. The structures

of neural networks can be divided into two main cate­gories, namely the feed-forward networks and the recurrent

networks. In feed-forward networks, the computing units,

neurons, are arranged in different layers, and there are no

Wei Yao is with the School of Computer Science. South-Central Univer­sity for Nationalities. Wuhan. China (email: [email protected]).

Zhigang Zeng (Corresponding Author) and Cheng Lian are with the

School of Automation. Huazhong University of Science and Technology. Wuhan, China (email: [email protected]@hust.edu.cn).

Huiming Tang is with the Faculty of Engineering, China University of Geosciences, Wuhan, China.

The work is supported by the Natural Science Foundation of China under Grants 61203286 and 61125303, the 973 Program of China under Grant 20 11 CB710606, the Specialized Research Fund for the Doctoral Program of Higher Education of China under Grant 20 I 00 142110021.

978-1-4673-6343-3/13/$3\.00 ©2013 IEEE 299

close loop connections in the structure. In recurrent networks,

neurons are more freely connected, and connections between neurons may form directed cycles. Difference in structures resulted in difference in function. Feed-forward networks are static models, while recurrent networks are dynamic.

There are many successful applications of feed-forward

neural networks in time series prediction tasks as well as other machine learning tasks [7-9]. However, the realistic systems behind the time series observations are in essence

dynamic processes. When applying feed-forward neural net­works, the time series prediction problem has to be trans­formed into a static regression problem, and the sequential

observations have to be rearranged as input-output couples, which will lose the natural temporal ordering property of a time series.

Therefore, using recurrent neural networks to establish

dynamic prediction models will be a more direct and proper solution. The main issue hindering a wider usage of recurrent networks in time series prediction tasks is the lack of efficient learning algorithms. In this paper, Echo State Network (ESN) [10] is employed to establish dynamic predictors. ESN pro­vide an efficient supervised learning principle for dynamic

modeling. Details will be explained in follow-up sections.

On the other hand, in order to obtain more stable and accu­rate predictions, ensemble approaches are usually adopted in time series tasks. In ensemble approaches, multiple predictors are established independently to predict the future values of the same time series. Multiple and usually various predictions for the same process are obtained. These predictions are then combined to generate a single prediction.

It is not surprising that the ensemble can give more stable prediction than its component predictors, since the combina­

tion can "neutralize" the performance between the good com­ponent predictors and the bad component predictors. And it is also widely recognized a fact that the combined prediction

can sometimes be superior to the best prediction achieved by all the component predictors. Ensembles of feedforward neural networks and statistic models are common, while ensembles of recurrent neural networks are less frequently

reported. In our study, ensembles of ESN are established and their performances are comparatively studied.

This paper is organized as follows. In Section II, a brief introduction about the structure and the learning algorithm for ESN is provided. In Section III, ensemble prediction ap­proaches are discussed and the most proper ensemble strategy

from some candidates for ESN predictors is proposed. In Section IV, the structure of the ESN ensemble predictor and the proposed ensemble prediction method are explained in

detail. Numerical experiments on Mackey-Glass benchmark data and case study on realistic landslide data are described

Page 2: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

in Section V and Section VI, respectively. The conclusions are drawn in Section VII.

II. ECHO STATE NETW ORK PREDICT OR

The structure of an echo state network is not obviously different from a common recurrent neural network. Neurons in the network can be divided into three groups: the input neurons, the output neurons and the internal neurons. Re­current connections mainly exist between internal neurons and these recurrently connected internal neurons build up the very component of the network which is in charge of producing dynamics. This component, namely the internal

neurons and the connections between them, is called a reservoir. Signals are conducted from the input neurons to the reservoir to provoke dynamic processes in internal neurons.

Then these dynamic processes are collected by the output neurons to combine into desired dynamics. This is how an echo state network simulates the development of realistic

dynamic processes. Training of an echo state network can be accomplished

by two steps. First, initial connection weights for all the connections in the network are randomly generated. All the connection weights inside the reservoir will not be changed at all hereafter, therefore the generating of the internal connections in the first place has to fulfill some conditions, which are called "the echo state property" [10, 11], to ensure the memory capability as well as the stability of the network. In the second step of training, only the output connections

will be adjusted, aiming at producing a best combination of

the internal dynamics to approximate the desired dynamic. The structure of the echo state network and its training

1. Initi alization

j !

d(t. l ) � Input connections

ReC"lUTent connect1 on s

o

o

o o

o

o

o

2 . Training 3. Testing

j j � d(t)

Output connections

Fig. l. Structure of the echo state network predictor

process is illustrated in Fig. 1. In a time series prediction task, the training data set for an echo state network can be formulated as follows. On one hand, the time series is divided into two parts. The first part is picked up to train the model,

while the second part is reserved as the testing set. On the other hand, a delayed version of the time series is created as the input to stimulate internal dynamics in the network,

and the original time series is kept as the desired output, according to which the output connections will be adjusted during training.

300

Assuming the length of the time series observations is T,

then it can be denoted as d(t), t = 1 rv T. The length of

the first part is h, therefore the training set is obtained as {d(t -1),d(t)}, t = 1 rv h. After initialization, the internal states, denoted as a matrix X(t), will be updated following

X(t) = f(Wx . X(t -1) + Wi· d(t -1)), t = 1 rv tl, (I)

where fO denotes the activation function of the internal neurons. Wx and Wi are the 'randomly' created internal connection weights and input connection weights. Then the output connection weights will be trained following

Wo = argming(d(t) - Wo· X(t), Wo), t = to rv h, (2)

where gO denotes the objective function, which is monotoni­cally decreasing as the deviation d( t) -Wo . X (t) decreasing, and usually a regularization term is also included in it. Notice that to > 1. When t = 1 rv to -1, the internal states X (t) will be considered as initial transient states and discarded. After training, the network will be able to produce output which is close to the realistic time series. And its accuracy can be assessed following

error = e(d(t) - Wo . X(t)), t = tl + 1 rv T, (3)

where eO is a certain kind of error, such as mean square

error. Once the accuracy of the model is acceptable, it can be used to predict future values.

III. PREDICT OR ENSEMBLES

Since the input and internal connection weights are ran­domly created, the echo state network predictor will hold

some randomness by nature, and its performance in pre­diction tasks tends to be unstable. In order to cope with this problem, ensemble prediction approach is employed. Multiple ESN predictors are created for the same prediction task, and their predictions are combined.

Diversity among component predictors and high quality of component predictors are the two requirements for a successful ensemble. If the component predictors are similar,

the combination can not bring obvious improvement. If the

performances of the component predictors are poor, the per­formance of the ensemble may be further deteriorated. ESN ensembles, on the other hand, meet these two requirements

quite well. The random internal connections of ESN ensure the diversity. As regard to the accuracy of the components, on a benchmark task of predicting Mackey-Glass time series,

the ESN predictor achieved dramatically improvements by a factor of 2400 over previous techniques [10].

Usually, the ensemble prediction is obtained by average or weighted average the predictions of the component pre­

dictors. Averaging is the simplest ensemble approach, and as argued by Andrawis et. al. [12], simple combination schemes tend to perform best. As for weighted linear combination, the most important issue is assigning proper weight for each component predictor. It is preferred that the better component can take a bigger weight in the ensemble. Under such a

consideration, more sophisticated and adaptive ensembles,

Page 3: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

usually based on neural networks [13 16], have been pro­posed.

In our work, a feed-forward neural network is used to make an adaptive ensemble of ESN predictors. Therefore, the combination weights of the component predictors are obtained in an additional step of training. The very effective

training algorithm of extreme learning machine (ELM) [17] is adopted to train a single layer feed-forward neural network into an ensemble predictor. This ensemble predictor will denoted as the ESN-ELM predictor hereafter.

IV. ESN -ELM PREDICTOR

T ime selles samples

naining Step!: ! naining Step2: 1 / dl {t+l)

d{f) .'-----1 ------, � -, ESN 2 1----... �

dn{t+1)

Fig. 2, Strncture of the ESN-ELM predictor

Testing: 1 d{l+1)

ELM is widely used in different kinds of pattern classi­fication and regression applications [18]. Although the ESN predictors are dynamic, the combination of their predictions is a static regression problem. Therefore, ELM's efficiency in regression applications make it a good choice for the ensemble process. The proposed ESN-ELM predictor fea­tures a parallel-cascade structure, as illustrated in Fig. 2. The component predictors and the ensemble unit combined into a

big network, which is a hybrid of multiple recurrent networks and a single layer feed-forward network.

There are two steps of training for the ensemble predic­tor. The first step of training is applied to the component predictors, following the algorithm which has already been

explained in Section II. The second step of training for the ELM ensemble unit is similar to the training of ESNs. The

input connections will be initialized previously then kept unchanged and only the output connections of the network

will be adjusted during the training process. In order to avoid

over-fitting, the training set is divided into two subsets. The first training set is used to train the component predictors,

namely the ESNs. Then the second training set is used to train the ELM ensemble unit.

The whole process of using the proposed ESN-ELM predictor to produce reliable predictions in a prediction task is as follow.

1) Initialization: n recurrent networks and a feed-forward network are connected to build a parallel-cascade structure. The input connections of all the component networks and

301

internal connections of the n recurrent networks are randomly initialized.

2) 1st training: The first training set is applied onto the recurrent networks, and the output connections of these recurrent networks are adjusted accordingly following the

training algorithm of ESNs. 3) 2nd training: The second training set is applied onto

the feed-forward part, and the output connections of this feed-forward network are adjusted accordingly, following the

training algorithm of ELMs. 4) Testing: After training, the reliability of the ESN-ELM

predictor can be evaluated quantitatively by computing the

deviations between its predictions and the real values in the testing set. If the predictor achieves acceptable prediction accuracy on the testing set, it may hopefully produce reliable

predictions for the really unknown future. 5) Predicting: Finally, the predictor can be used to predict

future values. For multi-step prediction, the iterative predic­tion strategy can be adopted. The first four of the above listed steps will be conducted in numerical experiments to verify the effectiveness of the proposed predictor. A realistic prediction task will also be reported in the following sections.

V. NUMERIC AL EXPERIMENTS

The benchmark Mackey-Glass time series is used in our

numerical predicting experiments. The purposes of the nu­merical experiments are threefold. First, since the training set will be divided into two subsets, how to divide it and how the dividing will affect the performance of the ensemble predictor need to be figured out. Second, the improvements

of the ensemble predictor, as compared to the component predictors as well as their average, are to be verified. Third, the effectiveness of the strategy that dividing the training set

into two parts will also be discussed. 1000 data points are selected out of the Mackey-Glass time

series. A one step delayed version of this data set is created to make up input-output couples. The first 50 data points are

used to provoke the dynamics in the ESNs, then follow-up 250 data couples are picked up to form the training set, and the left 700 couples are reserved for testing the predictors.

20 ESN predictors are established to build up the ensemble. The length of the two training sets sum up to be 250.

When the first training set gets shorter, the second will get longer. This is a tradeoff between the two steps of training. And when the first training set gets shorter, the training of the components predictors will be less sufficient, and

the performance of them will deteriorate. This explains the

phenomena as illustrated in Fig. 3 that the averaged root mean square errors (rmse) of the components' predictions raise up when the second training set gets longer. However, this deterioration is not obvious until the second training set

is longer them 120, which means that 130 training couples are sufficient for the ESNs. The best performance of the ensemble predictor is achieved when the second training set

is of a length of about 160. At this point, an optimal tradeoff is arrived that both the components (ESNs) and the ensemble unit (ELM) can be properly trained.

Page 4: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

0.9%

I c=:J mean ape I -_max ape 0.8%

0.7% -

0.6% -

0.5% -

0.4% -

0.3% -

0.2% -

I I I -

I I r I I I I I I ! r I I ! I I , J 0.1%

o 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 Av En

Fig. 4. Comparisons between component predictors and the ensemble predictor regarding absolute percentage error

Q) \ � 10-1 \ C \ "-55 \ 2 \ .. Ql .. :5 10-2 \ C \ o \ � \ } e \ I \ "- \ I \ � 10-3

\ " \ ..Q \"

\.......

'

... _---- ' " ........ " 1 0-4 c::=:::;:=:=:::::::;::==�-�-�-�- -.:- -=-- -=--�-�L--��J

o 20 40 60 80 100 120 140 160 180 200 220 240 length of training set 2

Fig. 3. Impacts of different separations between the two training sets on the testing results of the ensemble predictor

Absolute percentage error is another common measure­

ment for the predictions. The mean absolute percentage error can measure the general performance of the predictor on the testing set, while the maximum absolute percentage

error is also important because it indicate the stability of the predictor, and a predictor with low variance might be

preferred to a predictor with a better average error but with higher error peaks [19]. The ensemble predictor is compared with the 20 component predictors, as regarding to mean ape and max ape. The 20 component predictors are first trained on the whole 250 training couples to show their best, and their predictions are average. The apes of their predictions are

labeled as 1 rv 20 and ' Av' in Fig. 4 .Then these component predictors are reset and trained again on a subset with the length of 90, while the left 160 training couples are used to

302

train the ensemble unit. The apes of the prediction ensemble are labeled as 'En'. Both the averaging and the proposed ensemble can give more accurate and more stable predictions as compared to the components. And the proposed ensemble

is even better than the averaging.

TABLE I

COMPARISONS BETWEEN TWO ENSEMBLE STRATEGIES

length of training set 1 length of training set 2

average testing error of components training error of the ensemble

testing error of the ensemble

strategy 1 strategy 2

250 90

250 160

2.11e - 4 3.37e - 4

4.46e - 7 l.15e - 4

4.08e - 4 l.51e - 4

Since the length of the trammg set is usually limited, especially in realistic prediction tasks, there is a tradeoff between the two training steps in the proposed method. Then a question will arise naturally-why not reuse the training couples in the two training steps. Our reason for not doing so is that this strategy may lead to over-fitting problems. And our assumption is validated by a comparison between these

two training strategies during the numerical experiments, as reported in TABLE I. When reusing the 250 training couples

in both of the two training steps, better performances of component predictors are obtained. And for the training of

the ensemble, the training error in the form of rmse drops

dramatically. However, the ensemble predictor's performance on the testing set gets worse. The training error decreasing

and the testing error increasing, this is a typical sign of over-fitting. Therefore, it can be concluded that the training strategy adopted in the proposed ESN-ELM predictor is more

proper than the reusing strategy.

Page 5: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

It's worth mentioning that the time-consuming of neural computing process is not really a problem. The training of

all the component predictors can be accomplish in less than 2 seconds, as recorded during our experiments. Furthermore, the time-consuming of the training and the testing processes of the ensemble is less than 0.3 seconds. Since the length of the training set will be even shorter in the realistic prediction tasks we caring about, which will be explained in detail in

the following section, the efficiency of the proposed model will not be a problem in its applications.

VI. C ASE STUDY

After verifying its efficiency in numerical experiments, the

proposed method is used to predict landslide displacement time series. This is a case study that demonstrates the application of the proposed ensemble predictor in realistic prediction tasks.

We choose the Yuhuangge (Y HG) landslide as our study case. In order to give early warning of possible landslide disasters in the future, the displacements of the landslide body are recorded. The monthly recordings of Yuhuangge

landslide begin in May 2004, and a time series of lOJ month recordings are selected for our study. A denoising process is conducted. As illustrated in Fig. 5, the original time series

is severely polluted by noises, and the denoised time series is much smoother. The recordings of the first 76 months

35 �======�====�--��------�--------� ill ill 30 � � 25 en � 20 � C 15 Q) E � 10 co

Ci. :6 5

-- original recordings - - - de noised values

OL---------�----------�----------�----------� 2004-5 2006-6 2008-7

recording da1a 2010-9 20 12-10

Fig. 5. Original landslide displacement recordings and the corresponding denoised values

are used to train the ensemble predictor. After training, the ensemble predictor keeps running to produce prediction for the following up 25 months' displacements. The predictions

and the actual displacements are compared as illustrated in Fig. 6. Despite this realistic time series is quite short and the recordings are severely polluted by noises, which can hardly be removed completely, the predictions are basically acceptable.

VII. C ONCLUSIONS

A dynamic ensemble predictor, the ESN-ELM predictor,

is proposed in this paper. Since realistic systems behind time series observations are in essence dynamic, dynamic

models like the proposed ESN-ELM model, have larger

303

30

ill 28 ill � 26 E -;;; 24 Q) :::J � 22

� 20 E Q) J§ 18 Cl. 06 16

displacement predictions based on YHG recordings

-- actual displacements - e - displacement predictions

2 4 6 8 10 12 14 16 18 20 22 24 time step/month

Fig. 6. Landslide displacement predictions produced by the proposed ensemble predictor

potential to be trained into accurate predictors than static models. Successful applications of ESNs in task of predicting benchmark time series have already been reported, while the goal of our study is to further improve the performance of the dynamic predictor by building up ESN ensembles.

As verified by the results obtained in numerical experi­ments, the ensemble can not only suppress the instability of the ESN predictors but also improve the overall prediction

accuracy. And the adaptive ensemble based on a feed-forward neural network also shows improvement as compared to the averaging method. Another important find is that, reusing the training set may bring in over-fitting problems hence training

set separating will be a better strategy for training the ESN­

ELM predictor.

REFERENCES

[I] D. Orner Faruk, "A hybrid neural network and ARIMA model for water quality time series prediction," Engineering Applications of Artificial Intelligence, vol. 23, pp. 586-594, 2010.

[2] R.I. Povinelli and F. Xin, "A new temporal pattern identification method for characterization and prediction of complex time series events," IEEE Transactions on Knowledge and Data Engineering, vol. 15, pp. 339-352, 2003.

[3] E.w. Saad, DV Prokhorov and D.C. Wunsch, II, "Comparative study of stock trend prediction using time delay, recurrent and probabilistic neural networks," IEEE Transactions on Neural Networh, vol. 9, pp.

1456-1470, 1998.

[4] C.L. WU and K.w. Chau, "Data-driven models for monthly streamflow

time series prediction," Engineering Applications of' Artificial Intelli­gence, vol. 23, pp. 1350-1367,20 10.

[5] F.M. Pouzols, A. Lendasse and A.B. Barros, "Autoregressive time series prediction by means of fuzzy inference systems using nonparametric

residual variance estimation," Fuzzy Sets and Systems, vol. 161, pp. 471-497,2010.

[6] E. Kayacan, B. Ulutas and O. Kaynak, "Grey system theory-based

models in time series prediction," Expert Systems with Applications, vol. 37, pp. 1784-1789, 20 10.

[7] C.-M. Lee and C.-N. Ko" "Time series prediction using RBF neural

networks with a nonlinear time-varying evolution PSO algorithm," INeurocomputing, vol. 73, pp. 449-460, 2009.

[8] N.1. Sapankevych and R. Saukar, "Time series prediction using support vector machines: a survey," IEEE Computational Intelligence Magazine, 4, pp. 24-38, 2009.

[9] P. Yee and S. Haykin" "A dynamic regularized radial basis function network for nonlinear, nonstationary time series prediction," IEEE Transactions on Signal Processing, vol. 47, pp. 2503-2521,1999.

Page 6: [IEEE 2013 Sixth International Conference on Advanced Computational Intelligence (ICACI) - Hangzhou, China (2013.10.19-2013.10.21)] 2013 Sixth International Conference on Advanced

[10] H. Jaeger and H. Haas, "Harnessing nonlinearity: predicting chaotic systems and saving energy in wireless communication," Science, vol.

304, pp. 78-80, 2004. [II] M. Lukosevicius, H. Jaeger, "Reservoir computing approaches to

recurrent neural network training," Computer Science Review, vol. 3, pp. 127-149,2009.

[12] R.R. Andrawis, A.F. Atiya and H. EI-Shishiny, "Forecast combinations of computational intelligence and linear models for the NN5 time series forecasting competition," International Journal oj" Forecasting, vol. 27, pp. 672-688, 20 11.

[13] M. Heeswijk, Y. Miche, T. Lindh-Knuutila, PJ. Hilbers, T. Honkela, E. Oja and A. Lendasse, "Adaptive ensemble models of extreme learning machines for time series prediction," ICANN 2009, Springer Berlin Heidelberg, pp. 305-314,2009.

[14] J. Sun and H. Li, "Financial distress prediction using support vector machines: Ensemble vs. individual," Applied S(?/t Computing, vol. 12,

pp. 2254-2265, 20 12. [15] K. Siwek and S. Osowski, "Improving the accuracy of prediction of

PMlO pollution by the wavelet transformation and an ensemble of neural predictors," Engineering Applications oj" Artificial Intelligence, vol. 25, pp. 1246-1258,20 12.

[16] c.Y. Sheng, J. Zhao, W. Wang and H. Leung, "Prediction intervals for a noisy nonlinear time series based on a bootstrapping reservoir computing network ensemble," IEEE Transactions on Neural Networks and Learning Systems, vol. 24, pp. 1036-1048,2013.

[17] G.B. Huang, H.M. Zhou, XJ.Ding and R. Zhang, "Extreme learning machine for regression and multiclass classification," IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, vol. 42, pp. 513-529,2012.

[18] G.B. Huang, D. Wang, Y. Lan, "Extreme learning machines: a survey," International Journal of Machine Learning and Cybernetics, vol. 2, pp. 107-122, 20 1 1.

[19] M. De Felice and Y. Xin, "Short-Term load forecasting with neural network ensembles: a comparative study [application notes]," IEEE

Computational Intelligence Magazine, vol. 6, pp. 47-56, 20 11.

304