prediction of stock market index using neural networks an empirical study of bse

13
European Journal of Business and Manageme ISSN 2222-1905 (Paper) ISSN 2222-2839 (O Vol 4, No.12, 2012 Prediction of Stock R. Lakshman Naik *1 , B. 1. Department of Info 2. Department of MC 3. Department of CS * E-mail of the cor Abstract Predicting stock data with tradition neural network may be more suitab has to be made prior to forecasting. from large sets of data, which o Subsequently an Error Correction N as fundamental data are used as inp stocks of the BSE are predicted usi standard Error Correction Network predictions. The results on the sto strategy. Keywords: - Prediction of stock, EC 1. Introduction Different techniques are being used of neural networks has emerged as and consequently handle large sets patterns and irregularities as well quality is extremely useful for mo networks are frequently used for p Network (ECN) is built and implem network’s ability to make forecast successfully used as decision suppor and other trading issues are not cons Neural networks can be applied for forecasts of yield curves, excha approach in stock prediction is twofo Stock data is highly comple A large set of interacting i networks This approach is also applied fo to predict returns rather than actual over a long period of time [1], prior Documentation is made ignorin Frequently one has to handle missin A convenient way to work around th 2. Error Correction Networks Most dynamical systems contain bo relevant external forces may be hard the dynamics may be impossible to error as additional information to the 2.1. Mathematical Description Following set of equations consistin is very general form for discrete tim ent Online) 60 k Market Index Using Neural Ne Empirical Study of BSE Manjula 1 , D. Ramesh 1 , B. Sridhara Murthy 2 , Prof. SS ormatics, Kakatiya University, Warangal, A.P, India CA, KITS, Warangal, Andhra Pradesh, India. SE, Vaagdevi College of Engineering,Warangal, A.P, I rresponding author:[email protected] nal time series analysis has become one popular resea ble for the task, because no assumption about a suitab Furthermore, a neural network has the ability to ext often is required for a satisfying description of a Network is defined and implemented for an empirical s put to the network. One-step returns of the BSE stoc ing two separate network structures. Daily predictio whereas an extension of the Error Correction Netw ocks are less convincing; nevertheless the network CN, Backpropagation, Feedforward Neural Networks, in the trading community for prediction tasks. In re one of them. A neural network is able to work paral of data swiftly. The principal strength with the netwo as detecting multi-dimensional non-linear connectio odeling dynamical systems, e.g. the stock market. A pattern recognition tasks and non-linear regression. mented for an empirical study. Standard benchmarks a ts. The objective of this study is to conclude whet rt in a real trading situation. The matters of buy-sell si sidered. d to all sorts of financial problems, not only stock pre ange rates, bond rates etc. The principal motivation old: ex and hard to model; therefore a non-linear model is b input series is often required to explain a specific sto or the prediction task from the angle of economics. Su stock prices. Another reason is that it facilitates sta to model. ng needed data and importance is given to the data irre ng data at recreant intervals or even data emerging dis hese difficulties is by letting the network accept missin oth an autonomous part and a part governed by exter d to identify or data could be noisy. As a consequence o get. A remedy for a better model description is to u e system will take care of all the above data. ng of a state and output equations is a recurrent descrip me grids. A basic dynamical recurrent system depicte www.iiste.org etworks: An SVN Sarma 3 India arch issue. An artificial ble mathematical model tract useful information financial time series. study. Technical as well ck index and two major ons are performed on a work is used for weekly outperforms the naive Dynamic system. ecent years the concept llel with input variables ork is its ability to find ons in data. The latter Apart from that, neural An Error Correction are used to evaluate the ther an ECN could be ignals, transaction costs ediction tasks, also used for the neural network beneficial ock, which suites neural uch as stock forecasting abilization of the model elevant for the problem. scontinuous time series. ng data. nal forces. Many times a correct description of use the previous model ption of dynamic system ed in Fig. 1, at time t, it

Upload: alexander-decker

Post on 18-Dec-2014

226 views

Category:

Technology


3 download

DESCRIPTION

International Journals Call for Paper: http://www.iiste.org/Journals

TRANSCRIPT

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

Prediction of Stock Market Index Using Neural Networks: An

R. Lakshman Naik*1

, B. Manjula

1. Department of Informatics, Kakatiya University, Warangal, A.P, India

2. Department of MCA, KITS, Warangal, Andhra Pradesh, India.

3. Department of CSE, Vaagdevi College of Engineering,Warangal, A.P, India

* E-mail of the corresponding author:

Abstract

Predicting stock data with traditional time series analysis has

neural network may be more suitable for the task, because no assumption a

has to be made prior to forecasting. Furthermore, a neural network has the ability to extract useful information

from large sets of data, which often is required for a satisfying description of a financial time series.

Subsequently an Error Correction Network is defined and implemented for an empirical study. Technical as well

as fundamental data are used as input to the network. One

stocks of the BSE are predicted using t

standard Error Correction Network whereas an extension of the Error Correction Network is used for weekly

predictions. The results on the stocks are less convincing; nevertheless the net

strategy.

Keywords: - Prediction of stock, ECN,

1. Introduction

Different techniques are being used in the trading community for prediction tasks. In recent years the

of neural networks has emerged as one of them. A neural network is able to work parallel with input variables

and consequently handle large sets of data swiftly. The principal strength with the network is its ability to find

patterns and irregularities as well as detecting multi

quality is extremely useful for modeling dynamical systems, e.g. the stock market. Apart from that, neural

networks are frequently used for pattern recognition tasks an

Network (ECN) is built and implemented for an empirical study. Standard benchmarks are used to evaluate the

network’s ability to make forecasts. The objective of this study is to conclude whether an ECN could

successfully used as decision support in a real trading situation. The matters of buy

and other trading issues are not considered.

Neural networks can be applied to all sorts of financial problems, not only stock predicti

for forecasts of yield curves, exchange rates, bond rates etc. The principal motivation for the neural network

approach in stock prediction is twofold:

• Stock data is highly complex and hard to model; therefore a non

• A large set of interacting input series is often required to explain a specific stock, which suites neural

networks

This approach is also applied for the prediction task from the angle of economics. Such as stock forecasting

to predict returns rather than actual stock prices. Another reason is that it facilitates stabilization of the model

over a long period of time [1], prior to model.

Documentation is made ignoring needed data and importance is given to the data irrelevant for the problem.

Frequently one has to handle missing data at recreant intervals or even data emerging discontinuous time series.

A convenient way to work around these difficulties is by letting the network accept missing data.

2. Error Correction Networks

Most dynamical systems contain both an autonomous part and a part governed by external forces. Many times

relevant external forces may be hard to identify or data could be noisy. As a consequence a correct description of

the dynamics may be impossible to get. A remedy for a

error as additional information to the system will take care of all the above data.

2.1. Mathematical Description

Following set of equations consisting of a state and output equations is a recurrent descrip

is very general form for discrete time grids. A basic dynamical recurrent system depicted in Fig. 1, at time t, it

European Journal of Business and Management

2839 (Online)

60

Prediction of Stock Market Index Using Neural Networks: An

Empirical Study of BSE

, B. Manjula1, D. Ramesh

1, B. Sridhara Murthy

2, Prof. SSVN Sarma

Department of Informatics, Kakatiya University, Warangal, A.P, India

Department of MCA, KITS, Warangal, Andhra Pradesh, India.

Department of CSE, Vaagdevi College of Engineering,Warangal, A.P, India

il of the corresponding author:[email protected]

Predicting stock data with traditional time series analysis has become one popular research issue.

neural network may be more suitable for the task, because no assumption about a suitable mathematical model

has to be made prior to forecasting. Furthermore, a neural network has the ability to extract useful information

from large sets of data, which often is required for a satisfying description of a financial time series.

ubsequently an Error Correction Network is defined and implemented for an empirical study. Technical as well

as fundamental data are used as input to the network. One-step returns of the BSE stock index and two major

stocks of the BSE are predicted using two separate network structures. Daily predictions are performed on a

standard Error Correction Network whereas an extension of the Error Correction Network is used for weekly

predictions. The results on the stocks are less convincing; nevertheless the network outperforms the naive

Prediction of stock, ECN, Backpropagation, Feedforward Neural Networks, Dynamic system.

Different techniques are being used in the trading community for prediction tasks. In recent years the

of neural networks has emerged as one of them. A neural network is able to work parallel with input variables

and consequently handle large sets of data swiftly. The principal strength with the network is its ability to find

ities as well as detecting multi-dimensional non-linear connections in data. The latter

quality is extremely useful for modeling dynamical systems, e.g. the stock market. Apart from that, neural

networks are frequently used for pattern recognition tasks and non-linear regression. An Error Correction

Network (ECN) is built and implemented for an empirical study. Standard benchmarks are used to evaluate the

network’s ability to make forecasts. The objective of this study is to conclude whether an ECN could

successfully used as decision support in a real trading situation. The matters of buy-sell signals, transaction costs

and other trading issues are not considered.

Neural networks can be applied to all sorts of financial problems, not only stock predicti

for forecasts of yield curves, exchange rates, bond rates etc. The principal motivation for the neural network

approach in stock prediction is twofold:

Stock data is highly complex and hard to model; therefore a non-linear model is ben

A large set of interacting input series is often required to explain a specific stock, which suites neural

This approach is also applied for the prediction task from the angle of economics. Such as stock forecasting

ther than actual stock prices. Another reason is that it facilitates stabilization of the model

over a long period of time [1], prior to model.

Documentation is made ignoring needed data and importance is given to the data irrelevant for the problem.

equently one has to handle missing data at recreant intervals or even data emerging discontinuous time series.

A convenient way to work around these difficulties is by letting the network accept missing data.

tems contain both an autonomous part and a part governed by external forces. Many times

relevant external forces may be hard to identify or data could be noisy. As a consequence a correct description of

the dynamics may be impossible to get. A remedy for a better model description is to use the previous model

error as additional information to the system will take care of all the above data.

Following set of equations consisting of a state and output equations is a recurrent descrip

is very general form for discrete time grids. A basic dynamical recurrent system depicted in Fig. 1, at time t, it

www.iiste.org

Prediction of Stock Market Index Using Neural Networks: An

, Prof. SSVN Sarma3

Department of CSE, Vaagdevi College of Engineering,Warangal, A.P, India

become one popular research issue. An artificial

bout a suitable mathematical model

has to be made prior to forecasting. Furthermore, a neural network has the ability to extract useful information

from large sets of data, which often is required for a satisfying description of a financial time series.

ubsequently an Error Correction Network is defined and implemented for an empirical study. Technical as well

step returns of the BSE stock index and two major

wo separate network structures. Daily predictions are performed on a

standard Error Correction Network whereas an extension of the Error Correction Network is used for weekly

work outperforms the naive

, Feedforward Neural Networks, Dynamic system.

Different techniques are being used in the trading community for prediction tasks. In recent years the concept

of neural networks has emerged as one of them. A neural network is able to work parallel with input variables

and consequently handle large sets of data swiftly. The principal strength with the network is its ability to find

linear connections in data. The latter

quality is extremely useful for modeling dynamical systems, e.g. the stock market. Apart from that, neural

linear regression. An Error Correction

Network (ECN) is built and implemented for an empirical study. Standard benchmarks are used to evaluate the

network’s ability to make forecasts. The objective of this study is to conclude whether an ECN could be

sell signals, transaction costs

Neural networks can be applied to all sorts of financial problems, not only stock prediction tasks, also used

for forecasts of yield curves, exchange rates, bond rates etc. The principal motivation for the neural network

linear model is beneficial

A large set of interacting input series is often required to explain a specific stock, which suites neural

This approach is also applied for the prediction task from the angle of economics. Such as stock forecasting

ther than actual stock prices. Another reason is that it facilitates stabilization of the model

Documentation is made ignoring needed data and importance is given to the data irrelevant for the problem.

equently one has to handle missing data at recreant intervals or even data emerging discontinuous time series.

A convenient way to work around these difficulties is by letting the network accept missing data.

tems contain both an autonomous part and a part governed by external forces. Many times

relevant external forces may be hard to identify or data could be noisy. As a consequence a correct description of

better model description is to use the previous model

Following set of equations consisting of a state and output equations is a recurrent description of dynamic system

is very general form for discrete time grids. A basic dynamical recurrent system depicted in Fig. 1, at time t, it

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

can be expressed as follows:

),(s 1t tt usf −=

)(y t tsg=

The state transition is a mapping from the pervious internal hidden state of the system s

external inputs tu to the new state

Figure 1: A dynamical system with input u, output y and internal state s. The identification of dynamical system

using a discrete time description: Input

The system can be viewed as a partially observable autoregressive dynamic state

external disturbances tu . Without the external inputs the systems is called autonomous system[7].The

indentify the dynamic system of equations 1 & 2, can then be stated as task to find (parameterized) functions f,

g such that an average distanced measurement (Eq. 3 ) between the observed data

computed data ty of the model is minimal[5]

∑=

−←T

tty

1gf,

(T

1 min

Where T is the number of patterns (Patterns are data points in a time series). If we ignore the dependency on s

in the state transition equation by assuming

we are back in the frame work where the neural network approach without recurrence i.e. a feedforward neural

network, can solve the system identification task.

The Eq. 1 and 2 without external inputs

error at the previous time t-1 act as an additional input to the system. We get (

is the computed output and ts describes the state.)

,,( 1 ttt yusfs −=

European Journal of Business and Management

2839 (Online)

61

state transition (1)

output equation (2)

The state transition is a mapping from the pervious internal hidden state of the system s

ts . The output equation computes the observable output vector

Figure 1: A dynamical system with input u, output y and internal state s. The identification of dynamical system

using a discrete time description: Input kR∈tu hidden states

dR∈ts and output

The system can be viewed as a partially observable autoregressive dynamic state ts

Without the external inputs the systems is called autonomous system[7].The

indentify the dynamic system of equations 1 & 2, can then be stated as task to find (parameterized) functions f,

g such that an average distanced measurement (Eq. 3 ) between the observed data dt

y

of the model is minimal[5]

− dty 2) (3)

Where T is the number of patterns (Patterns are data points in a time series). If we ignore the dependency on s

in the state transition equation by assuming )(s t tuf= and )(y t tsg=

we are back in the frame work where the neural network approach without recurrence i.e. a feedforward neural

network, can solve the system identification task.

The Eq. 1 and 2 without external inputs tu would represent an autonomous system. Let the observed model

1 act as an additional input to the system. We get (dt

y denotes observed data,

describes the state.)

)11

d

tt yy −− − (4)

www.iiste.org

The state transition is a mapping from the pervious internal hidden state of the system st-1 and the influence of

. The output equation computes the observable output vector ty .

Figure 1: A dynamical system with input u, output y and internal state s. The identification of dynamical system

and output nR∈ty

which also driven by

Without the external inputs the systems is called autonomous system[7].The task of

indentify the dynamic system of equations 1 & 2, can then be stated as task to find (parameterized) functions f,

d, t=1,2,……T and the

Where T is the number of patterns (Patterns are data points in a time series). If we ignore the dependency on st-1

))(() tufg= ,

we are back in the frame work where the neural network approach without recurrence i.e. a feedforward neural

would represent an autonomous system. Let the observed model

denotes observed data, ty

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

)(y t tsg=

The identification task of Eq. 1, 2 & 3, can be implemented as a neural network (denoted

,,( 1 yuss ttt −ℵ=

);(y t ws tℵ=

By specifying the functions f and g as neural networks with parameter vectors v and w, we have transformed the

system identification task of Eq. 3 into a parameter optimization problem

∑=

−←T

tty

1wv,

(T

1 min

The dynamic system described in Eq. 9 & 10, can be model by Error Correction neural network architecture as

shown in fig 2. The weights are v = {A, B} and w = {C}. in general one may think of a matrix D instead of the

identity matrix (id) between the hidden layer

combines matrix A and D to a new matrix A’ between the input

ℵ ( ts ; w) is realized as a linear function. It is straight forward to show by using an augmented inner state vector

that this is nota functional restriction.

tanh( 1tt Ass −=

tCs=ty

The term 1−tCs recomputed the last output

transformation D is necessary in order to adjust different dimensionalities in the state transition equation.

Figure 2: The error correc

The identification of above model is ambiguously creating the numerical problem, because autoregressive

structures i.e. the dependency of ts

European Journal of Business and Management

2839 (Online)

62

(5)

The identification task of Eq. 1, 2 & 3, can be implemented as a neural network (denotedℵ

);11 vyy d

tt −− − (6)

(7)

By specifying the functions f and g as neural networks with parameter vectors v and w, we have transformed the

system identification task of Eq. 3 into a parameter optimization problem

− dty 2) (8)

cribed in Eq. 9 & 10, can be model by Error Correction neural network architecture as

shown in fig 2. The weights are v = {A, B} and w = {C}. in general one may think of a matrix D instead of the

identity matrix (id) between the hidden layer ts this is not necessary because for linear output layer

combines matrix A and D to a new matrix A’ between the input 1-ts and the hidden layer. The output equation

; w) is realized as a linear function. It is straight forward to show by using an augmented inner state vector

that this is nota functional restriction.

))( 111

d

ttt yCsDBu −− −++ (9)

(10)

recomputed the last output 1−ty and compares to the observed data

transformation D is necessary in order to adjust different dimensionalities in the state transition equation.

Figure 2: The error correction neural network.

The identification of above model is ambiguously creating the numerical problem, because autoregressive

on 1−ts , could either be coded in matrix A or DC. On this

www.iiste.org

ℵ ) then we get

By specifying the functions f and g as neural networks with parameter vectors v and w, we have transformed the

cribed in Eq. 9 & 10, can be model by Error Correction neural network architecture as

shown in fig 2. The weights are v = {A, B} and w = {C}. in general one may think of a matrix D instead of the

this is not necessary because for linear output layer ts can

and the hidden layer. The output equation

; w) is realized as a linear function. It is straight forward to show by using an augmented inner state vector

and compares to the observed datad

ty 1− . The matrix

transformation D is necessary in order to adjust different dimensionalities in the state transition equation.

The identification of above model is ambiguously creating the numerical problem, because autoregressive

either be coded in matrix A or DC. On this problem, one may

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

argue to transform Eq. 9, well define form (Eq. 11) utilizing

tanh( 1tt sAs −= &

Adding a non-linearity is a measure to avoid this problem [2]. This yield

tanh( 1tt BuAss − +=

tCs=ty

Distinguishing external inputs tu

denote the fixed identity matrix. As a consequence the target values of output clusters

,2,1,0z -t =ττ are zero. Only the difference between

offers forecasts based on the modelling of the recursive structure (matrix A), the external forces (matrix B)

the error correcting part (matrices C and D). The error correcting part can also be viewed as an external input

similar to tu .

2.2. Variants and Invariants

Predicting a high-dimensional dynamical system is difficult. A way of reducin

separate the dynamics into time variant and invariant structures. Let the system forecasts the variants then

eventually this will combine this forecast with the unchanged invariants. This can be done by connecting the

standard ECN (Fig. 2) to a compression

from invariants while matrix F reconstructs the dynamics. The actual forecasting is coded in G, [3].

Figure 3: Separation of variants and invarian

3. Training The Proposed Model

3.1. Backpropagation

After the description of the network structure, training set

overall objective in training is to minimise the discrepancy between real data and the output of t

This principle is referred to as supervised learning

or self-organized learning. One well known learning method in this class is the Self Organizing Map (SOM))

a step-by-step manner the error guides the network in the direction towards the target data. The backpropagation

algorithm belongs to this class and can be described as "an efficient way to calculate the partial derivatives of the

network error function with respect to the

3.2. Learning Algorithm

The backpropagation algorithm supplies information about the gradients. However, a learning rule that uses this

information to update the weights efficiently is also needed. A weight update from iteration k to k + 1 may l

like

kw .w 1k η+=+

Where dk describes the search direction and

European Journal of Business and Management

2839 (Online)

63

argue to transform Eq. 9, well define form (Eq. 11) utilizing DCAA +=&

))11

d

tt DyBu −++ (11)

linearity is a measure to avoid this problem [2]. This yield

))tanh() 11

d

ttt yCsDBu −− −+ (12)

(13)

affecting the state transition st from target inputs

denote the fixed identity matrix. As a consequence the target values of output clusters

are zero. Only the difference between ty and dt

y influences

offers forecasts based on the modelling of the recursive structure (matrix A), the external forces (matrix B)

the error correcting part (matrices C and D). The error correcting part can also be viewed as an external input

dimensional dynamical system is difficult. A way of reducing the complexity of the task is to

separate the dynamics into time variant and invariant structures. Let the system forecasts the variants then

eventually this will combine this forecast with the unchanged invariants. This can be done by connecting the

andard ECN (Fig. 2) to a compression-decompression network shown in Fig. 3. Matrix E separates variants

from invariants while matrix F reconstructs the dynamics. The actual forecasting is coded in G, [3].

Figure 3: Separation of variants and invariants

Training The Proposed Model

After the description of the network structure, training set-up has to be settled. As previously mentioned the

overall objective in training is to minimise the discrepancy between real data and the output of t

This principle is referred to as supervised learning (The other basic class of learning paradigms is unsupervised

organized learning. One well known learning method in this class is the Self Organizing Map (SOM))

ner the error guides the network in the direction towards the target data. The backpropagation

algorithm belongs to this class and can be described as "an efficient way to calculate the partial derivatives of the

network error function with respect to the weights" [2].

The backpropagation algorithm supplies information about the gradients. However, a learning rule that uses this

information to update the weights efficiently is also needed. A weight update from iteration k to k + 1 may l

kd. (14)

describes the search direction and η the learning rate (or step length). Issues that have to be

www.iiste.org

from target inputs dt

y is important. I

denote the fixed identity matrix. As a consequence the target values of output clusters

influences ts . The ECN

offers forecasts based on the modelling of the recursive structure (matrix A), the external forces (matrix B) and

the error correcting part (matrices C and D). The error correcting part can also be viewed as an external input

g the complexity of the task is to

separate the dynamics into time variant and invariant structures. Let the system forecasts the variants then

eventually this will combine this forecast with the unchanged invariants. This can be done by connecting the

decompression network shown in Fig. 3. Matrix E separates variants

from invariants while matrix F reconstructs the dynamics. The actual forecasting is coded in G, [3].

up has to be settled. As previously mentioned the

overall objective in training is to minimise the discrepancy between real data and the output of the network.

The other basic class of learning paradigms is unsupervised

organized learning. One well known learning method in this class is the Self Organizing Map (SOM)). In

ner the error guides the network in the direction towards the target data. The backpropagation

algorithm belongs to this class and can be described as "an efficient way to calculate the partial derivatives of the

The backpropagation algorithm supplies information about the gradients. However, a learning rule that uses this

information to update the weights efficiently is also needed. A weight update from iteration k to k + 1 may look

the learning rate (or step length). Issues that have to be

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

addressed are how to determine (i) the search direction, (ii) the learning rate and (ii

A familiar way of determining the search direction d

rule [4]. The major drawback though is that learning easily is caught in a local minimum. To avoid this hazar

the vario-eta algorithm can be chosen as learning rule [5, 6]. Basically it is a stochastic approximation of a

Quasi-Newton method. In the vario

arbitrary weight, the jth

weight-specific factor β is defined according to Eq. 15.

∑ = ∂∂

=N

t 1

j

(

∑= ∂

∂=N

tj

t

w

E

NE

1

1

Assuming there are p weights in the network. The search direction is det

component of the negative gradient with its weight

=00

0.0

00

dp

j

k

β

β

Above E denotes the error function and N the number of patterns. A benefit with the vario

increments kd.η become non-static. This property implies a potentially fast learning phase in [2]. Concerning

a reasonable value of the learning rate

Regarding pattern selection, a stochastic procedure can be used. This simply means that the gradient

of a subset M of all patterns are used as an approximation of the true gradient

∑∇=∇Mt

t

M EM

1

Where |M | denotes the number of elements of M. M is subset network

M can be composed in several ways. In our empirical study selection of M is made with equal probability, a

predefined number of patterns (less than 10 percent of the training

Eq. 17 was computed and used as input to Eq. 16. Once all weights were updated, out of the remaining training

patterns a new subset was picked for the next iteration. It is also noted that re

significant one, may prefer a non-uniform probability distribution, where it is more likely to choose a recent

pattern [6].

3.3. Cleaning

The dilemma of overfitting is deeply rooted in neural networks. One way to suppress overfitt

input data not to be exact (the case in the field of financial analysis). The total error of pattern t can be split into

two components associated with the weights and the erroneous input respectively. The corrected input data,

can be expressed as

ttt xxx ∆+=~

Where xt is the original data and

updated in parallel with the weights. To this end, the outp

using original and corrected input data has to be known which is only true for training data. Accordingly, the

European Journal of Business and Management

2839 (Online)

64

addressed are how to determine (i) the search direction, (ii) the learning rate and (iii) which patterns to include.

A familiar way of determining the search direction dk is to apply the gradient descent which is a relatively simple

rule [4]. The major drawback though is that learning easily is caught in a local minimum. To avoid this hazar

eta algorithm can be chosen as learning rule [5, 6]. Basically it is a stochastic approximation of a

Newton method. In the vario-eta algorithm a weight-specific factor, β, is related to each weight. For an

specific factor β is defined according to Eq. 15.

−∂j

t Ew

E 2)

1 (15)

Where E denotes the E average and E is error function

Assuming there are p weights in the network. The search direction is determined by multiplying each

component of the negative gradient with its weight-specific factor,

E∇

. (16)

Above E denotes the error function and N the number of patterns. A benefit with the vario

static. This property implies a potentially fast learning phase in [2]. Concerning

a reasonable value of the learning rateη . The learning rate is many times determined on an ad hoc basis.

ern selection, a stochastic procedure can be used. This simply means that the gradient

of a subset M of all patterns are used as an approximation of the true gradient ∇ E according to

(3.17)

Where |M | denotes the number of elements of M. M is subset network

M can be composed in several ways. In our empirical study selection of M is made with equal probability, a

predefined number of patterns (less than 10 percent of the training data) to represent M. The gradient

Eq. 17 was computed and used as input to Eq. 16. Once all weights were updated, out of the remaining training

patterns a new subset was picked for the next iteration. It is also noted that recent patterns are considered

uniform probability distribution, where it is more likely to choose a recent

The dilemma of overfitting is deeply rooted in neural networks. One way to suppress overfitt

input data not to be exact (the case in the field of financial analysis). The total error of pattern t can be split into

two components associated with the weights and the erroneous input respectively. The corrected input data,

(18)

tx∆ a correction vector. During training the correction vector must be

updated in parallel with the weights. To this end, the output target difference i.e. the difference in output from

using original and corrected input data has to be known which is only true for training data. Accordingly, the

www.iiste.org

i) which patterns to include.

is to apply the gradient descent which is a relatively simple

rule [4]. The major drawback though is that learning easily is caught in a local minimum. To avoid this hazard

eta algorithm can be chosen as learning rule [5, 6]. Basically it is a stochastic approximation of a

specific factor, β, is related to each weight. For an

denotes the E average and E is error function

ermined by multiplying each

Above E denotes the error function and N the number of patterns. A benefit with the vario-eta rule is that weight

static. This property implies a potentially fast learning phase in [2]. Concerning

. The learning rate is many times determined on an ad hoc basis.

ern selection, a stochastic procedure can be used. This simply means that the gradient ∇ EM

E according to

M can be composed in several ways. In our empirical study selection of M is made with equal probability, a

data) to represent M. The gradient ∇ EM

of

Eq. 17 was computed and used as input to Eq. 16. Once all weights were updated, out of the remaining training

cent patterns are considered

uniform probability distribution, where it is more likely to choose a recent

The dilemma of overfitting is deeply rooted in neural networks. One way to suppress overfitting is to assume

input data not to be exact (the case in the field of financial analysis). The total error of pattern t can be split into

two components associated with the weights and the erroneous input respectively. The corrected input data, tx~ ,

a correction vector. During training the correction vector must be

ut target difference i.e. the difference in output from

using original and corrected input data has to be known which is only true for training data. Accordingly, the

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

model might be optimised for training data but not for generalisation data because the la

distribution and an unknown output target difference. To work around this disadvantage the model

tx~ is composed according to

δ−∆+= ttt xxx~

δ is exactly one element drawn at random from {

training data). A composition with an additional noise term (Eq. 19) benefits from distribution properties which

is desirable for generalisation, [5].

The input modification "cleaning with noise" described above helps the network to concentrate on broader

structures in data. With this extent the model is prevented from establishing false causalities with considerable

effect.

3.4. Stopping Criteria

It is also underling to study the number of epochs (An epoch is completed when all training patterns have been

read in exactly once) are needed for a network to be trained. Mainly two paradigms exist, late and early

stopping. Late stopping means that the network is trained until a minimum error on the training set is reached, i.e.

the network is clearly overfitted. Then different techniques are used to exterminate nodes in the network (known

as pruning). By doing so eventually good generalisation

The concept of early stopping is a way of avoiding overfitting. During learning the progression is monitored and

training is terminated as soon as signs of overfitting appear. A clear advantage with early stopping is that the

time of training is relatively short. On the downside it is hard to know when to stop.

3.5. Error Function

When modelling one also has to be aware of outliers in data. Outliers typically appear when the economic or

political climate is unstable or unexpected informati

function the impact of outliers can be restrained.

)))(ln(cosh(1

ii toaa

The ln(cosh(·)) error function is a smooth approximation of the absolute error function |o

response from output neuron i and t

financial applications [5].

Compared to a quadratic error function the ln(cosh) error function has a reminiscent behaviour in

around zero but not for large positive or negative values as we see in Fig. 4. The advantage with this function

when modelling financial data is that a large difference in output and target yields a limited and

more reasonable error.

Figure

European Journal of Business and Management

2839 (Online)

65

model might be optimised for training data but not for generalisation data because the latter has a different noise

distribution and an unknown output target difference. To work around this disadvantage the model

(19)

δ is exactly one element drawn at random from { ix∆ i = 1, . . . , T } (memorised correction vectors from

training data). A composition with an additional noise term (Eq. 19) benefits from distribution properties which

The input modification "cleaning with noise" described above helps the network to concentrate on broader

structures in data. With this extent the model is prevented from establishing false causalities with considerable

It is also underling to study the number of epochs (An epoch is completed when all training patterns have been

read in exactly once) are needed for a network to be trained. Mainly two paradigms exist, late and early

that the network is trained until a minimum error on the training set is reached, i.e.

the network is clearly overfitted. Then different techniques are used to exterminate nodes in the network (known

as pruning). By doing so eventually good generalisation ability is reached.

The concept of early stopping is a way of avoiding overfitting. During learning the progression is monitored and

training is terminated as soon as signs of overfitting appear. A clear advantage with early stopping is that the

training is relatively short. On the downside it is hard to know when to stop.

When modelling one also has to be aware of outliers in data. Outliers typically appear when the economic or

political climate is unstable or unexpected information enters the market. By picking an appropriate error

function the impact of outliers can be restrained.

(20)

The ln(cosh(·)) error function is a smooth approximation of the absolute error function |o

response from output neuron i and ti the corresponding target. a ∈ [3, 4] has proven to be suitable for

Compared to a quadratic error function the ln(cosh) error function has a reminiscent behaviour in

around zero but not for large positive or negative values as we see in Fig. 4. The advantage with this function

when modelling financial data is that a large difference in output and target yields a limited and

Figure 4: The lncosh error function of Eq. 20, a = 3.

www.iiste.org

tter has a different noise

distribution and an unknown output target difference. To work around this disadvantage the model

i = 1, . . . , T } (memorised correction vectors from

training data). A composition with an additional noise term (Eq. 19) benefits from distribution properties which

The input modification "cleaning with noise" described above helps the network to concentrate on broader

structures in data. With this extent the model is prevented from establishing false causalities with considerable

It is also underling to study the number of epochs (An epoch is completed when all training patterns have been

read in exactly once) are needed for a network to be trained. Mainly two paradigms exist, late and early

that the network is trained until a minimum error on the training set is reached, i.e.

the network is clearly overfitted. Then different techniques are used to exterminate nodes in the network (known

The concept of early stopping is a way of avoiding overfitting. During learning the progression is monitored and

training is terminated as soon as signs of overfitting appear. A clear advantage with early stopping is that the

When modelling one also has to be aware of outliers in data. Outliers typically appear when the economic or

on enters the market. By picking an appropriate error

The ln(cosh(·)) error function is a smooth approximation of the absolute error function |oi − ti|. oi denotes the

[3, 4] has proven to be suitable for

Compared to a quadratic error function the ln(cosh) error function has a reminiscent behaviour in the region

around zero but not for large positive or negative values as we see in Fig. 4. The advantage with this function

when modelling financial data is that a large difference in output and target yields a limited and

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

4. Evaluation of the Study

A crucial part of financial forecasting is the evaluation of the prediction algorithm. Several performance

measures are widely used, but a performance measure in itself is not suffici

Relevant benchmarks are also needed. A benchmark is basically a different prediction algorithm used for

comparison. In this study, the quality of a prediction algorithm is judged with the help of existing offered data.

4.1. Benchmarks

Any prediction algorithm claiming to be successful should outperform the naive predictor defined as

tt yy =+1ˆ

Where ty is the current value of the stock and ŷ

21 states that the most intelligent suggestion of tomorrow’s price is today’s price which is a direct consequence

of the Efficient Market Hypothesis (EM

assimilation of all information available. Therefore no prediction of future changes in the price can be made

given this information). In the empirical study a comparison to the naive predictio

definition is

tt RR =+1ˆ

Where Rt is the last known return and

incorporated buy and sell signals it could be useful to

strategy expresses the profit made when making an investment at the start of a time period and selling n

time-steps into the future, i.e.

t

tnt

by

yyR

−= +.100

A comparison to the buy-and-hold strategy simply gives an indication of the quality of the signals. Is it more

profitable to be a "passive" investor?

4.2. Performance Measures

Three frequently used measures, namely hit rate, return on investment and realised potential, and are defined

below. Begin with the hit rate HR.

,0ˆ{

,0ˆ{

RRt

RRtH

k

t

k

t

k

t

k

t

R

>=

Where k

t

k

t RR ˆ is the actual (predicted) k

)/()( ktktt

k

t yyyR −−−= ). The norm simply is the number of eleme

often the algorithm produces a correct prediction. In this context a prediction is correct when the direction of the

stock k time-steps into the future is successfully predicted. The return on investment ROI takes i

not only the correctness of the sign, but also the quantity of the actual return. The definition is

∑=

=T

t

tt RsignRROI1

)ˆ(.

Finally, in Eq. 26 the definition of realised potential RP is given.

European Journal of Business and Management

2839 (Online)

66

A crucial part of financial forecasting is the evaluation of the prediction algorithm. Several performance

measures are widely used, but a performance measure in itself is not sufficient for a satisfying evaluation.

Relevant benchmarks are also needed. A benchmark is basically a different prediction algorithm used for

comparison. In this study, the quality of a prediction algorithm is judged with the help of existing offered data.

Any prediction algorithm claiming to be successful should outperform the naive predictor defined as

(21)

is the current value of the stock and ŷt+1 the predicted value one time-step into the future. The Eq.

21 states that the most intelligent suggestion of tomorrow’s price is today’s price which is a direct consequence

of the Efficient Market Hypothesis (EMH), (The EMH states that the current market price reflects the

assimilation of all information available. Therefore no prediction of future changes in the price can be made

. In the empirical study a comparison to the naive prediction of returns was made. The

(22)

is the last known return and 1ˆ

+tR the predicted one-step return. For a prediction algorithm with

incorporated buy and sell signals it could be useful to do a comparison with the buy-and

strategy expresses the profit made when making an investment at the start of a time period and selling n

(23)

hold strategy simply gives an indication of the quality of the signals. Is it more

profitable to be a "passive" investor?

Three frequently used measures, namely hit rate, return on investment and realised potential, and are defined

},......,

},......,

Nt

Nt (24)

is the actual (predicted) k-step return at time t. (According to the definition the k

The norm simply is the number of elements in the series. Eq. 24 indicates how

often the algorithm produces a correct prediction. In this context a prediction is correct when the direction of the

steps into the future is successfully predicted. The return on investment ROI takes i

not only the correctness of the sign, but also the quantity of the actual return. The definition is

(25)

Finally, in Eq. 26 the definition of realised potential RP is given.

www.iiste.org

A crucial part of financial forecasting is the evaluation of the prediction algorithm. Several performance

ent for a satisfying evaluation.

Relevant benchmarks are also needed. A benchmark is basically a different prediction algorithm used for

comparison. In this study, the quality of a prediction algorithm is judged with the help of existing offered data.

Any prediction algorithm claiming to be successful should outperform the naive predictor defined as

step into the future. The Eq.

21 states that the most intelligent suggestion of tomorrow’s price is today’s price which is a direct consequence

The EMH states that the current market price reflects the

assimilation of all information available. Therefore no prediction of future changes in the price can be made

n of returns was made. The

step return. For a prediction algorithm with

and-hold return Rb. This

strategy expresses the profit made when making an investment at the start of a time period and selling n

hold strategy simply gives an indication of the quality of the signals. Is it more

Three frequently used measures, namely hit rate, return on investment and realised potential, and are defined

step return at time t. (According to the definition the k-step return

nts in the series. Eq. 24 indicates how

often the algorithm produces a correct prediction. In this context a prediction is correct when the direction of the

steps into the future is successfully predicted. The return on investment ROI takes into account

not only the correctness of the sign, but also the quantity of the actual return. The definition is

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

∑∑

=

==T

t t

T

t t

R

RsignRRP

1

1(.

The realised potential states how large part of the total movement (upwards and downwards) the prediction

algorithm successfully identifies.

5. Empirical Study: Forecasting the INR/USD Fx

In the empirical study well traded stocks with a reasonable spread were

spread may be fatal if one wishes to incorporate the predictions in real trading. Hence a decision is further to

make one-day predictions (using daily data) of the Bombay stock exchange (BSE), AxisBank B and Cipla B

with a standard ECN. Also one-week predictions (using weekly data) of AxisBank B and Cipla B were

performed on an ECN separating variants and invariants. Basically, one assumes that certain time invariant

structures can be identified and learned quickly. This

with some weights frozen. The occurrence of invariant structures could prove to be more evident in a weekly

compared to a daily model, because patterns originate from the same day of the week in a w

5.1. Data Series

For all predictions the following four time series were used as raw input:

• Closing price y (price of the last fulfilled trade during the day)

• Highest price paid during the day, y

• Lowest price paid during the day, y

• Volume V (total amount of traded stocks during the day)

Additionally, external time series served as input. Table 1 gives a summary of time series considered to have a

significant impact on the behaviour of the AxisBank B and Cipla B.

another set of external inputs was used. Table 2 gives the full list. All data used in the modeling was

acquired from the daily service provider E

Table 1: external time series used predicting AxisBank B and Ci

SENSEX

6 months interest rate Bombay

Bombay INR/USD FX - rate

Table 2: external time series used predicting the BSE.

BSE- IT (Infotech)

BSE-BANK

6 months interest rate Bombay

SENSEX

Bombay INR/USD FX - rate

5.2. Technical Considerations

Prior to each training session relevant data series were transformed and preprocessed in d

external time series (Tab. 1 and 2) we calculated the normalized one

volume was computed in a running window of 30 days and 12 weeks for daily and weekly predictions

respectively. On inputs y, yH and yL, we was applied the log

changes the log-return is similar to the one

European Journal of Business and Management

2839 (Online)

67

tR )ˆ (26)

lised potential states how large part of the total movement (upwards and downwards) the prediction

Empirical Study: Forecasting the INR/USD Fx-Rate

In the empirical study well traded stocks with a reasonable spread were considered. Mainly because a large

spread may be fatal if one wishes to incorporate the predictions in real trading. Hence a decision is further to

day predictions (using daily data) of the Bombay stock exchange (BSE), AxisBank B and Cipla B

week predictions (using weekly data) of AxisBank B and Cipla B were

performed on an ECN separating variants and invariants. Basically, one assumes that certain time invariant

structures can be identified and learned quickly. This means that the latter part of the training is performed

with some weights frozen. The occurrence of invariant structures could prove to be more evident in a weekly

compared to a daily model, because patterns originate from the same day of the week in a w

For all predictions the following four time series were used as raw input:

Closing price y (price of the last fulfilled trade during the day)

Highest price paid during the day, yH

Lowest price paid during the day, yL

(total amount of traded stocks during the day)

Additionally, external time series served as input. Table 1 gives a summary of time series considered to have a

significant impact on the behaviour of the AxisBank B and Cipla B. Regarding the BSE predict

another set of external inputs was used. Table 2 gives the full list. All data used in the modeling was

acquired from the daily service provider E-Eighteen.com and yahoo finance

Table 1: external time series used predicting AxisBank B and Cipla B.

Stock Model Input

NIFTY

1 year interest rate Bombay

Bombay INR/USD FX - rate

Table 2: external time series used predicting the BSE.

Index Model Input

BSE- AUTO

BSE-OILGAS

1 year interest rate Bombay

NIFTY

Bombay INR/USD FX - rate

Prior to each training session relevant data series were transformed and preprocessed in d

external time series (Tab. 1 and 2) we calculated the normalized one-step return Rt is calculated. The Gaussian

volume was computed in a running window of 30 days and 12 weeks for daily and weekly predictions

, we was applied the log-return. The definition is given in Eq. 27. For small

return is similar to the one-step return Rt.

www.iiste.org

lised potential states how large part of the total movement (upwards and downwards) the prediction

considered. Mainly because a large

spread may be fatal if one wishes to incorporate the predictions in real trading. Hence a decision is further to

day predictions (using daily data) of the Bombay stock exchange (BSE), AxisBank B and Cipla B

week predictions (using weekly data) of AxisBank B and Cipla B were

performed on an ECN separating variants and invariants. Basically, one assumes that certain time invariant

means that the latter part of the training is performed

with some weights frozen. The occurrence of invariant structures could prove to be more evident in a weekly

compared to a daily model, because patterns originate from the same day of the week in a weekly model.

Additionally, external time series served as input. Table 1 gives a summary of time series considered to have a

Regarding the BSE prediction,

another set of external inputs was used. Table 2 gives the full list. All data used in the modeling was

Prior to each training session relevant data series were transformed and preprocessed in different ways. For all

is calculated. The Gaussian

volume was computed in a running window of 30 days and 12 weeks for daily and weekly predictions

return. The definition is given in Eq. 27. For small

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

1

(log)+

=t

t

y

yR

Data were also divided into three subsets: a training set, a valida

all patterns available were used for training and one

generalisation period ran over 12 months for both the daily and the weekly model. (In the daily mo

training sessions were performed on a shorter generalisation set due to missing data.). As error function we used

the ln(cosh) function of Eq. 20 with a=3 is used. The standard tanh(·) function served as activation function.

5.3. Training Procedure

Weights were initialized randomly (uniformly distributed in the interval [−1, 1]). The ECN was trained (on the

training set) for 500 and 1000 epochs in the daily model and weekly model respectively. It is noticed that this

was enough for the cleaning error (correction to the net inputs) to stabilize. After training, weights associated

with the best performance (in terms of hit rate) on the validation set were selected and applied to the

generalisation set to get the final results.

5.4. Implementation

Neuro Dimension provide a software for neural computing, NSNN (Neuro Solutions for Neurel Networks)is

designed to build ANN, based on which mimic the learning process of the brain in order to extract patterns

from historical data. Once these files are loaded into

6. Experimental Results

In the following sections, results from the empirical study are presented. Tables are based on generalisation data.

The naive predictions of returns constitute benchmark for each predi

realised potential (RP) using the ECN and the naive strategy are given in tables below.

6.1. Daily Predictions

Table 3: Daily predictions of BSE. Hit rate and realised potential using the ECN and the naive predictor.

Period

Jan.05 to Dec. 05

Jan. 06 to Dec. 06

Jan. 07 to Dec. 07

Jan. 08 to Dec. 08

Jan. 09 to Dec. 09

Jan. 10 to Dec. 10

Jan. 11 to Dec. 11

Jan. 12 to Jun. 12

Mean

Table 4: Daily predictions of Axis Bank B. Hit rate and realized potential using the ECN and the naive predictor

Period

Jan.05 to Dec. 05

Jan. 06 to Dec. 06

Jan. 07 to Dec. 07

Jan. 08 to Dec. 08

Jan. 09 to Dec. 09

Jan. 10 to Dec. 10

Jan. 11 to Dec. 11

Jan. 12 to Jun. 12

Mean

European Journal of Business and Management

2839 (Online)

68

(27)

Data were also divided into three subsets: a training set, a validation set and a generalisation set. Roughly, half of

all patterns available were used for training and one-quarter each for validation and generalisation. The

generalisation period ran over 12 months for both the daily and the weekly model. (In the daily mo

training sessions were performed on a shorter generalisation set due to missing data.). As error function we used

the ln(cosh) function of Eq. 20 with a=3 is used. The standard tanh(·) function served as activation function.

eights were initialized randomly (uniformly distributed in the interval [−1, 1]). The ECN was trained (on the

training set) for 500 and 1000 epochs in the daily model and weekly model respectively. It is noticed that this

(correction to the net inputs) to stabilize. After training, weights associated

with the best performance (in terms of hit rate) on the validation set were selected and applied to the

generalisation set to get the final results.

nsion provide a software for neural computing, NSNN (Neuro Solutions for Neurel Networks)is

designed to build ANN, based on which mimic the learning process of the brain in order to extract patterns

from historical data. Once these files are loaded into the software it is possible to start training.

In the following sections, results from the empirical study are presented. Tables are based on generalisation data.

The naive predictions of returns constitute benchmark for each prediction model. Values of hit rate (HR) and

realised potential (RP) using the ECN and the naive strategy are given in tables below.

Table 3: Daily predictions of BSE. Hit rate and realised potential using the ECN and the naive predictor.

1-Day forecast (%)

HR ECN RP ECN HR navie

52.4 10.8 54.7

59.1 32.4 52.8

54.5 11 51.4

57.9 24.5 50

60.1 26.1 53.9

57.9 24.4 53.5

56.3 17.8 52

56.2 21 48.5

56.8 21 52.1

Table 4: Daily predictions of Axis Bank B. Hit rate and realized potential using the ECN and the naive predictor

1-Day forecast (%)

HR ECN RP ECN HR navie

50.7 5.1 42.4

51.2 9.3 41.7

56.3 30.8 43.3

51.2 15.2 39.8

55.7 17.8 48.2

48 0 46.5

53.5 3.7 40.7

47.6 10.1 48

51.8 11.5 43.8

www.iiste.org

tion set and a generalisation set. Roughly, half of

quarter each for validation and generalisation. The

generalisation period ran over 12 months for both the daily and the weekly model. (In the daily model some

training sessions were performed on a shorter generalisation set due to missing data.). As error function we used

the ln(cosh) function of Eq. 20 with a=3 is used. The standard tanh(·) function served as activation function.

eights were initialized randomly (uniformly distributed in the interval [−1, 1]). The ECN was trained (on the

training set) for 500 and 1000 epochs in the daily model and weekly model respectively. It is noticed that this

(correction to the net inputs) to stabilize. After training, weights associated

with the best performance (in terms of hit rate) on the validation set were selected and applied to the

nsion provide a software for neural computing, NSNN (Neuro Solutions for Neurel Networks)is

designed to build ANN, based on which mimic the learning process of the brain in order to extract patterns

the software it is possible to start training.

In the following sections, results from the empirical study are presented. Tables are based on generalisation data.

ction model. Values of hit rate (HR) and

Table 3: Daily predictions of BSE. Hit rate and realised potential using the ECN and the naive predictor.

RP navie

22.3

12.5

8.9

1.7

12

18.3

15.2

4.3

11.9

Table 4: Daily predictions of Axis Bank B. Hit rate and realized potential using the ECN and the naive predictor.

RP navie

-5.8

2.1

-7.3

-6.4

4.5

-0.4

-0.5

22.6

1.1

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

Table 5: Daily predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predict

Period

Jan.05 to Dec. 05

Jan. 06 to Dec. 06

Jan. 07 to Dec. 07

Jan. 08 to Dec. 08

Jan. 09 to Dec. 09

Jan. 10 to Dec. 10

Jan. 11 to Dec. 11

Jan. 12 to Jun. 12

Mean

Weekly Predictions

Table 6: Weekly predictions of Axis bank B. Hit rate and realised potential using th

predictor.

Period

Jan. 06 to Dec. 06

Jan. 07 to Dec. 07

Jan. 08 to Dec. 08

Jan. 09 to Dec. 09

Jan. 10 to Dec. 10

Jan. 11 to Dec. 11

Jan. 12 to Jun. 12

Mean

Table 7: Weekly predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predictor.

Period

Jan. 06 to Dec. 06

Jan. 07 to Dec. 07

Jan. 08 to Dec. 08

Jan. 09 to Dec. 09

Jan. 10 to Dec. 10

Jan. 11 to Dec. 11

Jan. 12 to Jun. 12

Mean

European Journal of Business and Management

2839 (Online)

69

Table 5: Daily predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predict

1-Day forecast(%)

HR ECN RP ECN HR navie

44.9 9.7 41.3

44.1 -2.5 47.6

52.4 21 45.7

46.5 13.8 39

44.5 1.3 42.5

47.8 15 45.8

44.9 7.8 42.9

40.3 -2.5 37.9

45.7 8 42.8

Table 6: Weekly predictions of Axis bank B. Hit rate and realised potential using th

1-week forecast (%)

HR ECNN RP ECNN HR navie

60 25.9 44.4

48.9 22.1 44.4

47.8 -8 50

62.2 25.7 55.6

51.1 -0.9 60

57.8 -6.5 55.6

42.2 -12.9 42.2

52.9 6.5 50.3

Table 7: Weekly predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predictor.

1-week forecast (%)

HR ECN RP ECN HR navie

66.7 47.7 46.7

60 32.2 44.4

51.1 21.3 40

45.7 13.3 45.7

57.8 33.4 42.4

60 18 48.9

55.6 15.7 60

56.7 25.9 46.8

www.iiste.org

Table 5: Daily predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predictor.

RP navie

12.8

18.4

22.3

5.4

5.4

12.2

10.6

1.5

11.1

Table 6: Weekly predictions of Axis bank B. Hit rate and realised potential using the ECN and the naive

HR navie RP

44.4 -18.8

44.4 0.3

18.9

55.6 -19

2.9

55.6 -1

42.2 -34.1

50.3 -7.3

Table 7: Weekly predictions of Cipla B. Hit rate and realised potential using the ECN and the naive predictor.

RP navie

12.7

-33

-9.7

-17.8

-17.8

-0.5

12.5

-7.7

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

7. Conclusion

The most crucial issue of stock prediction is how to get stability over time in this respect neither the daily nor the

weekly model is not optimized and refinements have to be made. Nevertheless, there is no doubt about neural

networks potential in a trading environment. As already exhibited in previous section, the ECN occasionally

shows good results.

Intuitively, one may think that the weight i

Theoretically, this should not be the case though due to the stochastisity in the selection of patterns during

training. To confirm this notion a training scheme where was set up the net t

of hit rate) generalisation weights from the previous training period. The results gave indications of an even

negative impact (on the hit rate) using a biased initialization compared to a random one. This phenomenon

typically illustrates the challenge faced when trying to validate a dynamical system for a longer period of time.

Previously gained information about the system may be hard to utilize successfully in future time perspective.

A further development is to c

relevance of a possible detected trend a weighted sum could be calculated, where weights are associated with the

one-day forecast, two-day forecast and so on in descending order. Eventu

assessment of whether the trend id likely to turn out as predicted or not. Moreover, the forecasts of a committee

of models could be averaged. This will bring about a reduction of the forecast variance, thus a more re

output

References

[1] Hellström T. A Random Walk through the Stock Market, Licentiate Thesis, Department of Computing

Science, Ume a University, Sweden, 1998.

[2] Grothmann R. Multi-Agent Market Modeli ng based on Neural Net

Economics, University of Bremen, Germany, 2002.

[3] Zimmermann H.-G. System Identification & Forecasting by Neural Networks. Principles, Techniques,

Applications, Compilation of material presented to a conference held at Mälardalen University, September

2003.

[4] Heath M.T. Scientific Computing

[5] Neuneier R. and Zimmermann H.

Neural Networks: Tricks of the Trade,

[6] NSNN User Manual. Version 6.1, Neuro Dimensions, NewYork 2012.

http://www.neurosolutions.com/products/ns/

[7] Haykin S. Neural Networks. A Comprehensive Foundation, Macmillan College Publishi

1994.

Author:

R. Lakshman Naik received his B.Tech. (Electronics and Communication Engineering) from Sree Sarathi

Institute of Engineering and Technology, Jawaharlal Nehru Technological University (JNTUH), Krishna,

Andhra Pradesh, India and M.Tech.(Computer Science and Engineering) from Balaji Institute of Technology and

Sciences, JNTUH, Warangal, A.P, India. He served as a Systems Engineer in Beteel Business Solution and

Wipro Technologies, Hyderabad. Now he is working as Senior Systems E

a member of IAENG, IAEME, AIRCC and various organizations. He has Publications more than 10 papers in

area of computer networks, Data mining and Artificial neural networks.

B. Manjula received her BCA (Computer Appl

M.Sc. (Information System) from Osmania University, Hyderabad, A.P, India. She served as an Academic

Consultant in Department of Computer Science, Osmania Univesrsity, Hyderabad, A.P., India. And

working as Assistant professor in Department of Informatics, Kakatiya University, Warangal, A.P, India. She is

a Member of IAENG and IAEME. She has Publications more than 11 papers in area of computer networks, Data

mining and neural networks.

D. Ramesh received his B.Tech. (Computer Science and Engineering) from Srineedhi Institute of Science and

Technology, JNTU, Hyderabad, A.P, India and M.Tech. (Computer Science) from School of Information

Technology, JNTU, Hyderabad, A.P, India. He served a

India. Now he is working as Assistant Professor in Department of Informatics, Kakatiya University, Warangal,

European Journal of Business and Management

2839 (Online)

70

The most crucial issue of stock prediction is how to get stability over time in this respect neither the daily nor the

ized and refinements have to be made. Nevertheless, there is no doubt about neural

networks potential in a trading environment. As already exhibited in previous section, the ECN occasionally

Intuitively, one may think that the weight initialization phase is decisive for the outcome of the predictions.

Theoretically, this should not be the case though due to the stochastisity in the selection of patterns during

training. To confirm this notion a training scheme where was set up the net to initialized with the "best" (in terms

of hit rate) generalisation weights from the previous training period. The results gave indications of an even

negative impact (on the hit rate) using a biased initialization compared to a random one. This phenomenon

typically illustrates the challenge faced when trying to validate a dynamical system for a longer period of time.

Previously gained information about the system may be hard to utilize successfully in future time perspective.

A further development is to combine one-day predictions with multi-day predictions. To estimate the

relevance of a possible detected trend a weighted sum could be calculated, where weights are associated with the

day forecast and so on in descending order. Eventually, the aggregate is used to make an

assessment of whether the trend id likely to turn out as predicted or not. Moreover, the forecasts of a committee

of models could be averaged. This will bring about a reduction of the forecast variance, thus a more re

Hellström T. A Random Walk through the Stock Market, Licentiate Thesis, Department of Computing

Science, Ume a University, Sweden, 1998.

Agent Market Modeli ng based on Neural Net works, Ph.D. Thesis, Faculty o

Economics, University of Bremen, Germany, 2002.

G. System Identification & Forecasting by Neural Networks. Principles, Techniques,

Applications, Compilation of material presented to a conference held at Mälardalen University, September

Heath M.T. Scientific Computing- An Introductory Survey, McGraw Hill, New York, 1997.

Neuneier R. and Zimmermann H.-G. How to Train Neural Networks. In Orr G.B. and Müller K.

Neural Networks: Tricks of the Trade, pp. 373-423, Springer, Berlin, 1998.

NSNN User Manual. Version 6.1, Neuro Dimensions, NewYork 2012.

http://www.neurosolutions.com/products/ns/

Haykin S. Neural Networks. A Comprehensive Foundation, Macmillan College Publishi

received his B.Tech. (Electronics and Communication Engineering) from Sree Sarathi

Institute of Engineering and Technology, Jawaharlal Nehru Technological University (JNTUH), Krishna,

nd M.Tech.(Computer Science and Engineering) from Balaji Institute of Technology and

Sciences, JNTUH, Warangal, A.P, India. He served as a Systems Engineer in Beteel Business Solution and

Wipro Technologies, Hyderabad. Now he is working as Senior Systems Engineer in Unit4System Pvt. Ltd. He is

a member of IAENG, IAEME, AIRCC and various organizations. He has Publications more than 10 papers in

area of computer networks, Data mining and Artificial neural networks.

received her BCA (Computer Application) from Osmania University, Hyderabad, A.P, India and

M.Sc. (Information System) from Osmania University, Hyderabad, A.P, India. She served as an Academic

Consultant in Department of Computer Science, Osmania Univesrsity, Hyderabad, A.P., India. And

working as Assistant professor in Department of Informatics, Kakatiya University, Warangal, A.P, India. She is

a Member of IAENG and IAEME. She has Publications more than 11 papers in area of computer networks, Data

received his B.Tech. (Computer Science and Engineering) from Srineedhi Institute of Science and

Technology, JNTU, Hyderabad, A.P, India and M.Tech. (Computer Science) from School of Information

Technology, JNTU, Hyderabad, A.P, India. He served as an Academic Consultant in JNTU, Hyderabad, A.P,

India. Now he is working as Assistant Professor in Department of Informatics, Kakatiya University, Warangal,

www.iiste.org

The most crucial issue of stock prediction is how to get stability over time in this respect neither the daily nor the

ized and refinements have to be made. Nevertheless, there is no doubt about neural

networks potential in a trading environment. As already exhibited in previous section, the ECN occasionally

nitialization phase is decisive for the outcome of the predictions.

Theoretically, this should not be the case though due to the stochastisity in the selection of patterns during

o initialized with the "best" (in terms

of hit rate) generalisation weights from the previous training period. The results gave indications of an even

negative impact (on the hit rate) using a biased initialization compared to a random one. This phenomenon

typically illustrates the challenge faced when trying to validate a dynamical system for a longer period of time.

Previously gained information about the system may be hard to utilize successfully in future time perspective.

day predictions. To estimate the

relevance of a possible detected trend a weighted sum could be calculated, where weights are associated with the

ally, the aggregate is used to make an

assessment of whether the trend id likely to turn out as predicted or not. Moreover, the forecasts of a committee

of models could be averaged. This will bring about a reduction of the forecast variance, thus a more reliable

Hellström T. A Random Walk through the Stock Market, Licentiate Thesis, Department of Computing

works, Ph.D. Thesis, Faculty of

G. System Identification & Forecasting by Neural Networks. Principles, Techniques,

Applications, Compilation of material presented to a conference held at Mälardalen University, September

An Introductory Survey, McGraw Hill, New York, 1997.

G. How to Train Neural Networks. In Orr G.B. and Müller K.-R. (Eds.)

NSNN User Manual. Version 6.1, Neuro Dimensions, NewYork 2012.

Haykin S. Neural Networks. A Comprehensive Foundation, Macmillan College Publishing, New York,

received his B.Tech. (Electronics and Communication Engineering) from Sree Sarathi

Institute of Engineering and Technology, Jawaharlal Nehru Technological University (JNTUH), Krishna,

nd M.Tech.(Computer Science and Engineering) from Balaji Institute of Technology and

Sciences, JNTUH, Warangal, A.P, India. He served as a Systems Engineer in Beteel Business Solution and

ngineer in Unit4System Pvt. Ltd. He is

a member of IAENG, IAEME, AIRCC and various organizations. He has Publications more than 10 papers in

ication) from Osmania University, Hyderabad, A.P, India and

M.Sc. (Information System) from Osmania University, Hyderabad, A.P, India. She served as an Academic

Consultant in Department of Computer Science, Osmania Univesrsity, Hyderabad, A.P., India. And now she is

working as Assistant professor in Department of Informatics, Kakatiya University, Warangal, A.P, India. She is

a Member of IAENG and IAEME. She has Publications more than 11 papers in area of computer networks, Data

received his B.Tech. (Computer Science and Engineering) from Srineedhi Institute of Science and

Technology, JNTU, Hyderabad, A.P, India and M.Tech. (Computer Science) from School of Information

s an Academic Consultant in JNTU, Hyderabad, A.P,

India. Now he is working as Assistant Professor in Department of Informatics, Kakatiya University, Warangal,

European Journal of Business and Management

ISSN 2222-1905 (Paper) ISSN 2222-2839 (Online)Vol 4, No.12, 2012

A.P, India. He has Publications more than 4 papers in area of Data mining and neural networks.

B. Sridhara Murthy received his MCA (Computer Application) from Osmania University, Hyderabad, A.P,

India and M.Tech (Computer Science and Engineering) from Balaji Institute of Technology and Sciences,

JNTUH, Warangal, A.P, India. He served as an Assistant

Technology and Sciences, Warangal, A.P., India. And now he is working as Assistant professor in Department

of MCA, KITS, Warangal, A.P, India. He is a Member of ISTE. His current research interests in ar

computer networks and business data mining.

S.S.V.N. Sarma is a Professor and working as a Dean, Placement cell officer in Vaagdevi College of

Engineering, Warangal (A.P), India. He has more than 45 years of experience in teaching area. He served as

Head in Department of informatics, Chairman in Board of Studies, and Dean in Faculty of Science, Kakatiya

University, Warangal, A.P, India. His current research interests in the area of Data mining, Distributed systems,

Mobile computing, Computer networks

presented more than 120 articles. He has organized several workshops, delivered number of lecture at major and

minor conferences, and serves as Editor of several International/National j

European Journal of Business and Management

2839 (Online)

71

A.P, India. He has Publications more than 4 papers in area of Data mining and neural networks.

received his MCA (Computer Application) from Osmania University, Hyderabad, A.P,

India and M.Tech (Computer Science and Engineering) from Balaji Institute of Technology and Sciences,

JNTUH, Warangal, A.P, India. He served as an Assistant professor in Department of CSE, Balaji Institute of

Technology and Sciences, Warangal, A.P., India. And now he is working as Assistant professor in Department

of MCA, KITS, Warangal, A.P, India. He is a Member of ISTE. His current research interests in ar

computer networks and business data mining.

is a Professor and working as a Dean, Placement cell officer in Vaagdevi College of

Engineering, Warangal (A.P), India. He has more than 45 years of experience in teaching area. He served as

Head in Department of informatics, Chairman in Board of Studies, and Dean in Faculty of Science, Kakatiya

University, Warangal, A.P, India. His current research interests in the area of Data mining, Distributed systems,

Mobile computing, Computer networks, Computer security, and Performance evaluation. He has published and

presented more than 120 articles. He has organized several workshops, delivered number of lecture at major and

minor conferences, and serves as Editor of several International/National journals and magazines.

www.iiste.org

A.P, India. He has Publications more than 4 papers in area of Data mining and neural networks.

received his MCA (Computer Application) from Osmania University, Hyderabad, A.P,

India and M.Tech (Computer Science and Engineering) from Balaji Institute of Technology and Sciences,

professor in Department of CSE, Balaji Institute of

Technology and Sciences, Warangal, A.P., India. And now he is working as Assistant professor in Department

of MCA, KITS, Warangal, A.P, India. He is a Member of ISTE. His current research interests in area of

is a Professor and working as a Dean, Placement cell officer in Vaagdevi College of

Engineering, Warangal (A.P), India. He has more than 45 years of experience in teaching area. He served as

Head in Department of informatics, Chairman in Board of Studies, and Dean in Faculty of Science, Kakatiya

University, Warangal, A.P, India. His current research interests in the area of Data mining, Distributed systems,

, Computer security, and Performance evaluation. He has published and

presented more than 120 articles. He has organized several workshops, delivered number of lecture at major and

ournals and magazines.

This academic article was published by The International Institute for Science,

Technology and Education (IISTE). The IISTE is a pioneer in the Open Access

Publishing service based in the U.S. and Europe. The aim of the institute is

Accelerating Global Knowledge Sharing.

More information about the publisher can be found in the IISTE’s homepage:

http://www.iiste.org

The IISTE is currently hosting more than 30 peer-reviewed academic journals and

collaborating with academic institutions around the world. Prospective authors of

IISTE journals can find the submission instruction on the following page:

http://www.iiste.org/Journals/

The IISTE editorial team promises to the review and publish all the qualified

submissions in a fast manner. All the journals articles are available online to the

readers all over the world without financial, legal, or technical barriers other than

those inseparable from gaining access to the internet itself. Printed version of the

journals is also available upon request of readers and authors.

IISTE Knowledge Sharing Partners

EBSCO, Index Copernicus, Ulrich's Periodicals Directory, JournalTOCS, PKP Open

Archives Harvester, Bielefeld Academic Search Engine, Elektronische

Zeitschriftenbibliothek EZB, Open J-Gate, OCLC WorldCat, Universe Digtial

Library , NewJour, Google Scholar