var and causality tests

34
Multivariate time series models VAR’s and Causality Tests Chapter 12&15 Asteriou (apprx 20-pages) Chapter 6 Brooks Chapter 5 Enders

Upload: oisin-o-cionaoith

Post on 21-Jan-2016

22 views

Category:

Documents


0 download

DESCRIPTION

Time Series Econometrics

TRANSCRIPT

Page 1: VaR and Causality Tests

Multivariate time series modelsVAR’s and Causality Tests

Chapter 12&15 Asteriou (apprx 20-pages)

Chapter 6 Brooks

Chapter 5 Enders

Page 2: VaR and Causality Tests

Simultaneous Equations Models

All the models we have looked at thus far have dealt with single dependent variables and estimations of single equations

All of the independent variables have been EXOGENOUS and y has been an ENDOGENOUS variable.

However interdependence is often encountered in economics, where several dependent variables are determined simultaneously

An example from economics to illustrate - the demand and supply of a good:  (1)

(2)(3)

 where = quantity of the good demanded = quantity of the good supplied

St = price of a substitute good Tt = some variable embodying the state of technology[Note: We’re trying to estimate the parameters (i.e. greek letters)]

Q P S udt t t t Q P T vst t t t Q Qdt st

Qdt

Qst

Page 3: VaR and Causality Tests

Assuming that the market always clears, and dropping the time subscripts for simplicity

(4)(5)

This is a simultaneous STRUCTURAL FORM of the model.

The point is that price and quantity are determined simultaneously by the market process (price affects quantity and quantity affects price). Another way to think of this: when P changes all we observe is changes in

equilibrium, we don’t know whether this is due to a shift of only demand or only supply or some combination. i.e. we can’t separately estimate the effect of a change in price on demand or on supply

P and Q are endogenous variables (they are determined by the model), while S and T are taken to be exogenous.

We can obtain REDUCED FORM equations corresponding to (4) and (5) by solving equations (4) and (5) for P and for Q (separately).

Simultaneous Equations Models: The Structural Form

Q P S u Q P T v

Page 4: VaR and Causality Tests

The steps before we see the maths1. Solve the equations simultaneously to find

an expression for P that does not involve Q This is the reduced form for P

2. Rearrange supply and demand to get them in the form P=…. Then solve simultaneously to get an expression for Q that does not involve P

This is the reduced form for Q

3. We can estimate the reduced forms

Page 5: VaR and Causality Tests

Solving to find an expression for P,

(6)

Rearranging (6),

(7)

Obtaining the Reduced Form

P S u P T v

P P T S v u

( ) ( ) ( ) P T S v u

P T Sv u

Page 6: VaR and Causality Tests

Rearranging (4) and (5) to get P on the left hand side and then equating them.

Solving to find an expression for Q, (8)

Multiplying (8) through by ,    

(9) 

Obtaining the Reduced Form (cont’d)

Q S u Q T v

Q Q T S u v

( ) ( ) ( ) Q T S u v

Q T Su v

Q S u Q T v

Page 7: VaR and Causality Tests

Obtaining the Reduced Form (cont’d)

 (7) and (9) are the reduced form equations for P and Q.

Note that each of the reduced form equations include the error terms from both (4) and (5)

(i.e. μ and υ)

P T Sv u

Q T Su v

Page 8: VaR and Causality Tests

 

But what would happen if we had estimated equations (4) and (5), i.e. the structural form equations, separately using OLS?

  Both equations (4) and (5) depend on P. One of the

CLRM assumptions was that the error term should be uncorrelated with the explanatory variables.

  It is clear from (7) that P is related to the errors in (4) and

(5) - i.e. it is stochastic. Another way of saying this is that a shock affecting one

equation will have an effect on the second equation via the influence on P!

Simultaneous Equations Bias

Page 9: VaR and Causality Tests

This can be seen more clearly in the following general form of a simultaneous equation model.

A) y1t = b10 + b11y2t+b12x2t+b14x4t +e1t

B) y2t = b20 + b21y1t+b23x3t+b24x4t +e2t

Let one of the error terms e1t increase and hold all else constant Then:

This will cause y1t to increase due to equation A In turn y1t will cause y2t to increase (if b21 is positive) due to equation B) if y2t increases in B) then it also increases in A) where it is an explanatory variable. Hence an increase in e1t caused an increase in an explanatory variable in the same equation

(meaning the error and the explanatory variable are correlated with each other!!).

(I have shown this with the arrows: Blue leads to Red leads to Green

Simultaneous Equation Bias

Page 10: VaR and Causality Tests

Simultaneous Equations Bias

What would be the consequences for the OLS estimator, , if we ignore the simultaneity?

Application of OLS to structural equations which are part of a simultaneous system will lead to biased and inconsistent estimators

Hence it would not be possible to estimate simultaneous equations validly using OLS.

Page 11: VaR and Causality Tests

So What Can We Do? Taking the normal equations (7) and (9), we can rewrite them as

(10) 

(11)  We CAN estimate equations (10) & (11) using OLS since all the RHS

variables are exogenous [(10) doesn’t have Q while (11) doesn’t have P => feedback has been controlled for!].

But ... we probably don’t care what the values of the coefficients are; what we wanted were the original parameters in the structural equations - , , , , , .

Can We Retrieve the Original Coefficients from the ’s?Short answer: sometimes.This is known as the identification problem

Avoiding Simultaneous Equations Bias

P T S 10 11 12 1

Q T S 20 21 22 2

Page 12: VaR and Causality Tests

Identification:

We re-wrote:

as

and

as

So: if we ‘know’ the π’s it

may be possible to solve these equations to find what we care about i.e. α, β, etc. – but not always!!

P T Sv u

Q T Su v

P T S 10 11 12 1

Q T S 20 21 22 2

etc,, 1110

Page 13: VaR and Causality Tests

VAR’s

Page 14: VaR and Causality Tests

Vector Autoregressive (VAR) Models

Identification of exogenous and endogenous variables was heavily criticised by Sims(1980).

According to Sims if there is simultaneity among a number of variables, then all these variables should be treated in the same way.

All variables should be treated as endogenous This means that in its general reduced form each equation has

the same set of regressors which leads to the development of VAR models

VARs have been often advocated as an alternative to large scale simultaneous equations structural models

Page 15: VaR and Causality Tests

• A natural generalisation of autoregressive models popularised by Sims

• A VAR is in a sense a systems regression model i.e. there is more than one dependent variable.

 • Simplest case is a bivariate VAR(k)

where u1t and u2t are uncorrelated white noise error terms

 • This is a bivariate VAR(k) since there are 2 variables and the longest lag is k• The analysis could be extended to a VAR(k) model with g variables and g

equations but still k lags.

• (Note: when looking back at previous slides remember Var is used for Variance!

Vector Autoregressive Models

y y y y y u

y y y y y ut t k t k t k t k t

t t k t k t k t k t

1 10 11 1 1 1 1 11 2 1 1 2 1

2 20 21 2 1 2 2 21 1 1 2 1 2

... ...

... ...

Page 16: VaR and Causality Tests

• One important feature of VARs is the compactness with which we can write the notation. For example, consider the case from above where k=1.

 • We can write this as  

or using matrix notation as  

or even more compactly as

yt = 0 + 1 yt-1 + ut

21 21 22 21 21

Vector Autoregressive Models: Notation and Concepts

y y y u

y y y ut t t t

t t t t

1 10 11 1 1 11 2 1 1

2 20 21 2 1 21 1 1 2

y

y

y

y

u

ut

t

t

t

t

t

1

2

10

20

11 11

21 21

1 1

2 1

1

2

Page 17: VaR and Causality Tests

• This model can be extended to the case where there are k lags of each variable in each equation and g endogenous variables:

yt = 0 + 1 yt-1 + 2 yt-2 +...+ k yt-k + ut

g1 g1 gg g1 gg g1 gg g1 g1

• We can also extend this to the case where the model includes first difference terms and cointegrating relationships (a VECM) – we’ll see this later .

Vector Autoregressive Models: Notation and Concepts (cont’d)

Page 18: VaR and Causality Tests

Advantages of VAR Modelling- Do not need to specify which variables are endogenous or exogenous - all are endogenous- Allows the value of a variable to depend on more than just its own lags or combinations of white noise terms, so more general than ARMA modelling- Provided that there are no contemporaneous terms on the right hand side of the equations, can simply use OLS separately on each equation. For OLS to be used efficiently it is also required that there is the same lag length in each equation- Forecasts are often better than “traditional structural” models.

Problems with VAR’s- VAR’s are a-theoretical (as are ARMA models)- How do you decide the appropriate lag length?- So many parameters! If we have g equations for g variables and we have k lags of each of the variables in each equation, we have to estimate (g+kg2) parameters. e.g. g=3, k=3, parameters = 30 so need a lot of data- Do we need to ensure all components of the VAR are stationary?- How do we interpret the coefficients?

Vector Autoregressive Models Compared with Structural Equations Models

Page 19: VaR and Causality Tests

Does the VAR Include Contemporaneous Terms?

So far, we have assumed the VAR is of the form

 

But what if the equations had a contemporaneous feedback term?

 

We can write this as

This VAR is in primitive (structural) form. This cannot be estimated using OLS

y y y u

y y y ut t t t

t t t t

1 10 11 1 1 11 2 1 1

2 20 21 2 1 21 1 1 2

y y y y u

y y y y ut t t t t

t t t t t

1 10 11 1 1 11 2 1 12 2 1

2 20 21 2 1 21 1 1 22 1 2

y

y

y

y

y

y

u

ut

t

t

t

t

t

t

t

1

2

10

20

11 11

21 21

1 1

2 1

12

22

2

1

1

20

0

Page 20: VaR and Causality Tests

Primitive versus Standard Form VARs

We can take the contemporaneous terms over to the LHS and write

or

B yt = 0 + 1 yt-1 + ut

  We can then pre-multiply both sides by B-1 to give

  yt = B-10 + B-11 yt-1 + B-1ut

or

yt = A0 + A1 yt-1 + et

This is known as a standard form VAR, which we can estimate using OLS. In general the approach is to estimate the VAR equation in standard form

1

122

12 1

2

10

20

11 11

21 21

1 1

2 1

1

2

y

y

y

y

u

ut

t

t

t

t

t

Page 21: VaR and Causality Tests

Choosing the Optimal Lag Length for a VAR

An important decision as too few leg will mean the model is misspecified and too many lags will waste degrees of freedom.

To check lag length begin with the longest feasible length given degrees of freedom considerations.

Can use single equation and multi equation information criteria The AIC and SBC can be applied to each equation in the

VAR

Recall that in both the AIC and SBC increasing the lag length will reduce the RSS at the ‘expense’ of increasing k.

k= number of parameters in the equation, which equals g.p +1 where g is the number of variables, p is the number of lags of each variable and 1 is for the intercept term

Choose the AIC or SBC with the lowest value

nkeT

RSSAIC /2)( nke

T

RSSSBC /)(

Page 22: VaR and Causality Tests

Choosing the Optimal Lag Length for a VAR

Multivariate versions of the information criteria are available. These can

be defined as:

MAIC = T ln│∑│ + 2N

MSBC = T ln│∑│ + Nln(T)

where │∑│is the determinant of the var/cov matrix of residuals μi in the standard VAR system.

N is the total number of parameters estimated in all equations, which will be equal to g2k + g for g equations, each with k lags of the g variables, plus a constant term in each equation. The values of the information criteria are constructed for 0, 1, … lags (up to some pre-specified maximum).

Select the model with the lowest MAIC or MSBC

Page 23: VaR and Causality Tests

Diagnostic Checking

Once we have estimated a VAR(p) standard form model we can check the residuals from the equation to see if they are white noise error terms, in particular to see if the residuals are serially correlated Apply the Lagrange Multiplier test of serial

corelation. LM~χ2q where q is the number of lags

you wish to test in the μt series

Apply the Jarque-Bera test for normality. JB~χ22

Page 24: VaR and Causality Tests

Causality

Page 25: VaR and Causality Tests

Block Significance and Causality Tests

It is likely that, when a VAR includes many lags of variables, it will be difficult to see which sets of variables have significant effects on each dependent variable and which do not. For illustration, consider the following bivariate VAR(3):

This VAR could be written out to express the individual equations as

We might be interested in testing the following hypotheses, and their implied restrictions on the parameter matrices:

t

t

t

t

t

t

t

t

t

t

u

u

y

y

y

y

y

y

y

y

2

1

32

31

2221

1211

22

21

2221

1211

12

11

2221

1211

20

10

2

1

tttttttt

tttttttt

uyyyyyyy

uyyyyyyy

2322231212222212112221121202

1321231112212211112121111101

Page 26: VaR and Causality Tests

Block Significance and Causality Tests (cont’d)

Each of these four joint hypotheses can be tested within the F-test framework, since each set of restrictions contains only parameters drawn from one equation.

These tests are also be referred to as Granger causality tests. Granger causality tests seek to answer questions such as “Do changes in y1

cause changes in y2?” If y1 causes y2, lags of y1 should be significant in the equation for y2. If this is the case, we say that y1 “Granger-causes” y2.

If y2 causes y1, lags of y2 should be significant in the equation for y1. If both sets of lags are significant, there is “bi-directional causality” If both sets of lags are not significant in either equation then y1 and y2 are

independent

Hypothesis Implied Restriction1. Lags of y1t do not explain current y2t

21 = 0 and 21 = 0 and 21 = 02. Lags of y1t do not explain current y1t

11 = 0 and 11 = 0 and 11 = 03. Lags of y2t do not explain current y1t

12 = 0 and 12 = 0 and 12 = 04. Lags of y2t do not explain current y2t

22 = 0 and 22 = 0 and 22 = 0

Page 27: VaR and Causality Tests

Vector Autoregression (VAR) Vector Autoregression (VAR) AnAn ExampleExample

Granger causality tests – which have been interpreted as testing, for example, the validity of the monetarist proposition that autonomous variations in the money supply have been a cause of output fluctuations.

Granger-causality requires that lagged values of variable A are related to subsequent values in variable B, keeping constant the lagged values of variable B and any other explanatory variables

Page 28: VaR and Causality Tests

Vector Autoregression (VAR)Vector Autoregression (VAR)Granger CausalityGranger Causality

In a regression analysis, we deal with the dependence of one variable on other variables, but it does not necessarily imply causation. In other words, the existence of a relationship between variables does not prove causality or direction of influence.

In the relationship between GDP and Money supply, the often asked question is whether GDP M or M GDP. Since we have two variables, we are dealing with bilateral causality.

The structural VAR equations are given as:

yttttt

mttttt

ymmy

ymym

1221212

1121111

Page 29: VaR and Causality Tests

Vector Autoregression (VAR)Vector Autoregression (VAR)Granger CausalityGranger Causality

We can distinguish four cases:

Unidirectional causality from M to GDP Unidirectional causality from GDP to M Feedback or bilateral causality Independence

Assumptions: Stationary variables for GDP and M Number of lag terms, can be established using AIC/SBC Error terms are uncorrelated

Page 30: VaR and Causality Tests

Vector Autoregression (VAR)Vector Autoregression (VAR)Granger Causality Example – Estimation (t-test)Granger Causality Example – Estimation (t-test)

mt does not Granger cause yt relative to an information set consisting of past m’s and y’s iff 21 = 0.yt does not Granger cause mt relative to an information set consisting of past m’s and y’s iff 12 = 0.

In a model with one lag, as in our example, a t-test can be used to test the null hypothesis that one variable does not Granger cause another variable. In higher order systems, an F-test is used.

tttt

tttt

ymy

ymm

2122121

1112111

Page 31: VaR and Causality Tests

1. Regress current GDP on all lagged GDP terms but do not include the lagged M variable (restricted regression). From this, obtain the restricted residual sum of squares, RSSR.

2. Run the regression including the lagged M terms (unrestricted regression). Also get the residual sum of squares, RSSUR.

3. The null hypothesis is Ho: i = 0, that is, the lagged M terms do not belong in the regression.

5. If the computed F > critical F value at a chosen level of significance, we reject the null, in which case the lagged m belong in the regression. This is another way of saying that m causes y.

Vector Autoregression (VAR)Vector Autoregression (VAR)Granger Causality – Estimation (F-test)Granger Causality – Estimation (F-test)

)/(

/)(

knRSS

mRSSRSSF

UR

URR

Page 32: VaR and Causality Tests

The Sims Causality Test

Sims (1980) put forward an alternative test for causality

The key to this test is that it is not possible for the future to cause the present

Suppose we want to find out does y cause m. For Sims causality testing consider the following model

ttttt

ttttt

mymy

yymm

2123122121

1113112111

Page 33: VaR and Causality Tests

The Sims Causality Test

We can test if yt causes mt by testing if 13=0. If 13≠ 0 then yt+1 does not cause mt as the future cannot cause the present

Again this hypothesis can be tested using a t test if there is 1 lead term. If there are more than one lead terms then an F test is applied to jointly test the hypothesis that they are all equal to zero

Page 34: VaR and Causality Tests

Overview of course: