Geweke, J Meese, R and Dent, W (1983) - Comparing Alternative Tests of Causality in Temporal Systems
Geweke, J Meese, R and Dent, W (1983) - Comparing Alternative Tests of Causality in Temporal Systems
John GEWEKE
University of Wisconsin, Madison, WI 53706, USA
Richard MEESE
University of California, Berkeley, CA 94720, USA
Warren DENT
Eli Lilly International Corporation, Indianapolis, IN 46206, USA
This paper discusses eight alternative tests of the absence of causal ordering, all of which are
asymptotically valid under the null hypothesis in the sense that their limiting size is known.
Their behavior under alternatives is compared analytically using the concept of approximate
slope, and these results are supported by the outcomes of Monte Carlo experiments. The
implications of these comparisons for applied work are unambiguous: Wald variants of a test
attributed to Granger, and a lagged dependent variable version of Sims’ test introduced in this
paper, are equivalent in all relevant respects and are preferred to the other tests discussed.
1. Introduction
*Thanks to Susan Hudak for research assistance and R.R. Bahadur for helpful discussions on
many points raised in the paper. Responsibility for remaining errors resides with the authors.
Views expressed are not necessarily those of the Board of Governors or the Federal Reserve.
Most of the work reported here was undertaken while the second author was a graduate student
and the third was Visiting Associate Professor in the Economics Department, University of
Wisconsin, Madison. The first author acknowledges financial support from the National Science
Foundation.
criterion for comparison of predictors is mean square error, then the question
of whether {y,} causes (xt} is equivalent to the question of whether there
exists an autoregressive representation,
(l.la)
(l.lb)
in which e(L) = 0.
Since (1.1) is a set of population regression equations with serially
uncorrelated disturbances, a test of the proposition that y does not cause x is
immediately suggested: estimate (l.la) by ordinary least squares, and test the
hypothesis e(L) = 0 in the conventional way. We shall refer to such a test as a
‘Granger test’ since the restriction e(L)=0 stems directly from Granger’s
definition. Unless one has special knowledge which restricts d(L) and e(L) to
be known functions of parameters whose number is reasonably small relative
to sample size, however, this cannot be done. In actual applications [e.g.,
Sargent (1976) and Geweke (1978)] some arbitrary restrictions on the form of
d(L) and e(L) must be made before estimation and testing can proceed; for
example, d(L) and e(L) might be assumed to be polynomials of finite order,
or ratios of polynomials of finite order. If the restrictions are true, then the
asymptotic properties of conventional tests of e(L)=0 are well known. If they
are false, such tests will in general reject the null hypothesis e(L) =0
asymptotically even if the null is true.
Sims (1972) considered the implications of the condition e(L)=0 for the
projection of y, on current and past xt,
s=o
(1.2)
COV(Ut,Xt_-s)=O, SZO,
S=-03
(1.3)
cov (v,, x,-J = 0, vs.
He showed that e(L)=0 - i.e., {y,> does not cause {xt> - if and only if the
second projection involves only current and past x,, so that f,(L)=f,(L). We
shall refer to a test of the hypothesis f,(L)=f,(L) as a ‘Sims test’ of the
hypothesis that (y,> does not cause {xt}. If such a test is to be implemented,
f,(L) andf,(L) must be restricted for the same reasons that d(L) and e(L) were
J. Geweke et al., Alternative tests of causality 163
where
c = cov (E,,6,),
to yield a new system in which the first equation is (l.la) and the second can
be expressed as
where g+(L) and f:(L) are functions of the parameters of (1.1). Since qt=
S,-(c/Q,~)E~,it is uncorrelated with E,, and hence Q is uncorrelated with current
x, as well as past x, and y,. The variables on the right side of (1.4) are
therefore predetermined, and qt is serially uncorrelated. Now suppose that v,
in (1.3) has autoregressive representation h(L)v,=w,, an assumption which is
no stronger than those employed in rigorous justifications for the use of
feasible generalized least squares in this equation. Let h(L) = 1 -h+(L) and
f:(L)= h(L)f,(L). Multiplying through (1.3) by h(L) and rearranging results in
(1.5)
combinations of all x,_, and past w,). Since h(L) is invertible, fr,=O for all
s >O if and only if& = 0 for all s > 0. A ‘Sims test’ of the hypothesis that {y,}
does not cause {xt} may therefore be conducted by basing a test off”;, = 0 for
all s<O on the estimates of some finite parameterization of (1.4) and (1.5) in
the conventional way. Again, the number of parameters in g+(L), h+(L),ff(L)
and f z(L) which are estimated must increase with sample size.
In this paper we report the results of an investigation into the actual
properties of several tests of the hypothesis that y does not cause x. This
investigation was undertaken for three reasons. First, as is evident from the
foregoing discussion, the theory which underlies these tests is only
asymptotic. Because of the need to work with a finite number of parameters
in a sample of finite size, the theory is even less operational than in the more
familiar case in which the parameterization is known: in the latter case we
know exactly how to construct the test statistic but are in doubt about its
precise interpretation in a sample of a given size, whereas in the tests
discussed above it is not known exactly how the test statistic should be
constructed or interpreted in any given situation.
Second, once a parameterization has been adopted, there remain several
ways in which linear restrictions may be tested. The most common are the
Wald, likelihood ratio, and LaGrange-multiplier tests discussed by Silvey
(1959); as Berndt and Savin (1977) have demonstrated, there are practical
situations in which the outcome depends on the test used.
Finally, and most important, as the results of more and more tests for the
absence of causal orderings are reported in the applied literature, it is
becoming evident [see Pierce and Haugh (1977)] that these results may
depend on which test is used.
The remainder of the paper is organized as follows. In the next section we
introduce eight alternative tests of the hypothesis that y does not cause x. All
are asymptotically valid in the sense that under the null hypothesis the
difference between the distribution of the test statistic and that of a sequence
of chi-square distributions vanishes as sample size increases. In section 3 we
compare the asymptotic behavior of all eight tests using the concept of
approximate slope [Bahadur (1960) and Geweke (1981 b)]. The design of a
Monte Carlo experiment conducted to assess the behavior of these
alternative tests in situations typical of those encountered in economic
applications is discussed in section 4. The outcome of the experiment is
reported in section 5, and a final section contains recommendations for
empirical work based on the results of our investigation.
In what follows, we shall assume that {x,} and {y,} are jointly stationary,
Gaussian, and have mean zero. The restriction that the set of predictors be
J. Geweke et al., Alternative tests of causality 165
T,GR
= n log (6$/r?;), (2.5)
The restriction that f,(L)=f,(L) in (1.2) and (1.3) on which Sims tests are
based is equivalent to {u,>= {u,}. If under the maintained hypothesis f,(L)
were restricted to a polynomial in the terms Lpp, . . ., L- ‘, 0, L,. . ., Lq, then a
test could be based on the regression equations
When an ad hoc prelilter [say, R(L)] is used to cope with serial correlation,
y, and x, are replaced by yt = R(L)y, and xt = R(L)x,, respectively, in (2.7) and
(2.8). The equations are then estimated by ordinary least squares and the test
F,(L)= F,(L) is conducted using a conventional Wald test statistic. In more
sophisticated feasible generalized least squares procedures, the serial
correlation structure of the disturbances is estimated. In a typical two-step
estimator [e.g., Hannan (1963) or Amemiya (1973)], (2.8) is first estimated by
ordinary least squares. It is assumed that Szi= var (V,, . . ., V,) = C$(a,), where
the function s2”(.) is known but the m, x 1 vector a, is unknown. The
maximum likelihood estimate oi, (or its asymptotic equivalent) of a, is
computed by ignoring differences between ordinary least squares residuals for
(2.8) and I$ The vector of coefficients E, is then estimated by generalized
least squares, replacing 52: with ai(ai,), which provides a vector of residuals
I? To construct the Wald test statistic, (2.7) is then estimated by generalized
least squares, replacing 52: with Q;(&), yielding a vector of residuals 0”. The
test statistic is
FSW)= 0#2,“(oi”))
- lo” - V’(Q@“))- l P, (2.9)
Under the null hypothesis the distributions of (2.9) and (2.10) converge to
that of a x’(p) variate. When mt,, m,, p and q are unknown and perhaps not
even known to be finite, m,, m,, and q may be increased with n: as functions
of n, however, proper choices of m,, mV, and q have not been worked out.
In implementations of (1.4) and (1.5) the parameterization problem may be
managed just as it was in the Granger test based on (1.1). A test off;,=0 for
all s<O can be based on the sums of squared residuals from ordinary least
J. Geweke et al., Alternative tests of causality 167
y,=G+(L)y,+FT(L)x,+H,=
i G,'ytms+ f
s=o
F~,x,_,+H,, (2.11)
s=l
If ~5; is the maximum likelihood estimate of var(H,) in (2.11) and 8& is the
maximum likelihood estimate of var (IQ in (2.12) then Wald, likelihood ratio
and LaGrange-multiplier tests may be based on the respective statistics
(2.13)
(2.15)
all of which have a limiting x’(p) distribution under the null hypothesis.
their comparative behavior under the alternative in the following way. Let
n’(t*,/?; fI) denote the minimum number of observations required to insure
that the probability that the test statistic Ti exceeds a specified critical point
t* is at least 1 -p. Then
where S,(A) denotes the spectral density of {xr} at frequency A [Whittle (1963,
ch. 2)]. Consequently the approximate slopes of the tests on (2.4)-(2.6) are
lnlvar(st)l=&_i lnlS,(l)ldA
x
we also have
whence
It follows that
TGW = Tsw 9 TGR = TSR> TGL = T=
170 .I. Geweke et al., Alternative tests of causality
Consider now Tcsw’ and T (sL),the probability limits of TisW’/n and ThsL)/n.
It is instructive to consider successively sophisticated variants of tests in (2.7)
and (2.8). Since there are fewer regressors in (1.2) than in (1.4), and fewer
regressors in (1.3) than in (1.5), g,’ ZD,” and o~Z& Eqs. (1.2) and (1.4)
coincide, and 0.”=o; if and only if u, is serially uncorrelated; similarly,
ai=aE if and only if u, is serially uncorrelated. If u, and V, are serially
uncorrelated, then a test of F,(L)=F,(L) may be conducted in classical
fashion
In this case,
Since uf may be serially correlated, a$? o,$ = c,‘. Consequently, Tcsw) >=TSW,
with equality if and only if the one-sided projection of {yJ> on {XI} turns out
to have a serially uncorrelated residual.
In a generalized least squares variant of the LaGrange-multiplier Sims test,
with R(L) chosen so that R(L)u, is serially uncorrelated, TcsL)= 1 -azt/a$. In
general, R(L), azt, and aft will differ from their counterparts in the Wald
Sims test. Because R(L)u,#v~, neither u! nor ut need be serially uncorrelated.
The approximate slope of a Sims LaGrange-multiplier test is also sensitive to
prefiltering. Treating {y,} and {xJ as the prefiltered series, we see that f,(L)
J. Geweke et al., Alternative tests of causality 171
and 0,’ vary with the prefilter, and consequently so will writ and a,$. [This
problem does not arise in the Wald Sims test, in which R(L) is always chosen
so that the disturbance in the two-sided projection is w, after serial
correlation correction.]
There appears to be no ordering between Tcsw) and TcsL’, and numerical
examples in the next section provide instances in which TcSL)> Tsw and
T’SL’< TsL
In practice, R(L) cannot be chosen so that R(L)u, or R(L)v, is known to be
serially uncorrelated. Either R(L) is chosen to be an arbitrary but reasonable
approximation of the autoregressive representation of u, or u, [Sims (1972)]
or more elaborate procedures are used to estimate R(L) explicitly [Amemiya
(1973)] or implicitly [Hannan (1963)]. In all instances the parameterization
problem again arises. In the mathematical appendix of this paper we show
that for sufficiently slow expansions of the relevant parameter spaces, sums of
squared residuals (after correction for serial correlation) divided by sample
size will converge to cft and ~zt in the restricted and unrestricted models,
respectively. The approximate slopes of the tests based on feasible
generalized least squares estimators are therefore the same as those based on
generalized least squares estimators.
4. Experimental design
(1) In a given population, what is the relative behavior of the tests discussed
above? Are the asymptotic distributions under the null adequate? Under
the alternative, do the approximate slopes of the different tests provide a
reliable indication of their relative abilities to reject the null hypothesis?
(2) Under the null, or under different alternatives in which approximate
slope is the same, is the performance of the tests sensitive to changes in
population characteristics such as autocorrelation or the lengths of
distributed lags?
(3) Under the alternative what is the effect of changing approximate slope
on test performance?
(4) How are the tests affected when the assumed parameterization is not that
of the population?
(5) In conducting a Sims test in which a correction for serial correlation is
necessary, what is the effect of preliltering?
the performance of all the tests, thus dealing thoroughly with the first
question. These tests were applied in six different populations, providing
some evidence on the second point. In a few cases the chosen
parameterization was not compatible with the population parameterization,
so limited information about point (5) may be forthcoming. All Sims tests
requiring correlation for serial correlation were conducted with and without
pretilters. In every case we used 100 observations.
The six models used in the experiments are summarized in table 1. The
parameter b was chosen in each case so that the common approximate slopes
of the Wald variants of the Granger and Sims lagged dependent variable
tests would be about 0.109. This value was selected to provide an interesting
frequency of rejection for the hypothesis that x does not cause y, using
Wald’s (1943) result on the distribution of Wald test statistics under the
alternative hypothesis. In our notation, his result states that the asymptotic
distribution of the test statistic under the alternative is non-central chi-square
with degrees of freedom equal to the number of restricted parameters and
non-centrality parameter equal to the product of the approximate slope and
the number of observations. The implied rejection frequencies for the
Granger and Sims lagged dependent variable Wald tests in our experiments
are about 0.76 when four parameters are restricted under the null and about
0.55 when twelve parameters are restricted, if the significance level of the test
is five percent. The approximate slopes of the Sims Wald and LaGrange-
multiplier tests vary across the six experiments.
Table 1
Design of experiments.
Mode1A:(1-0.75L)*y,=1.0+bx,_,+6~,6~-1N(0,1).
{a:}, {a:}, {it} are individually and mutually uncorrelated at all leads and lags.
Model A” Model B”
BApproximate slopes are given for respective tests of the hypothesis that {xc} does not cause
IYJ.
J. Geweke et al., Alternative tests of causality 173
Table 2
Parameterizations of (l.lHl.5) in the experiments for H,: {yt} does not cause {x,}.
Model A Mqdel B
0.2910
0.0000 0.2150
0.0000 0.1130
0.0000 0.2280
0.0000 0.2580
0.0000 0.2610
0.0000
0.4911
0.4365 0.3628
0.3225 0.1907
0.1695 - 0.1520
0.0760 - 0.1720
0.0860 -0.0870
0.1740
0.4604
0.4911 0.3401
0.3628 0.1788
0.1907 0.0000
1 0.0000
1 0.0000
1
0.3625
0.3108
0.4143 0.2679
0.2296
0.3061 0.1408
0.1207
0.1609
0.2185
0.2622 0.1614
0.1937 0.0849
0.1018
;I:; 1
0.0000
0.2910 1
0.0000
0.2150 1
0.0000
cx1130 -0.1758
-0.1900
0.2280 -0.1989
-0.2150
0.2580 -0.2012
0.2175
0.2610
$
:&1.7 1
- 0.0428
0.0000 1
- 0.0492
0.0000 1
-0.0489
0.0000
Table 3
Parameterizations of (l.lHl.5) in the experiments for H,: {.q} does not cause {y,}.
Model A Model B
k:: -0.0100
0.0167 -0.0021
0.036 1 0.0243
0.0853 - 0.0039
0.0079 - 0.0049
0.0080 - 0.0034
0.0026
k: - 0.0007
0.0030 -0.0010
-0.0015 0.0224
0.0234 - 0.0145
0.0046 - 0.0105
0.0027 0.0119
0.0023
;:’1.6 _a
0.0001 --0.0007
0.0004
0.0005 0.0192
0.0203
0.0214 - 0.0002
0.0011
0.0014 ~ 0.0005
0.0001
0.0012 0.0047
0.0054
LaGrange:
k -0.0100
0.0167 -0.0021
0.0361 0.0243
0.0853 - 0.0064
0.0074 -0.0085
0.0106 -0.0193
0.0448
k --0.0007
0.0030 -0.0010
-0.0015 0.0224
0.0234 0.0037
0.0014 -0.0036
0.0041 -0.0068
0.0182
k _a
0.0001 --0.0007
0.0005 0.0203
0.0214 -0.0011
0.0004 -0.0011
0.0005 0.0026
0.0076
f: 0.0001 - 0.0004 0.0192 -0.X03 - 0.0002 0.0060
Table 3 (continued)
Model A Model B
Table 3 (continued)
Model A Model B
the disturbances are serially uncorrelated. It is also less flexible than the one
for Amemiya’s method, employing a spectral estimator with only about five
effective degrees of freedom. The Hannan efficient estimator is especially ill-
suited to cope with the large variation in S,(n) and S,(2) near zero.
Both methods were also used after application of the prehlter (l -0.75L)’
to the data. For tests of the hypothesis that {y,} does not cause {xt> this
renders the disturbance term serially uncorrelated in all models. For tests of
{xt} not to cause { y,j the parameterization of Amemiya’s test is then inexact
in both experiments, but is a very good approximation. For those tests the
pretilter removes the peaks in S,(A) and S,(i) at 3.=O, whence the
parameterization in the Hannan efficient estimator should be more suitable.
Sims tests
A,B=OS \
44 F(4,91) F(4.90) 1.11 (0.08) 1.03 (0.07) 1.20 (0.08) I.1 I (0.07) 1.22 (0.09) 1.07 (0.07) 2.25 (0.27) 1.82 (0.15) 1.92 (C !4) 1.43 (r).@9) 1.73 (O.15) 1.39 (0.12)
&x F(4,87) F(4.86) 1.12 (0.08) 1.04 (0.07) 1.22 (0.08) 1.12 (0.07) I.1 I (0.09) 0.97 (0.07) 2.15 10.24) 1.76 (0.15) 1.37 (0.11) 1.05 (0.07) 1.68 (0.13) 1.36 (0.09)
4-12 F(4.83) F(4,82) I.15 (0.09) 1.06 (0.07) 1.24 (0.09) 1.13 (0.07) 1.16 (o.loj 1.01 io.oxj I.87 io.lxj 1.57 {o.lzj 1.48 (0.13) I.14 (0.08) 1.69 (0.12) 1.34 (0.09)
12-12 F(l2.75) F(l2.74) Lo6 (0.05) 0.89 (0.04) 1.19 (0.06) 1.06 (0.04) 1.12 (0.06) 0.86 (0.03) 1.55 (0.12) 1.11 (0.05) 1.47 (0.07) 0.96 (0.03) I.57 (0.10) I.01 (0.04)
8,0=0.5
4-4 F(4.91) F(4,9o) I.00 (0.07) 0.94 (0.06) 0.96 (0.96) 0.90 (0.06) 1.02 (0.07) 0.92 (0.06) 1.61 (0.13) 1.49 (0.11) 1.26 (0.10) 1.01 (0.07) 1.31 (0.10) 1.05 (0.07)
4-8 F(4,87) F(4.86) 0.98 (0.07) 0.92 (0.06) 0.95 (0.06) 0.88 (0.05) 1.04 (0.08) 0.92 (0.06) I.40 (0.12) 1.29 iO.O9) 1.19 (0.10) 0.95 (0.07) 1.26 (0.09) 1.03 (0.07)
4-12 F(4,83) F(4.82) 0.98 (0.07) 0.92 (0.06) 0.97 (0.05) 0.w (0.05) 1.06 (0.08) 0.94 (0.06) 1.25 (0.08) 1.13 (0.07) 1.21 (0.10) 0.95 (0.07) I .25 (0.09) I.00 (0.06)
12-12 F(l2.75) F(l2,74) 0.95 (0.04) 0.81 (0.03) I.01 (0.04) 0.94 (0.03) 1.07 (0.05) 0.83 (0.03) 1.32 (0.08) 0.98 (0.04) 1.28 (0.07) 0.87 (0.03) 1.37 (0.08) 0.89 (0.03)
A,B=O.X
4-4 F(4.91) F(4.9’3) I. I I (0.08) I.04 (0.07) 1.09 (0.07) 1.01 (0.06) I.17 (0.08) I.04 (0.07) 2.35 (0.29) 1.89 (0.17) 1.63 (0.15) 1.23 (0.09) I.45 (0.14) I.13 (0.09)
4-8 F(4.87) F(4.86) I.09 (0.08) I.01 (0.07) 1.08 (0.08) I.00 (0.07) I.18 (0.09) 1.03 (0.07) 1.91 (0.17) I.65 (0.13) 1.33 (0.10) I.04 10.07) 1.38 (0.11) 1.14 (0.09)
‘%I2 F(4.83) F(4,82) I.14 (0.06) I.06 (0.07) 1.13 (0.08) 1.03 (0.07) 1.22 (0.10) 1.05 (0.08) 1.70 (0.15) 1.48 (0.11) 1.4o~o.tlj 1.09~o.oxj 1.46 (0.11) I.21 (0.08)
12-12 F(l2,75) F(12,74) 1.09 (0.04) 0.91 (0.03) 1.10 (0.04) I.00 (0.03) 1.20 (0.07) 0.92 (0.M) 1.46 (0.09) I.1 I (0.05) I.41 (0.06) 0.95 (0.03) I.44 (0.07) 0.99 (0.03)
8,8=0.X
4-4 F(4.91) F(4.90) 1.19 (0.07) I.1 I (0.06) I.09 (007) I 01 (0.06) I.21 (0.09) 1.06 (0.07) 1.96 (0.19) 1.72 (0.14) 1.48 (0.12) 1.11 (0.07) 1.49 (0.11) I.17 (0.08)
4-X F(4,xn F(4,86) 1.19 (0.08) I.10 (0.07) 1.05 (0.071 0.97 CO.061 1.22 (0.08) 1.07 (0.07) 1.69 (0.16) 1.49 (0.12) 1.37 (0.10) I.04 (0.07) 1.42 (0.10) I.15 (0.07)
4-12 F(4,83) F(4, x2) I.17 (0.07) 1.08 (0.06) 1.07 (0.07j 0.99 io.o6j 1.27 (0.08) I.11 (0.07) 1.46 (0 13) 1.31 (0.10) 1.46 (0.11) 1.12 (0.07) 1.45 (0.1 I) I.15 (0.08)
12-12 F(l2.75) F(l2,74) 1.15 (0.05) 0.96 (0.03) I.06 (0.05) 0.97 (0.03) 1.27 (0.06) 0.98 (0.04) 1.52 (0.1 I) 1.14 (0.05) 1.50 (0.06) 1.01 (0.03) 1.49 (0.07) I.01 (0.04)
A.H=O.99
&4 F(4.91) F(4,9o) 1.26 (0.10) 1.16 (0.08) 1.17 (0.09) 1.07 (0.08) 1.57 (0.15) 1.32 (0.111 1.47 (0.16) 1.29 (0.12) 1.61 (0.13) I.15 (0.09) 1.68 (0.15) 1.32 (0.10)
&x F(4.87) F(4.86) 1.26 (0.09) 1.16 (0.08) I.18 (0.09) I.08 (0.07) 1.43 io.lzj 1.22 io.loj I.41 (0.14) I24 (0.11) I.52 (0.13) 1.16 (0.09) I.60 (0.13) 1.30 (0.10)
‘Cl2 F(4.83) F(4.82) 1.27 (0.091 1.17 (0.08) I .I9 (0.09) I ox (0.08) 1.26 (0.10) 1.07 (0.08) 1.46 (0.14) I 28 (0.1 I) I.51 (0.13) I.13 (0.09) 1.59 (0.13) 1.32 (0.10)
12-12 F(12.75) F(l2.74) 1.14 (o.osj 0.94 (0.04) 1.31 (0.05) 1.16 (0.04) 1.30 (0.06) 0 97 (0.04) 1.65 (0.14) 1.16 (0.06) I.68 (0.08) I.05 (0.04) 1.90 (0.11) I.19 (0.04)
5, u = 0.99
‘+I F(4,91) F(4,9o) 1.08 (0.07) 1.01 (0.07) I .05 (0.07) 0.98 (0.06) 1.96 (0.21) 1.63 (0.15) I71 (0.14) I.52 (0.12) 1.95 (0.19) I 35 (0.10) 1.18 (0.08) I.10 (0.07)
4-x F(4.87) F(4,86) 1.10 (0.07) 1.02 (0.07) I.09 (0.08) I.00 (0.07) I.25 (0.09) 1.11 (0.07) 1.63 CO.131 1.45 CO.111 1.48 (0.10) I IX (0.08) 1.26 (0.09) I.16 (0.08)
4-12 F(4.83) F(4, X2) 1.10 (0.07) 1.02 (0.07) 1.12 (0.08) 103 (0.07) 1.22 (0.W) 1.08 (0.07) 1.53 io.13j 1.36 io.lij I.41 (0.10) 1.10 (0.07) 1.27 (0.09) 1.16 (0.08)
12-12 F(l2,75) F(12.74) I.09 (0.05) 0.91 (0.03) I.17 (0.06) 1.05 (0.04) I .25 (0.06) 0.96 (0.03) I.44 (0.08) I 11(0.05) I.59 (0.10) 1.02 (0.04) 1.30 (0.06) 1.04 (0.04)
Sims tests
A,B=0.5
4-l F(491) F(4,90) 7, 14 4, I1 5. 14 5, 9 10,2l 6, 14 25, 37 20, 32 29, 38 14,23 28, 33 12,25
4-8 F(4,87) F(4,86) 7, 15 3. 11 7, I2 6, 8 7, 12 5. 7 24. 31 20.24 12.19 4, II 24, 32 11,20
4-12 F(4.83) F(4.82) IO.15 5, 12 9, 14 5, 10 9, 14 5:9 21.29 18, 21 15.25 6, 15 19, 33 10, I5
12-12 F(12,75) F(12, 74) 8, 11 2, 3 7, 16 5, 7 10, 13 1. 6 27, 31 12.15 24, 31 0, 5 24, 37 3, 6
B,B=OS
A,0=0.8
A,tJ=0.99
4-l F(4, 91) F(4190) IO, 18 8, I4 7, 17 6, 10 16, 23 11, 18 16, 23 13.19 19.29 7, 12 17. 25 5. 18
4-8 F(4,87) F(4,86) 9, I4 7. 12 8, 18 5. IO 16. 24 11,17 13,20 10, 18 15, 23 5, I5 17,28 l1,20
4-12 F(4,83) F(4,82) 11.16 7, 11 8, I4 5;11 11; 19 6, 11 16, 25 10, 19 16.25 7, 17 19,30 13,19
12-12 F(12.75) F(12.74) 7, 17 1, 2 15.23 6, 13 14, 23 1, 8 32, 40 9, 17 34, 38 2, 4 41. 52 6, I4
B, 0 =0.99
4-4 F(4.91) F(4,90) 2, 14 2, 8 2, 11 2, 7 27, 31 24, 27 23, 28 17,24 25, 31 17.22 12.21 5, 13
4-8 F(4,87) F(4, 86) 5, 15 3, 8 8, 13 5, 12 IO, 16 7, 13 22, 29 14, 24 20,29 7, 17 15,21 8, 15
4-12 F(4.83) F(4>82) 7, 16 2, 11 7, 13 5, 11 IO, 14 6, 11 17.25 II.22 16,22 5, 13 10, 23 6, 12
12-12 F(l2, 75) F(l2,74) 7, I3 4 3 9, 19 3, 8 13,21 5, 5 23, 34 9, 14 28, 38 2, 5 29, 39 3. 6
J. Geweke et al., Alternative tests of causality 181
experiments being too large and for others too small but with no apparent
pattern.
Overall, the results of the experiments reflect unfavorably on the
performance under the null hypothesis of tests requiring correction for serial
correlation. The Wald tests reject too frequently and are sensitive to
pretiltering. LaGrange-multiplier tests are also sensitive to preliltering. It is
discouraging to note that the LaGrange-multiplier tests did not perform very
well unless the parameterization of the distribution of the disturbance was
exact (i.e., Amemiya’s method) or the prelilter was used, in which case the
disturbance is then serially uncorrelated before correction for serial
correlation. If one of these conditions is indeed required for an adequate
distribution of the test statistic then these methods are useless in applied
work since then neither the serial correlation nor the functional form of its
parameterization is known a priori. On the other hand, it should be noted
that none of these tests are based on the exact maximum likelihood estimates
presumed in the discussion in section 2. Hannan’s and Amemiya’s procedures
in the Wald tests each constitute the first step in an iterative scheme leading
to maximum likelihood estimates [Oberhofer and Kmenta (1974)], and in the
sample size of 100 used in these experiments this first step may not be a very
close approximation to exact maximum likelihood. This question might merit
further investigation.
By comparison, the behavior of the Granger tests and Sims tests
incorporating lagged dependent variables is excellent. Rejection frequencies
for the Granger tests lie outside the 95% confidence intervals (constructed
under the assumption that the asymptotic distribution theory is exact) in 25
of 96 cases, and for the Sims lagged dependent variable tests in 15 of 96
cases. There is some tendency for the LaGrange-multiplier to be more
adequate than the Wald variant in the case of the Granger tests, and
conversely for the Sims lagged dependent variable tests. In both instances
rejection frequencies run a little too high for Wald tests and too low for
LaGrange-multiplier tests. The distribution of these statistics is as good as or
better than the distribution of the LaGrange-multiplier statistics after
pretiltering in the tests requiring a serial correlation correction. These test
statistics have the further advantages that they are cheaper to compute and
their interpretation is unclouded by the preliltering problem. However, there is
an analog of the problem of parameterizing the serial correlation of the
disturbance in choosing the number of lagged values of the dependent
variable to be used. There is no indication that test results are sensitive to
this choice as was the case in the tests requiring - a serial correlation
correction, even though in some cases the parameterization chosen was not
the correct one (as may be seen by consulting table 2).
The distributions of the test statistics for the hypothesis that (x~} does not
cause {y,> are summarized in tables 6 and 7. Given the results presented in
tables 4 and 5, the behavior of the Granger and Sims lagged dependent
Table 6
Mean va1ucs oi Fs under alternative hypotheses.”
44 F(4.91) F(4,90) 3.67 (0.19) 3.75 (0.18) 3.73 3.07 (0.13) 3.10 (0.12) 6.71 (0.30) 3.81 4.63 (0.15)
4-X F(4.87) F(4.86) 3.53 (0.18) 3.63 (0.18) 3.73 2.95 (0.12) 2.99 (0.12) 6.50 (0.29) 3.81 4.48 (0.15)
4-12 F(4,83) F(4,82) 3.45 (0.17) 3.58 (0.18) 3.73 2.88 10.121 2.93 CO.12) 6.28 (0.28) 3.81 4.30 (0.14)
12-12 F(12, 75) 1.77 (0.07) 1.87 (0.07) 3.73 2.88 (0.1 I) 1.94 5
F(12. 74) 1.34 io.04) 1.54 io.zoj 1.78 (0.04)
44 F(4.91) F(4,90) 3.47 (0.17) 3.31 (0.18) 3.73 2.93 (0.13) 2.77 (0.13) 4.66 (0.25) 3.89 3.49 (0.15)
4-8 F(4.87) F(4.86) 3.38 (0.17) 3.23 (0.18) 3.73 2.84 (0.13) 2.69 (0.13) 4.59 (0.24) 3.89 3.37 (0.15)
4-12 F(4,83) F(4.82) 3.30 (0.17) 3.18 (0.18) 3.73 2.76 (0.12) 2.63 (0.13) 4.52 (0.24) 3.89 3.31 (0.14)
12-12 F(12,75) F(R 74) 1.81 (0.07) 1.82 (0.17) 1.91 1.37 (0.04) 1.50 (0.05) 2.48 (0.11) I.96 1.54 (0.05)
A,8=0.5
4-4 F(4.91) F(4,90) 76,X2 75.82 0.76,0.85 66,76 72.79 97,9x 0.77,0.86 93,98
&X F(4,X7) F(4.861 76.85 73,83 0.76.0.85 62.75 65.77 96.98 0.77,0.86 93,96
4-12 F(4.83) F(4,xzj 79:X6 15,x1 0.76;0.85 56,78 69.79 96;98 0.77,0.86 91,96
12-12 F(12,75) F(l2.74) 62.72 45.62 0 55,0.68 8,20 19,3x X2.92 0.57,0.69 38.56
B,B=O.S
4-4 F(4.91) F(4.90) 73.79 70,81 0.76.0.84 69.78 66,78 94,96 0.77,0.86 X5.94
‘a F(4,87) F(4,86) 72,77 74.79 0.76,0.84 63,72 64.75 90,94 0.77,0.86 x4,93
4-12 F(4,83) F(4,82) 68,75 67.76 0.76,0.84 57.72 59.74 92,95 0.77,0.86 x0,94
12-12 F(l2,75) F(12.74) 50.60 47.61 0.55,0 67 14,31 26,47 70.85 0.56,0.69 31.42
A,Q=O.8
44 F(4,91) F(4.90) 59.76 60.70 0.76,0.85 55,67 51,65 90,96 0.82,0.89 X2.92
4-x F(4,87) F(4.86) 63,71 64.76 0.76,0.85 54.68 49,64 88.94 0.82,0.89 78.89
4-12 .Y4,831 F(4.82) a,70 58,6X 0.76,0.85 55.69 50.64 X5.94 0.89,0.89 77,X6
12-12 F(l2.75) F(12,74) 36,50 41,56 0.55,0.68 15,24 72,37 Xl,86 0.62,0.73 32,52
B,B=O.S
44 F(4,91) F(4,90) 69.79 71,79 0.76,0.85 59.74 63.74 X7,83 0.78,0.86 78.88
‘&8 F(4.87) F(4,86) 64.73 73.77 0.76,0.85 58,70 65,75 87.91 0.78,0.86 7X.89
4-12 F(4,83) F(4,82) 64,76 71,76 0.76,0.85 58.70 65,72 85,90 0.78,0.86 74,84
12~12 F(l2.75) F(12.74) 39.59 49.61 0.55,0.68 8,24 25.44 62,79 0.57,0.69 17,33
A, 8=0.99
4-4 F(4,91) F(4.90) 12,X0 32.51 0.76,0.85 67.82 24.43 69.74 0.96,0.98 55.70
&8 F(4.87) F(4,X6) 72.81 33,51 0.76,0.85 65,X0 23.37 61,75 0.96,0.98 55.67
4-12 F(4.83) F(4,82) 68,X3 30,4x 0.76,0.85 61,76 20,39 63,72 0.96,0.98 56.65
12-12 F(l2,75) F(12.74) 42,53 4X,56 0.55.0.6X 8,24 28.46 64,70 0.85.0.91 17.37
B, 8 = 0.99
4-4 F(4.91) F(4.90) 68.80 59,72 0.76,0.85 58.73 54.66 79.87 0.78,0.87 70,83
&X F(4,X7) F(4.86) 66,76 56.70 0.76,0.85 53,69 51,60 81,89 0.78,0.87 67,84
4-12 F(4.83) F(4,82) 62,76 56,68 0.76,0.85 52,65 47,60 80,X9 0.78,0 87 65,81
12-12 F(12,75) F(12,74) 40,52 37,52 0.55,0.68 10.26 22.36 66,7X 0.58,0.70 22,40
'Asymptoticdistribution
under nullhypothesis.
184 J. Geweke et al., Alternative tests of causality
variable tests are the most interesting. However, we also present some
information on the distribution of the filtered Hannan efficient Sims test of
the hypothesis that {xi} does not cause {y,}, since its distribution under the
null hypothesis was among the best of the tests involving serial correlation
correction. The Wald variant of this test rejects the null very reliably, and its
mean value is much higher than those of the other tests whose behavior is
summarized in these two tables. The tendency for this test statistic to be
‘too large’ is again evident in the first four experiments where its rejection
frequency is greater than the asymptotic distribution theory would lead one
to expect, and its mean value exceeds the limiting value (calculated as the
mean of a non-central chi-square distribution with degrees of freedom equal
to ‘F’ numerator degrees of freedom and non-centrality parameter equal to
the approximate slope multiplied by 100 - the number of observations
divided by the numerator degrees of freedom). In the fifth experiment the
parameterization of the Hannan efficient estimator cannot approximate the
behavior of the spectral density of the disturbance of (1.3) at low frequencies
very well, and evidently this prevents the test statistic from attaining the
large values suggested by the asymptotic theory set forth above. In the sixth
experiment the distribution over the 100 replications matches the asymptotic
distribution nicely, a result which is probably a fortuitous coincidence due to
the tendency of Wald Hannan efficient test statistics to be too large
combined with a downward bias due to an inability to parameterize the
spectral density of the disturbance adequately.
The behavior of the LaGrange-multiplier variant of this test is more
difficult to explain. This test rejects the null with a frequency greater than
that of the Granger and Sims lagged dependent variable Wald tests in most
instances when four restrictions are implied by the null, and usually with a
frequency greater than that of the LaGrange-multiplier variants of those tests
when twelve restrictions are implied. For the samples generated here, small
sample considerations evidently dominate the asymptotic behavior
demonstrated in section 2.
The behavior of both the Wald- and LaGrange-multiplier variants of the
Granger and Sims lagged dependent variable tests under the alternative
seems to be well described by asymptotic theory. In most cases rejection
frequencies run a little less than their limiting values [Wald (1943)], a
tendency which becomes more pronounced as serial correlation in {x,}
increases. The parameterization of the Sims lagged dependent variable test is
ill-suited in Model B, 8 =0.99 (see table 3) for essentially the same reason
that the form of this test requiring correction for serial correlation was not
well parameterized, a fact which again leads to a rate of rejection well below
the asymptotic norm. No systematic differences between Granger and Sims
lagged dependent variable tests emerge in the experiments beyond those
which clearly seem due to the inadequacy of the parameterization of the
J. Geweke et al., Alternative tests of causality 185
6. Conclusion
In this paper we have compared tests due to Granger (1969) and Sims
(1972) for the absence of a Granger causal ordering in stationary Gaussian
time series. Unlike previous efforts [e.g., Nelson and Schwert (1979)] our
comparison is based on asymptotic properties of these tests under the
alternative as well as the limiting behavior under the null and sampling
experiments. The analytic results on limiting behavior under the alternative
employ the concept of approximate slope, which is inversely proportional to
the number of observations required for a given power against the alternative
at small significance levels.
The implications of this study for empirical work are unambiguous: one
ought to use the Wald variant of either the Granger or Sims lagged
dependent variable tests described in the introduction. Both have three
important advantages over all other tests for the absence of Granger
causality which we have studied:
(1) The approximate slopes of these tests are at least as great as those of the
other tests (with the exception of the Wald Sims test) under all
alternatives.
(2) The sampling distribution of these tests under the null was very
satisfactory in our experiments, and under the alternative rejection
frequencies corresponded very closely to those indicated by the
asymptotic theory. By contrast, the sampling distribution of statistics for
Sims tests requiring correction for serial correlation conformed very
poorly to its limiting distribution and was sensitive to prefiltering under
the null hypothesis: under the alternative, rejection frequencies also
departed from their limiting values.
(3) The Granger and Sims lagged dependent variable tests are inexpensive to
compute and require only an ordinary least squares computer algorithm.
The variants of the Sims test which require correction for serial
correlation require much more computation time and more decisions
about parameterizations.
Our results are, of course, confined to time series which are stationary and
Gaussian, although the results of sections 2 and 3 will admit relaxation of
the normality assumptions employed in their derivation. It is difficult even to
186 J. Geweke et al., Alternative tests of causality
Mathematical appendix
IIAIII=(PI(A’A))‘, 11
All11
=n-’ fi (PL~(A’A))~.
Note that if A is symmetric,
lIAllr=y~I&‘% IIAIIII=n-’jilIP~L~;
J
(vii) If R(L) = Is”= 0 r,L” is an absolutely summable lag operator, its Fourier
transform is denoted @(n)= cs ,, r, exp (- ih).
(viii) For any function f(A) > 0 defined on [ - rc,n] and symmetric about zero,
Q.[f(A)] denotes the n x n diagonal matrix with jth entryf(271jln).
J. Geweke et al., Alternative tests ofcausality 187
A.2. Preliminaries
=IICAI+((II+((CBI+~~~~
= IIA~III’
+ IIBllrr~
Lemma 5. Let (Y”}, Y,zO, Vn, be a random sequence. Zf lim,,, E(Y,) =O, then
plim Y,= 0.
Proof Partition [0, cc) into the intervals [0, E] and (E,cc), where E is any
specified positive constant. Then E( Y,)2 E. P[ Y, > E), whence plim Y,= 0.
tj=& _i e’j*g@)d&
I
and suppose
tj =&5xeijAg(l)dl,
and let
Then
lim IIT,-J,llll=O.
n-tm
Proof. We closely follow Grenander and Szegii (1958, pp. 112-113) who
used weaker assumptions to show convergence in a weaker norm. Introduce
the Cesaro sums
(A.11
J. Geweke et al., Alternative tests of causality 189
The right side of (A.l) is bounded above by 2x7= r (itjl/p +2xj”=,ltj(, which
can be made arbitrarily small by choosing p suffkiently large.
Observe that
This expression is zero for lj-kl <n-p. Hence K,-L, has at most 2p non-
zero eigenvalues, and the absolute value of no eigenvalue can exceed
x:y”=_lt,l. Hence
converges to zero because the right side of (A.l) converges to zero. The result
follows from Lemma 4.
Proof Since {zf*j is ergodic, plim n- ’ I:= r z:’ =var (zr). The process z: has
spectral density lR((n)12S,(1)[Fishman (1969, p. 41)], whence
var(zT)=& _i I&l)("S,(A)dA.
II
190 J. Geweke et al., Alternative tests ofcausality
then
n-‘z’P’/i
nn n
p nnz =n-lzt’A n Zt =n-l
n n iil (di)'G 5 n- l i&
(z?i)21nll.
Lemma 10. Let {zt} and R(L) be the process and lag operator defined in
Lemma 8,
Proof Let Z;,= (~7;).. ., zT)‘=&,, the finite Fourier transform of {zt}. Let zy
have moving average representation zy= Q(L)E~, in which the E, are i.i.d.
normal. Let En= (ET,.. ., E:)‘= FOE,; the ~7 are i.i.d. complex normal. We have
By virtue of Hannan (1970, corollary 5, p. 214) and the fact that the
(R”(2j/n)(2 are bounded above and below by positive constants uniformly in
n, the difference between the right side of (A.3) and
vanishes in probability. But the $! are i.i.d. complex normal and R”and Q”are
J. Geweke et al., Alternative tests of causality 191
Theorem 1. Let (zt}, {z:}, and R(L) be the processes and lag operator defined
in Lemma 8, and define fi,, as in Lemma 9. Given (A.2),
64.4)
Zf there exists E>O such that P[n,(fi,J <c]+O, P[,I,(C&) > E- ‘l-+0, then (A.2)
may be replaced by
Proof The conditions on &(fi,) and ,I,(@,) guarantee that llb,,llr and II&;‘//,
are bounded away from zero with probability 1, in the limit. The conditions
on R(L) guarantee that IR(in)l’ is bounded above and below by positive
constants for AE[-rr,rc]. Lemma 4 then guarantees the equivalence of (A.2)
and (A.5).
The result itself follows from Lemmas 8, 9, and 10, with application of
Lemma 4.
Remark 1. Theorem 1 is sufficient for the claim of the text that tests based
on a generalized least squares estimator with known variance matrix, and on
a feasible generalized least squares estimator with estimated fi,, [satisfying
(A.2) or (A.5)], have the same approximate slope; let z,= y,-c:= 1 bjxjf, where
x1*,. . ., xkt are the regressors and bj is any consistent estimator of Bj in the
projection Cj”=r ajxj, of yt on xIt,. . ., xkt.
Proof Suppose initially that p is fixed, and let R,(L) denote the lag operator
from the autoregression of u, on u,_ 1,. . ., u,-~. The roots.of R,(L) =EEO r,PLs
lie outside an open annulus of the unit circle [Grenander and Szegii (1958,
192 1. Geweke et al., Alternative tests of causality
We have
(i) lim1152;‘-G~G$,=O,
(ii) plirnII~~l-~~~~lI,,=O,
The assertion (i) follows from the fact that 0;’ and GrGI: differ only in the
p x p submatrix consisting of their first p rows and columns. Hence
I;= I IP,~(K 1- G’G:)l is a fixed constant not depending on n. Similarly, (ii).
From Lemma 7,
Il~~cl~-G~G~lll,~;lGl~-G~lI,,.IIG~ll,+Il~~-Glllrr.IIG~llr.
>I
2 3
whence
plim
(
sup &F,p-rf)~Sp+~
[ X*+1s=o
=O,
lim- ((F,~~Cl~~(n)(21F:,-Fn~~CI~(~)121FhJJrr=0.
References
Amemiya, T., 1973, Generalized least squares with an estimated autocovariance matrix,
Econometrica 41, 723-732.
Baliadur, R.R., 1960, Stochastics comparison of tests, Annals of Mathematical Statistics 31, 276
295.
Bellman, R.A., 1960, Introduction to matrix analysis (McGraw-Hill, New York).
Berndt, E.R. and N.E. Savin, 1977, Conflict among criteria for testing hypotheses in the
multivariate linear regression model, Econometrica 45, 1263-1278.
Box, G.E.P. and M.E. Muller, 1958, A note on the generation of random normal deviates,
Annals of Mathematical Statistics 27, 610-611.
Breusch, T.S., 1979, Conflict among criteria for testing hypothesis: Extension and comment,
Econometrica 47, 203-208.
Fishman, G., 1969, Spectral methods of econometrics (Harvard University Press, Cambridge,
MA).
Geweke, J., 1978, Testing the exogeneity specification in the complete dynamic simultaneous
equation model, Journal of Econometrics 7, 163-185.
Geweke; J., 1981a, A comparison of tests of the independence of two covariance-stationary time
series, Journal of the American Statistical Association 76, 363-373.
Geweke, J., 1981b, The approximate slopes of econometric tests, Econometrica 49, 1427-1442.
JE- B
194 J. Geweke et al., Alternative tests ofcausality
Granger, C.W.J., 1969, Investigating causal relations by econometric models and cross spectral
methods, Econometrica 37, 428438.
Grenander, U. and G. Szego, 1958, Toeplitz forms and their applications (University of
California Press, Berkekely, CA).
Hannan, E.J., 1963, Regression for time series, in: M. Rosenblatt, ed., Proceedings of a
symposium on time series analysis (Wiley, New York).
Hannan, E.J., 1970, Multiple times series (Wiley, New York).
Knuth, D.E., 1969, The art of computer programming, Vol. 2 (Addison-Wesley, New York).
Nelson, C.R. and G.W. Schwert, 1979, Tests for Granger/Wiener causality: A Monte Carlo
investigation, University of Rochester Graduate School of Management working paper 7905.
Oberhofer, W. and J. Kmenta, 1974, A general procedure for obtaining maximum likelihood
estimates in generalized regression models, Econometrica 42, 5799590.
Pierce, D.A. and L.D. Haugh, 1977, Causality in temporal systems: Characterizations and a
survey, Journal of Econometrics 5, 2655293.
Rozanov, Y.A., 1969, Stationary random processes (Holden-Day, San Francisco, CA).
Sargent, T.J., 1976, A classical macroeconometric model of the U.S., Journal of Political
Economy 87,207-237.
Silvey, N.E., 1959, The LaGrangian multiplier test, Annals of Mathematical Statistics 30, 389-
407.
Sims, C., 1972, Money, income and causality, American Economic Review 62, 54@552.
Sims, C., 1974, Distributed lags, in: M.D. Intrihgator and D.A. Kendrick, eds., Frontiers of
quantitative economics (North-Holland, Amsterdam).
Wald, A., 1943, Tests of statistical hypotheses concerning several parameters when the number
of observations is large, Transactions of the American Mathematical Society 54, 426485.
Whittle, P., 1963, Prediction and regulation by linear least-squares method (Van Nostrand,
Princeton, NJ).