Demonstrating the Use of Vector Error Correction M
Demonstrating the Use of Vector Error Correction M
net/publication/267405478
Demonstrating The Use Of Vector Error Correction Models Using Simulated Data
CITATIONS READS
0 993
2 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Carl B. Mcgowan on 06 November 2014.
ABSTRACT
In this paper, we demonstrate the use of time series analysis, including unit roots tests, Granger
causality tests, cointergation tests and vector error correction models. We generate four time
series using simulation such that the data has both a random component and a growth trend. The
data are analyzed to demonstrate the use of time series analysis procedures.
Keywords: Time Series Analysis; Simulation; Unit Root Tests; Granger Causality; Cointegration; Vector Error
Correction Models
INTRODUCTION
R egressions between levels of variables may have high covariation because of persistence in the base
levels of the variables rather than persistence in the changes in the values of the variables. Taking
the first differences of the variables may eliminate, or at least reduce, the dependence between the
variables. Gross national income from period to period is an integrated process, but the changes in GNI are not an
integrated process. The first differences of GNI are an independent, identically distributed process which are only
weakly dependent. An alternative transformation to differencing is to take the natural logarithm of the ratio of the
two levels to generate the percentage rate of change which generates a continuously compounded rate of change.
Ordinary Least Squares regression requires that the time series being evaluated be stationary. Otherwise,
OLS is no longer efficient, the standard errors are understated, and the OLS estimates are biased and inconsistent.
Stationarity requires that the time series values for the mean, the standard deviation, and the covariance, be invariate
over time1.
That is, the mean for any time (t-1) will equal the mean for any time (t), the standard deviation for any time (t-1) will
equal the standard deviation for any time (t), and the covariance for any time (t-1) will equal the covariance for any
time (t).
One method to test for stationarity is the unit root test of Dickey-Fuller (1979). To test for a unit root of a
stochastic time series, the value of the random variable is regressed against lagged values of the same random
variable
xt = + xt-1 + t [1]
1
See Wooldridge (2003) for a more detailed discussion of the theoretical models discussed in this paper.
© 2012 The Clute Institute http://www.cluteinstitute.com/ 377
Journal of Business Case Studies – July/August 2012 Volume 8, Number 4
where, xt is the value of the time series at time (t), is the intercept term, is the regression coefficient, xt-1 is the
lagged value of the time series, and t is the residual. If is equal to one, then the process generating the time series
is non-stationary. The null hypothesis is that H0: =1 and the alternative hypothesis is that is less than one, H1:
<1. The actual test is run after subtracting xt-1 from both sides of Equation [1]. The regression is
where the (*) indicates the parameters from the regression adjusted by subtracting x t-1. The null hypothesis is that
H0: *=0 and the alternative hypothesis is that is less than zero, H1: *<0.
This model is only valid for AR(1) processes. If the underlying return generating process exhibits serial
correlation of order greater than one, Augmented Dickey-Fuller tests must be used. Higher order terms are included
in the regression
xt = * + *xt-1 + 1 xt-1 + 2 xt-2 +…. + n xt-n + *t [3]
where, the additional terms are derived from the higher order AR() terms. The null hypothesis is that H 0: *=0 and
the alternative hypothesis is that * is less than one, H1: *<0.
Co-integrated processes are random in the short-term but tend to move together in the long-term.
Wooldridge (2003) shows that six-month Treasury bill rates and three-month Treasury bill rates are both unit root
processes that are independent in the short-term but do not drift too far apart in the long-term. If either rate moves
too far from equilibrium, either too high or too low, investors move money from the low (high) rate alternative to
the high (low) rate alternative. This process will raise (lower) the rate in the low (high) rate market.
Engle and Granger (1987) show that if a linear combination of non-stationary time series is stationary, the
time series is co-integrated. If two time series are integrated of order one, the time series resulting from adding the
two is integrated of order one. If yt ~ I(1) and xt ~ I(1), then (yt + xt) ~ I(1). However, if a beta, , exits such that (yt
- xt) ~ I(0), then yt and xt are said to be co-integrated. This co-integration equation reflects the long-term
relationship between yt and xt.
If we can construct a linear combination of yt and xt such that the difference of the two variables has a unit
root, the two variables are co-integrated and the regression coefficient is the co-integration parameter.
yt = 0 + 1xt + ut
If ut is I(0), then yt and xt are co-integrated. The model for testing for co-integration with a time trend includes a
time variable.
yt = 0 + 2(t)+ 1xt + ut
Error correction models are a class of models that provide insight into the long-term relationship between
variables in terms of the “impact propensity, long run propensity, and lag distribution for y as a distributed lag in
x.”2 The independent variable is x and the dependent variable is y. An error correction term is computed based on
the past values of both x and y. If past values of y are over-estimated, future values will be moved back toward
2
Wooldridge (2003), page 621.
378 http://www.cluteinstitute.com/ © 2012 The Clute Institute
Journal of Business Case Studies – July/August 2012 Volume 8, Number 4
equilibrium by the error correction factor. In the example of the six-month and three-month Treasury bill rates, the
error correction term is computed from the difference of the one period lagged six-month rate and the two-period
lagged three-month rate. Thus, if either of the two rates drift too far from the long-term rate, the error correction
term shows the tendency of the rates to return to the long-term rate.
If two variables are cointegrated, we can construct a variable, st, which is I(0). The resulting error
correction equation is
where, st-1, equals (yt-1 - *xt-1) and is the error correction term.
We can analyze the short-term effects of the relationship between the two variables. If the value of < 0,
the error correction term serves to return the process to the long-run value. That is, if (yt-1 > *xt-1), the process was
above the long-run value in the previous period and has been moved back by the error correction process.
We use an Excel spreadsheet and the Excel function Rand() to generate four times series of numbers of
1,000 observations each. Rand() generates a number from zero to one. In order to create a random number series
with a value of zero, the random number generated by Rand() is transformed into a zero value function by
subtracting 0.50 from each Rand() value, Rand(*)=(Rand()-0.50). This random number generated by Rand() and
transformed to a zero value number is used to create an Index value with the following equation:
Index (i,t) is the index value for each period “t” that is calculated from the previous Index (i,t) value plus a randomly
generated value with an expected value of zero plus the trend. The trend is a long-run trend added to the random
index change in order to create both a random component of the Index plus a trend. Four Indexes are generated
using this function with 1001 observations each.
Returns are calculated from each Index (i,t) using the natural logarithm function. Return(i,t) is the natural
logarithm of the ratio of Index(i,t) divided by Index (i,t-1).
Each return series has 1,000 observations that have both a random component and a trend component. The random
component is the value of Rand(*)(Return) that is added to each previous Index (i,t) plus a trend.
The four return series are analyzed using EViews. Figures 1 to 5 show the probability distribution for each
of the four return series. Figure 1 shows the sample statistics and analysis for Return(1,t) which has a mean value of
0.04981 with a standard deviation of 0.005108. The skewness statistic equals -0.022106 and the kurtosis statistic
equals 2.8937. The Jarque-Bera statistic to measure normality is 0.55, indicating that the probability distribution for
the Return(1,t) is normal. All four Return(I,t) series have expected values and standard deviations that are similar
and Jarque-Bera statistics that do not reject normality. That is, all four Return(i,t) series exhibit the probability
distribution statistics that one would expect given the method used to construct each of the four Return(I,t) series.
160
140
I 120
n
d
100
e
x
80
V
a 60
l
u
e 40
20
0
1 101 201 301 401 501 601 701 801 901 1001
Time Periods
Figure 1: Graph of Four Indexes
Time Series Analysis Simulation
100
S eries: ROR01
S ample 1 1000
80 Observations 1000
Mean 0.004981
60 Median 0.005108
Maximum 0.019589
Minimum -0.009770
40 S td. Dev. 0.004965
S kewness -0.022106
K urtosis 2.893713
20
Jarque-B era 0.552149
P robability 0.758757
0
-0.010 -0.005 0.000 0.005 0.010 0.015 0.020
100
S eries: ROR02
S ample 1 1000
80 Observations 1000
Mean 0.004980
60 Median 0.004945
Maximum 0.018570
Minimum -0.009741
40 S td. Dev. 0.004987
S kewness -0.025674
K urtosis 2.774439
20
Jarque-B era 2.229754
P robability 0.327956
0
-0.010 -0.005 0.000 0.005 0.010 0.015
100
S eries: ROR03
S ample 1 1000
80 Observations 1000
Mean 0.004978
60 Median 0.005077
Maximum 0.019341
Minimum -0.010689
40 S td. Dev. 0.004985
S kewness -0.117149
K urtosis 3.031226
20
Jarque-B era 2.327943
P robability 0.312244
0
-0.010 -0.005 0.000 0.005 0.010 0.015 0.020
100
S eries: ROR04
S ample 1 1000
80 Observations 1000
Mean 0.004979
60 Median 0.004838
Maximum 0.021767
Minimum -0.010496
40 S td. Dev. 0.004988
S kewness -0.018847
K urtosis 2.991698
20
Jarque-B era 0.062075
P robability 0.969439
0
-0.010 -0.005 0.000 0.005 0.010 0.015 0.020
Table 2 contains the correlation matrix for the four Return(i,t) series. The four Return(i,t) series are
constructed with a short-run random component and a long-run trend component. The correlation coefficients for
the four Return(i,t) series reflect the short-run relationship between each of the Return(i,t) series. Thus, we see in
Table 1 that the correlation coefficients for the four Return(i,t) series are all low and none are statistically
significant.
Generally, the first step in analyzing the relationships between time series is to determine if each Return(i,t)
series has a unit root. The Augmented Dickey-Fuller test for a unit root is performed for each of the four Return(i,t)
series and the empirical results are detailed in Table 3, Table 4, Table 5, and Table 6 for each simulated return
series. For the Return(1,t) series, the ADF test statistic is -14.63 and the critical value for the ADF test statistic is -
3.97 which indicates that Return(1,t) series does not have a unit root. None of the four lagged Return(1,t) series
variable regression coefficients are statistically significant, but the intercept term is and equals 0.5014. The adjusted
R2 for the regression is 0.4798 and the F-statistic is 152. These results reject the presence of a unit root. That is,
Return(1,t) series does not have a unit root which is consistent with the method of creating the Return(i,t) series.
The results for all four Return(i,t) series are similar to the results for Return(1,t) series.
The next step in the time-series analysis process is to determine if the four Return(i,t) series Granger cause
each other. Table 7 shows the Granger causality statistics for the four Return(i,t) series. There are six combinations
of Granger causality between the four Return(i,t) series, such as a determination if Return(1,t) series Granger causes
Return(2,t) series and vice versa. In all six cases, Granger causality is rejected, as would be expected since the
short-run component for each of the four Return(i,t) series are randomly generated.
Once one has determined that the four Return(i,t) series are normally distributed with no statistically
significant correlation, that the four Return(i,t) series are stationary with no unit roots, and that the four Return(i,t)
series do not Granger cause each other, the four Return(i,t) series are tested for cointegration. Cointegration tests
determine if the four Return(i,t) series have a long-run relationship that is not random as is the short-run
relationship. Given that the four Return(i,t) series are constructed with an equal trend, we expect that the four
Return(i,t) series will exhibit cointegration, which means that the four Return(i,t) series have a long-run relationship;
i.e., the four Return(i,t) series follow the same long-run trend. Table 8 contains the results of the Johansen
cointegration test. The test results indicate that there are four cointegrating equations at the 1% level of statistical
significance as would be expected by the process by which in indices were constructed.
Normalized
Cointegrating
Coefficients: 1
Cointegrating
Equation(s)
ROR01 ROR02 ROR03 ROR04 @TREND(2) C
1.000000 -1.678586 0.333607 0.035206 1.11E-08 0.001565
(0.23610) (0.12572) (0.11973) (9.3E-07)
Table 8: Continued
Normalized
Cointegrating
Coefficients: 2
Cointegrating
Equation(s)
ROR01 ROR02 ROR03 ROR04 @TREND(2) C
1.000000 0.000000 69.87104 -98.73838 7.35E-06 0.137262
(653.897) (925.581) (9.3E-05)
0.000000 1.000000 41.42619 -58.84332 4.37E-06 0.080840
(389.716) (551.637) (5.6E-05)
Normalized
Cointegrating
Coefficients: 3
Cointegrating
Equation(s)
ROR01 ROR02 ROR03 ROR04 @TREND(2) C
1.000000 0.000000 0.000000 0.072012 -4.85E-08 -0.005315
(0.11737) (5.5E-07)
0.000000 1.000000 0.000000 -0.259132 -1.44E-08 -0.003693
(0.11115) (5.2E-07)
0.000000 0.000000 1.000000 -1.414182 1.06E-07 0.002041
(0.19130) (8.9E-07)
Table 9 contains the empirical results for the VEC model with an intercept and with an intercept but no
trend in the error correction model. This empirical results for this model show that the error correction equation is
not statistically significant except in one case, ROR01 and ROR03(-2). The error correction variables are mostly not
statistically significant and the signs are random. The adjusted R 2 for the models are 0.003339 or less and the F-
statistics are not statistically significant. Table 10 contains the empirical results for the VEC model with a trend in
the data and both an intercept and a trend in the error correction model. Given that the four Return(i,t) series are
constructed with an intercept and a trend, the model with a trend in the data and a VEC model with both an intercept
and a trend would seem to be most appropriate. This empirical results for this model show that the error correction
equation is statistically significant but the trend is not statistically significant because the regression model accounts
for the long-run trend effect across the four Return(i,t) series. Although the error correction variables are mostly
statistically significant, the signs are random. This supports the hypothesis that cointegration is statistically
significant but random in effect. The adjusted R2 for the models are 0.33 or greater and the F-statistics are not
statistically significant.
ROR02(-1) -3.107689
(0.56475)
(-5.50280)
ROR03(-1) 0.912382
(0.23212)
(3.93063)
ROR04(-1) -0.294429
(0.18021)
(-1.63381)
@TREND(1) 2.80E-07
(1.7E-06)
(0.16173)
C 0.007315
Error Correction: D(ROR01) D(ROR02) D(ROR03) D(ROR04)
CointEq1 -0.103659 0.292230 -0.065245 -0.004799
(0.01815) (0.01644) (0.01838) (0.01839)
(-5.71090) (17.7705) (-3.55066) (-0.26097)
Vector AutoRegression technique cannot be applied to the four Return(i,t) series because the four
Return(i,t) series are cointegrated; that is, the four Return(i,t) series follow the same long-run trend, but the short-run
trend is random. There are eight options for running the VEC model. The VEC model can be run with no trend in
the VEC but with an intercept included or not. The VEC model can be run with a trend in the VEC and an intercept
and/or a trend in the cointegration equation. The vector error correction equation uses lagged deviations for each of
the four Return(i,t) series as independent variables for each of the four Return(i,t) series in a regression that also
include lagged deviation variables for each of the four Return(i,t) series. Each set of VEC estimated regression
includes the cointegrating equation plus a series of deviations from past changes in the four Return(i,t) series with up
to two lags, unless more lags are specified. In addition, each VEC analysis can include a trend in the VEC and/or an
intercept or a trend for each VEC. Table 10 contains the empirical results for the VEC model with a trend in the
data and both an intercept and a trend in the error correction model. Given that the four Return(i,t) series are
constructed with an intercept and a trend, the model with a trend in the data and a VEC model with both an intercept
and a trend would seem to be most appropriate. The empirical results for this model show that the error correction
equation is statistically significant but the trend is not statistically significant because the regression model accounts
for the long-run trend effect across the four Return(i,t) series. Although the error correction variables are mostly
statistically significant, the signs are random. This supports the hypothesis that cointegration is statistically
significant but random in effect. The other three models provide similar results.
In this paper, we generated four Return(i,t) series using Excel that have both a random component and a
trend component for each of the four Return(i,t) series. We applied a series of tests for time series analysis –
correlation, normality, unit root, Granger causality, cointegration, and vector error correction regressions.
The empirical results are consistent with the method used to create the four Return(i,t) series. Each of the
four Return(i,t) series has the same expected value and standard deviation, a low correlation with the other
390 http://www.cluteinstitute.com/ © 2012 The Clute Institute
Journal of Business Case Studies – July/August 2012 Volume 8, Number 4
Return(i,t) series, which reflects the short-run random effect built into the four Return(i,t) series, no unit roots, and
cointegration between the four Return(i,t) series, which Return(i,t) series is consistent with the method of
constructing the four with a trend. Since the four Return(i,t) series are cointegrated by construction, a vector error
correction model is appropriate for analysis of the long-run relationship between each of the four Return(i,t) series.
The coinetegration equation is statistically significant as are the error correction variables, but in a random fashion
with some of the regression coefficients being positive and some being negative.
In this paper, we show how to use the time series paradigm currently being used to conduct time series
analysis. The basis of this analysis is the work in time series analysis done by noble laureate Engle and Granger.
We demonstrate each of the steps designed to allow the researcher to determine if a relationship exists between two
time series and to define the nature of that relationship.
AUTHOR INFORMATION
Carl B. McGowan, Jr., PhD, CFA is a Faculty Distinguished Professor and Professor of Finance at Norfolk State
University, has a BA in International Relations (Syracuse), an MBA in Finance (Eastern Michigan), and a PhD in
Business Administration (Finance) from Michigan State. From 2003 to 2004, he held the RHB Bank Distinguished
Chair in Finance at the Universiti Kebangsaan Malaysia and has taught in Cost Rica, Malaysia, Moscow, Saudi
Arabia, and The UAE. Professor McGowan has published in numerous journals including American Journal of
Business Education, Applied Financial Economics, Decision Science, Financial Practice and Education, The
Financial Review, International Business and Economics Research Journal, The International Review of Financial
Analysis, The Journal of Applied Business Research, The Journal of Business Case Studies, The Journal of Diversity
Management, The Journal of Real Estate Research, Managerial Finance, Managing Global Transitions, The
Southwestern Economic Review, and Urban Studies. E-mail: [email protected]. Corresponding author.
Izani Ibrahim, PhD, is a Professor of Econometrics and Corporate Finance at the Graduate School of Business,
National University of Malaysia. He received his B.Sc. (Actuarial Science) in 1985, from Pennsylvania State
University, USA, and both his M.Sc. (Actuarial Science), in 1992 and Ph.D. (Finance), in 1996 from University of
Nebraska, USA. Professor Izani has been giving lectures in quantitative methods at graduate level and to public and
private sectors especially in Multivariate Analysis and Econometrics. His research interest in the finance area
includes topics on corporate finance, derivatives and investment. Professor Izani has published in Capital Market
Review, International Business & Economics Research Journal, International Research Journal, Journal of Current
Research in Global Business, Journal of International Trade and Economic Development, Journal Analisis, and
Journal Pengurusan. E-mail: [email protected]
REFERENCES
1. Dickey, D.A. and W.A. Fuller (1979) “Distribution of the Estimators for Autoregressive Time Series with a
Unit Root,” Journal of the American Statistical Association, 74, 427–431.
2. Engle, Robert F. and C.W.J. Granger (1987) “Co-integration and Error Correction: Representation,
Estimation, and Testing,” Econometrica 55, 251–276.
3. Jarque, C. and A. Bera (1980) "Efficient Tests for Normality, Homoskedasticity, and Serial Independence
of Regression Residuals," Economics Letters, 6, 255–259.
4. Johansen, Soren (1991) “Estimation and Hypothesis Testing of Cointegration Vectors in Gaussian Vector
Autoregressive Models,” Econometrica, 59, 1551–1580.
5. Johansen, Soren and Katarina Juselius (1990) “Maximum Likelihood Estimation and Inferences on
Cointegration—with applications to the demand for money,” Oxford Bulletin of Economics and Statistics,
52, 169–210.
6. Wooldridge, Jeffrey M. Introductory Econometrics: A Modern Approach, Second Edition, Thomson
South-Western, Mason, OH, 2003.
NOTES