Financial Econometrics - #4
Financial Econometrics - #4
ASSUMPTIONS
OF CLASSICAL
REGRESSION
MODEL
The assumptions made for theLS
O
Classical Regression Model…
F
O
BASIC ASSUMPTIONS: S
N
I O
Zero Mean of the Disturbance: E[i] = 0 for all i;
P T
Homoscedasticity: Var[ ] = , a constant
for all i;
M
i
Nonautocorrelation: Cov[ , U
S Si ] = 0 if i j;
j
A
Uncorrelatedness of regressor and disturbance: Cov[X , ] = 0 if
F
i j
all i andj; O
N ]; and
Normality: ~ON[0,
I
i
A T
Non-Stochastic Regressor: the value of X is a known constant in
L
i
I O
the probability distribution of Y . And, X ’s are not linear function
i i
V
of other explanatory variables.
Violation #1:
The error term
does not have
a zero mean.
Error Term does notLS
have a zero mean!!!O
F
O
S
If a constant term is included N in the
Regression Model, thenTIthis O assumption will
never be violated and Phence, we should not
M
worry about it.
SU
AS
F
But, if the regression model does not require
the intercept
O
N and the error term is having
non-zeroI O mean, then it may lead to severe
A
biasesT in the slope coefficient estimates.
L
I O
V
Violation #2:
The error term
does not follow
Normal Probability
Distribution.
Error Term does not follow
Normal Probability LS
Distribution!!! O
F
O
It is assumed that Error TermSfollows Normal
Probability Distribution and Nas a
I O
consequence, we are able T to use t-Test, F-
P for Hypothesis
Test and 2 DistributionsM
Testing and Interval SU Estimation.
AS
F
O
If this assumption is not true, then our
estimates N will be still BLUE but we would find
I
difficultyO in performing hypothesis-testing and
AT
interval
L estimation.
estimation
I O
V
Error Term does not follow
Normal Probability LS
Distribution!!! O
F
O
S
How to test whether Error Term follows
eˆt +1 +
eˆt +
eˆt Time
- -
-
• No pattern in residuals at all.
all
• And, this is what we would like to see
If the Error Terms are correlated or have covariance
among them which is not zero, then it is called a problem
of SERIAL CORRELATION or AUTOCORRELATION !!!
ˆt
e
ût
eˆt +1
+
+
û t
-
ˆt
e
+
uˆ t 1
Time
- -
-
ˆt
e
+
uˆ t 1 T
ime
- -
Autocorrelation
problem is more
observed in Time Series
Data rather than Cross
Sectional Data.
How to detect
Autocorrelation …?
residuals.
No
Autocorrelation
!
Positive
Autocorrelation
!
How to detect
Autocorrelation …?
• Other Methods are –
• Run Test
• The Durbin-Watson test (d)
• It is test of autocorrelation of first order.
• The ratio of the sum of squared differences in successive residuals to the
RSS.
• No autocorrelation is its null hypothesis.
• But d has no unique critical value.
• Sample size and the number of regressors are used to calculate upper
(dU) and lower bounds (dL) to determine rejection regime.
• Rule of thumb: d 2(1-), d = 0 if = 1, d = 2 if = 0, d = 4 if = -1
(This relation provides better approximation as the sample size increases)
Durbin-Watson test for
Autocorrelation
d
i i 1 , for n and K -1 d.f.
(e e ) 2
i e 2
0
RESID
-2
-4
-6
-6 -4 -2 0 2 4
RESID(-1)
From the Table, we obtain critical values of Durbin-
Watson Statistic for 1 explanatory variable, 40
observations and level of significance 5% - dL=1.44 and
dU=1.54.
C 6.111328 P T
0.168570 36.25397 0.0000
LOG(P) 0.970582 M
0.110629 8.773336 0.0000
SU
R-squared A
0.706345S Mean dependent var 4.707273
Adjusted R-squared F
0.697168 S.D. dependent var 0.561094
S.E. of regression O
0.308771 Akaike info criterion 0.544589
Sum squared resid N 3.050865 Schwarz criterion 0.634375
Log likelihood I O -7.258010 Hannan-Quinn criter. 0.575208
F-statistic AT 76.97143 Durbin-Watson stat 1.291242
L
Prob(F-statistic) 0.000000
I O
V
residuals –Residual LS
over observation F O
O
S
RESID
N
.8
I O
.6
PT
M
.4
SU
.2
AS
.0
F
-.2 O
N
-.4
I O
A
-.6
T
L
I O -.8
V 5 10 15 20 25 30
residuals – Residual(-
LS
1) and Residual F O
O
.8
S
N
.6
I O
.4
PT
M
.2
SU
S
RESID
.0
A
-.2
F
O
-.4
N
I O-.6
AT
L -.8
I O -.8 -.4 .0 .4 .8
V RESID(-1)
From the Table, we obtain critical values of Durbin-
Watson Statistic for 1 explanatory variable, 34
observations and level of significance 5% - dL=1.393 and
dU=1.514.
Dependent Variable: LOG(A)
Method: Least Squares
Date: 08/13/10 Time: 06:09 What’s the conclusion?
Sample: 1 34
Included observations: 34
eˆt +1 +
eˆt +
eˆt Time
- -
-
• No pattern in residuals at all.
all
• And, this is what we would like to see
If the Error Terms are correlated or have covariance
among them which is not zero, then it is called a problem
of SERIAL CORRELATION or AUTOCORRELATION !!!
ˆt
e
ût
eˆt +1
+
+
û t
-
ˆt
e
+
uˆ t 1
Time
- -
-
ˆt
e
+
uˆ t 1 T
ime
- -
Autocorrelation
problem is more
observed in Time Series
Data rather than Cross
Sectional Data.
How to detect
Autocorrelation …?
residuals.
No
Autocorrelation
!
Positive
Autocorrelation
!
How to detect
Autocorrelation …?
• Other Methods are –
• Run Test
• The Durbin-Watson test (d)
• It is test of autocorrelation of first order.
• The ratio of the sum of squared differences in successive residuals to the
n
RSS. (e e
t t 1 )2
d t2
n
e
t 1
2
t
d
i i 1 , for n and K -1 d.f.
(e e ) 2
i e 2
0
RESID
-2
-4
-6
-6 -4 -2 0 2 4
RESID(-1)
From the Table, we obtain critical values of Durbin-Watson
Statistic for 1 explanatory variable, 40 observations and
level of significance 5% - dL=1.44 and dU=1.54.
C 6.111328 P T
0.168570 36.25397 0.0000
LOG(P) 0.970582 M
0.110629 8.773336 0.0000
SU
R-squared A
0.706345S Mean dependent var 4.707273
Adjusted R-squared F
0.697168 S.D. dependent var 0.561094
S.E. of regression O
0.308771 Akaike info criterion 0.544589
Sum squared resid N 3.050865 Schwarz criterion 0.634375
Log likelihood I O -7.258010 Hannan-Quinn criter. 0.575208
F-statistic AT 76.97143 Durbin-Watson stat 1.291242
L
Prob(F-statistic) 0.000000
I O
V
residuals –Residual LS
over observation F O
O
S
RESID
N
.8
I O
.6
PT
M
.4
SU
.2
AS
.0
F
-.2 O
N
-.4
I O
A
-.6
T
L
I O -.8
V 5 10 15 20 25 30
residuals – Residual(-
LS
1) and Residual F O
O
.8
S
N
.6
I O
.4
PT
M
.2
SU
S
RESID
.0
A
-.2
F
O
-.4
N
I O-.6
AT
L -.8
I O -.8 -.4 .0 .4 .8
V RESID(-1)
From the Table, we obtain critical values of Durbin-Watson
Statistic for 1 explanatory variable, 34 observations and
level of significance 5% - dL=1.393 and dU=1.514.
Dependent Variable: LOG(A)
Method: Least Squares
Date: 08/13/10 Time: 06:09 What’s the conclusion?
Sample: 1 34
Included observations: 34
1 2
*
11
y 1 y1; x 1 ; x 1 x1
* 2
*
12 2
for the first observation; and
y yt yt1; x 1 ; x xt xt1
*
t
*
t1
*
t2
for the remaining observations t = 2, 3, …
Once we get the transformed
variables, we use Least Square
Method without intercept using
variables –
* * *
y, x ,
t t1 xt2
Using GLS, we get the following:
Dependent Variable: YSTAR
Method: Least Squares
Date: 08/13/10 Time: 06:56
Sample: 1 34
Included observations: 34
Koutsoyiannis
1955 4,569 24,893
ItIt means
means that Cov(xxi,,xx j)) ¹00
that Cov(
i j
for some
for some values of ii ¹ jj..
values of
Why should I
bother about
Multicollinearity
?
CONCERNS OF MULTICOLLINEARITY!!!
What MULTICOLLINEARITY does impact?
Coefficientsa
Unstandardized Standardized
Coefficients Coefficients Collinearity Statistics
Model B Std. Error Beta t Sig. Tolerance VIF
1 (Constant) 6.377 5.629 1.133 .268
X .557 .075 .850 7.448 .000 .144 6.948
RY .154 .191 .154 .808 .426 .052 19.386
RX -.009 .153 -.013 -.058 .954 .034 29.070
a. Dependent Variable: Y
Solutions for Multicollinearity
…
• Ignore
• Drop Variable which are causing correlation.
• Transforming the variables
• Identify underlying factors.
What next…?