Econometrics Chapter Three (1)
Econometrics Chapter Three (1)
2
• One difficulty with R is that it can be made
large by adding more and more variables, even
if the variables added have no economic
justification. Algebraically, it is the fact that as
the variables are added the sum of squared
errors (RSS) goes down (it can remain
2
unchanged, but this is rare) and thus R goes up.
• An alternative measure of goodness of fit, called
the adjusted and often symbolized as , is
usually reported by regression programs. It is
computed as:
• =1-(1-
• This measure does not always goes up when a
variable is added because of the degree of
freedom term n-k decrease. As the number of
variables k increases, RSS goes down, but so
does n-k.
• It losses its interpretation; is no longer the
percent of variation explained. This modified is
sometimes used and misused as a device for
selecting the appropriate set of explanatory
variables.
Interpreting Regression Coefficients
1 Level-Level Models
• Y = b0 + b1X1 + · · · + bkXk + e
• The coefficient b1 is interpreted as follows ∂Y ∂X1 =
b1. The partial derivative of Y with respect to X1, ∂Y
∂X1 , can be interpreted as the rate of change in Y
associated with a change in X1 (holding constant
the variables X2 through Xk in the model).
• In other words if we change X1 by 1 unit (i.e.
∆X1 = 1, then we can expect Y to change by b1
units. That is: we expect that ∆Y = b1).
• So the interpretation of b1 in a level-level
regression is that a 1 unit change in X1 is
associated with a b1 unit change in Y holding
constant all other variables in the model.
2 Log-Log Models
A log-log model is a model where both the
dependent variable (Y) and the right hand side
variables (i.e., X1, ..., Xk) have been transformed
by the natural logarithm. These models can be
expressed as follows:
ln(Y ) = b0 + b1ln(X1) + · · · + bkln(Xk) + e
• So the interpretation in a log-log model is that a
1% change in X1 is associated with a b1 %
change in Y holding constant all other variables
in the model.
3 Level-Log Models
A level-log model is a model where the
dependent variable (Y) is in level form and the
right hand side variables have been transformed
by the natural logarithm. These models can be
expressed as follows:
Y = b0 + b1ln(X1) + · · · + bkln(Xk) + e
• Therefore, in a level-log model, the
interpretation is that a 1% increase in X is
associated with a b1/100 unit change in Y
holding constant all other variables in the model.
4 Log-Level Models
• A log-level model is where the dependent
variable (Y) has been transformed by the natural
logarithm and the right hand side variables are
in level form. These models can be expressed as
follows:
ln(Y ) = b0 + b1X1 + · · · + bxXk + e
• So the interpretation is that a 1 unit change in X1
is associated with a 100 · b1 % change in Y
holding constant all other variables in the model.
Hypothesis Testing in Multiple Regression Model
A.H 0 : 1 0
H1 : 1 0
B.H 0 : 2 0
H1 : 2 0
• The null hypothesis (A) states that, holding X2
constant X1 has no (linear) influence on Y.
Similarly hypothesis (B) states that holding X1
constant, X2 has no influence on the dependent
variable Yi.To test these null hypothesis we will
use the following tests:
Standard error test: under this and the following
testing methods we test only for ˆ .The test for ˆ
will be done in the same way.
• If SE(ˆ1 ) 1/2 ˆ1 , we accept the null
hypothesis that is, we can conclude that the
estimate i is not statistically significant.
• If SE(ˆ1) <1/2 ˆ 1, we reject the null
hypothesis that is, we can conclude that the
estimate i is statistically significant.
•Note: The smaller the standard errors, the
stronger the evidence that the estimates are
statistically reliable.
ii. The student’s t-test: We compute the t-ratio
for each ˆI
• If t*<t (tabulated), we accept the null hypothesis,
i.e. we can conclude that ˆ2 is not significant
and hence the regressor does not appear to
contribute to the explanation of the variations in
Y.
• If t*>t (tabulated), we reject the null hypothesis
and we accept the alternative one; ˆ2 is
statistically significant. Thus, the greater the
value of t* the stronger the evidence that i is
statistically significant.
3.5.2 Test of Overall Significance
Through out the previous section we were
concerned with testing the significance of the
estimated partial regression coefficients
individually, i.e. under the separate hypothesis that
each of the true population partial regression
coefficient was zero.
•In this section we extend this idea to joint test of
the relevance of all the included explanatory
variables. Now consider the following:
• Y 0 1 X 1 2 X 2 ......... k X k Ui
H 0 : 1 2 3 ............ k 0
H1 : at least one of the k is non-zero
•This null hypothesis is a joint hypothesis that
to X 1 , X 2 ,........X k .
• The test procedure for any set of hypothesis can
be based on a comparison of the sum of squared
errors from the original, the unrestricted
multiple regression model to the sum of squared
errors from a regression model in which the null
hypothesis is assumed to be true.
• When a null hypothesis is assumed to be true,
we in effect place conditions or constraints, on
the values that the parameters can take, and the
sum of squared errors increases.
• Let the Restricted Residual Sum of Square
(RRSS) be the sum of squared errors in the
model obtained by assuming that the null
hypothesis is true and URSS be the sum of the
squared error of the original unrestricted model
i.e. unrestricted residual sum of square (URSS).
• It is always true that RRSS - URSS 0.
• Consider
• Y ˆ0 ˆ1X1 ˆ X2 ......... 3ˆ X3 e .
•This model is called unrestricted. The test of
joint hypothesis is that:
H 0 : 1 2 3 ............ k 0
H1 : at least one of the k is different from zero.
• We know that: Y X ˆ X ......... ˆ X
• Yi Yˆ e
•e Y Y
• e2 (Y Yˆ )^2
•This sum of squared error is called unrestricted
residual sum of square (URSS). This is the case
when the null hypothesis is not true. If the null
hypothesis is assumed to be true, i.e. when all the
slope coefficients are zero.
• Y ˆ0 ei
By applying OLS
• e=