Exam preparation
Exam preparation
CHAPTER THREE
MULTIPLE LINEAR REGRESSIONS
In our firm example, for instance, it was assumed implicitly that only labour hour (X)
affects output (Y). But besides labour – hour, a number of other variables/factors are
also likely to affect output, such as capital. Therefore, we need to extend our simple two-
variable regression model into multi-variable regression model.
These include:
o Linearity in parameters
o Zero mean value of the disturbance term, ( )= 0
o No serial correlation between disturbance terms, ( )= ( ) ( )=0
o Homoscedasticity, ( ) - constant variance of the error term
o Zero covariance between error term and each explanatory variable: ( )=
( )= 0
o No specification bias. The model is correctly specified
o No perfect multicollinearity between the explanatory variables.
The coefficients and are called the Partial Regression Coefficients. The
interpretations of the βs coefficients are different from the case in simple regression. In
multiple regressions Model single coefficient can only be interpreted under ‘ceteris
paribus conditions’ (other things being constant).
To find the OLS estimators, let us first write the SRF corresponding to the PRF as
follows:
̂ + ̂ +̂ +̂
The OLS estimators are obtained where the squared sum of the residuals (RSS) from the
estimation is as minimum as possible.
Problem:
min ∑ ̂ = ∑( ̂ ̂ ̂ )
FOC: first take the partial derivative w.r.t. ̂ ̂ and ̂ set it equal to zero
∑̂
= ∑[ ̂ ̂ ̂ ]( )
̂
∑̂
[∑( ̂ ̂ ̂ )( )]
̂
u
2
2 Yi ˆ0 ˆ1 X 1 ˆ 2 X 2 X 2 0
ˆ 2
Y ˆ
i 0 ˆ1 x1 ˆ2 x2
Finally after mathematical manipulation, we get the following formulas for each
estimator.
∑ ∑ ∑ ∑
̂
∑ ∑ (∑ )
Alternatively in deviation form we can write the following equation:
∑ ̂ ∑
̂
∑
∑ 𝒚𝒊 𝒙𝟐 ∑ 𝑥 ∑ 𝒚𝒊 𝒙𝟏 ∑ 𝒙𝟏 𝒙𝟐
̂𝟐
𝜷
∑𝑥 ∑𝑥 (∑ 𝒙𝟏 𝒙𝟐 )𝟐
Chapter Three Multiple Linear Regression
Yi 30 20 36 24 40
X1 4 3 6 4 8
X2 10 8 11 9 12
In deviation form:
∑ = ∑ - ̅ ̅ = 62, ∑ = ∑ -n ̅ ̅ = 52, ∑ = ∑ -
̅ = 16, ∑ = ∑ - ̅ = 10, ∑ ∑ - ̅ ̅ = 12, ∑ =
∑ - ̅ = 272
Thus,
∑ ∑ ∑ ∑
̂ = = = = - 0.25
∑ ∑ (∑ ) ( )
∑ ∑ ∑ ∑
̂
∑ ∑ (∑ )
= ( )
= = = 5.5
𝑬𝑺𝑺 ̂ 𝟏 ∑ 𝒙𝟏 𝒚𝒊
𝜷 ̂ 𝟐 ∑ 𝒙𝟐 𝒚𝒊
𝜷
𝑹𝟐 𝑻𝑺𝑺 ∑ 𝒚𝒊 𝟐
and 0 <𝑅 < 1
Note also:
For our numerical example 3, is computed as follows:
̂ ∑ ̂ ∑
∑
= = = 0.994485
Adjusted (̅ )
It is important to note that never decreases, it usually increases, when another
independent variable is added into a regression. As the number of explanatory
(independent) variables increases, always increases. The makes it a poor tool for
deciding whether one variable or several variables should be added to a model. This
implies that the goodness-of-fit of an estimated model depends on the number of
independent (explanatory) variables regardless of whether they are important or not.
To eliminate this dependency, we compute the adjusted (̅ ) as:
The partial correlation coefficients measure the strength of linear correlations between
variables independent of the impudence of other variables existed within the model.
Partial correlation coefficients are used in multiple regression analysis to determine the
relative importance of each explanatory variable in the model.
In a three variable regression model, we can compute three pair- wise correlation
coefficients;
= partial correlation between Y and , while holding is constant
= partial Correlation between Y and holding X1 is constant
= partial Correlation between and , holding Y constant
∑
= simple/pair wise correlation coefficient b/n Y & : = = 0.9394
√(∑ )(∑ )
∑
= simple/pair wise correlation coefficient b/n Y & : = = 0.997
√(∑ )(∑ )
∑
= simple/pair wise correlation coefficient b/n & : = = 0.9487
√(∑ )(∑ )
The three partial correlation coefficients, for our numerical example, are computed as
follows:
= : Partial correlation between Y and while holding constant
√( )( )
=
√( )( )
=
√( )( )
Partial correlation coefficients range in value from -1 to +1. For example, -1 refers to the
case where there is an exact or perfect negative linear relationship. However, +1
indicates a perfect positive linear net relationship And zero partial correlation
indicates no linear relationship between variables. Partial correlation coefficients are
used to determine the relative importance of different explanatory variables in a
multiple regression. For our example above we conclude that is more important than
in explaining the variation of Y, since > .
In above example, .
∑ ∑ (̂ ∑ ̂ ∑ )
̂ = =
( )
̂ = 0.7 ̂ √̂ = √ = 0.866
̅ ∑ ̅ ∑ ̅ ̅ ∑
b) (̂ ) [ ∑ ∑ (∑ )
] Hence: (̂ ) √ (̂ )
(̂ ) * += 30.62 (̂ ) √ = 5.533
∑
c) (̂ ) ∑ ∑ (∑ )
Hence: (̂ ) √ (̂ )
∑
d) (̂ ) ∑ ∑ (∑ )
Hence: (̂ ) √ (̂ )
The - statistics
Its most common use is to test the statistical significance of the joint effects of
regression coefficients/ explanatory variables. The ratio or statistics is used to test the
overall significance of the regression model. A high value for the F - statistic suggests a
significant relationship between the dependent and independent variables, leading to
the rejection of the null hypothesis which asserts the slope coefficients of all
explanatory variables are jointly zero.
;
The F- test is used to determine the adequacy of the regression model for prediction
purposes. If the F- test showed the regression coefficients are jointly insignificant, the
model can’t be used for prediction. The F – test is equivalent to testing the statistical
significance of of a model.
Decision rule:
o If F – statistics > F- critical value, reject
o If F – statistics > F- critical value; accept or can’t be rejected
⁄
; k - 1 and n- k are the degree of freedoms.
⁄
= =
Decision: Since F – statistics > F- critical value, we reject or can’t accept , hence,
we conclude that the model statistically significantly explain the variation in the
dependent variable.
ANOVA Table
Source SS df MS
Model 270.5 2 135.25 F( 2, 2) = 180.33] Prob > F 0.0055
Residual 1.5 2 0.75 R2 = 0.9945, ̅ = 0.9890
Total 272 4 68 Root MSE = 0.86603
The upper part of the table could be called the ANOVA table (since it tabulates the
Analysis of Variances: ESS, RSS, TSS, and MS (mean sum, which obtained by dividing the
Sum Square (the SS) by the corresponding df. Root MSE is standard error of disturbance
term. Note the result perfectly much with our manual computation. However, the uses
of statistical software tremendously reduce the computational complexity involved. As
Chapter Three Multiple Linear Regression
the number of explanatory variables increases manual computation will become very
difficult and unpractical.