Multiple Regression Analysis - Estimation
Multiple Regression Analysis - Estimation
Estimation
Multiple Regression Model
y = 0 + 1x1 + 2x2 + + kxk + u
Regression: Q = 0 + 1P + 2I + 3A + u
Interpretation:
- measures the effects of the tile price on the tile consumption,
1
-Goal: Estimate , , , 0 1 2 3
min
n n
ui
2
yi 0 1 Pi 2 I i 3 Ai
2
min
0, 1, 2, 3 i 1
0, 1, 2, 3 i 1
The Generic Multiple Regression
Model Y X Y X ... X
i 0 1 i 1,..., n
1i 2 2i k ki i
i
j
X ji
Y1 1 X11 X12 ... X1k
Y 1 X X 22 ... X 2 k
Y 2 X 21
Y
n nx1 1 X n1 X n 2 ... X nk nx ( k 1)
0 1
1 2
k ( k 1) x1 n nx1
Estimation of regression parameters:
-Least Squares (no knowledge of the distribution of the error or disturbance terms is required).
-The use of the matrix notation allows a view of how the data are housed in software programs.
Components of the Model
Captures:
1. Omission of the influence of other
variables.
2. Measurement error.
0
xT is the transpose of x ((k 1) xn) 1 ((k 1) x1)
k
The Gauss-Markov Theorem
Given the assumptions below, it can be shown that the OLS
estimator is BLUE.
- Best
- Linear
- Unbiased
- Estimator
Assumptions:
- Linear in parameters
- Corr (i, j) = 0
- Zero mean
- No perfect collinearity
- Homoscedasticity
Communication and
Aims for the Analyst
Communication
- A technician can run a program and get output.
- An analyst must interpret the findings from examination of this output.
- There are no bonus points to be given to terrific hackers but poor
analysts.
Aims
1. Improve your ability in developing models to conduct structural analysis and to
forecast with some accuracy.
2. Enhance your ability in interpreting and communicating the results, so as to
improve your decision-making.
Bottom Line
1. The analyst transforms the economic model/idea to a mathematical/statistical
one.
2. The technician estimates the model and obtains a mathematical/statistical
answer.
3. The analyst transforms the mathematical/statistical answer to an economic one.
Goodness-of-Fit
yi y i ui
Definitions :
iy y 2
is the total sum of squares (SST)
i
y y 2
is the regression sum of squares (SSR)
R = SSR/SST = 1 SSE/SST
More about R-Squared
2 SSE /( n k 1)
Adjusted R R 1
SST /(n 1)
Questions:
(a)Why do we care about the adjusted R ?
(b)Is adjusted R always better than R ?
(c)Whats the relationship between R and adjusted R ?
Model Selection Criteria
T
t
e 2
MSE t 1
T
T
T e 2
t
T
s 2 t 1
MSE
T p T T p
T
Is the " penalty factor"
Tp
2P
Akaike Information Criterion (AIC) AIC e T
MSE
(SIC) or (BIC)
p
SIC ( BIC ) T T
MSE
p is the number of parameters to be estimated
Model Selection Criteria Example
s
2 2
u n k 1 SSE df
2
i
-df = n (k + 1), or df = n k 1
-df (i.e. degrees of freedom) is the (number of
observations) (number of estimated parameters)
Variance of OLS Parameter
Estimates
Var s 2 (x T x) -1
T 97
s2 MSE (16.29) 17.56
T-P 90
2P 14
AIC e T MSE e 97 (16.29) 18.82
P 7
SIC T T MSE 97 97 (16.29) 22.67