0% found this document useful (0 votes)
61 views

Lecture 11

This document introduces matrix notation for linear regression analysis. Key points include: - Random vectors and matrices can be used to represent regression models and parameters. - The covariance matrix describes the variances and covariances between elements in a random vector. - Linear regression equations, least squares estimation of parameters, fitted values, and residuals can all be expressed compactly using matrices. - Matrix properties like idempotence help simplify expressions for quantities like residuals and their covariance. - Common regression outputs like total sum of squares, regression sum of squares, and error sum of squares can be derived and expressed using matrices.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
61 views

Lecture 11

This document introduces matrix notation for linear regression analysis. Key points include: - Random vectors and matrices can be used to represent regression models and parameters. - The covariance matrix describes the variances and covariances between elements in a random vector. - Linear regression equations, least squares estimation of parameters, fitted values, and residuals can all be expressed compactly using matrices. - Matrix properties like idempotence help simplify expressions for quantities like residuals and their covariance. - Common regression outputs like total sum of squares, regression sum of squares, and error sum of squares can be derived and expressed using matrices.
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 36

Matrix Approach to Linear

Regression
Dr. Frank Wood

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 1


Random Vectors and Matrices
• Let’s say we have a vector consisting of three
random variables

The expectation of a random vector is defined

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 2


Expectation of a Random Matrix
• The expectation of a random matrix is defined
similarly

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 3


Covariance Matrix of a Random Vector
• The collection of variances and covariances of and
between the elements of a random vector can be
collection into a matrix called the covariance matrix

remember

so the covariance matrix is symmetric

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 4


Derivation of Covariance Matrix
• In vector terms the covariance matrix is
defined by

because

verify first entry

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 5


Regression Example
• Take a regression example with n=3 with
constant error terms σ{ǫi} = σ and are
uncorrelated so that σ{ǫi, ǫj} = 0 for all i ≠ j
• The covariance matrix for the random vector ǫ
is

which can be written as

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 6


Basic Results
• If A is a constant matrix and Y is a random
matrix then

is a random matrix

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 7


Multivariate Normal Density
• Let Y be a vector of p observations

• Let µ be a vector of p means for each of the p observations

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 8


Multivariate Normal Density
• Let § be the covariance matrix of Y

• Then the multivariate normal density is given


by

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 9


Example 2d Multivariate Normal Distribution
mvnpdf([0 0], [10 2;2 2])

0.04
0.02
0

10

2
10
0
8
6
-2
4
-4 2
0
-6 -2
-4
y -8
-8
-6
Run multivariate_normal_plots.m
-10 x
-10
Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 10
Matrix Simple Linear Regression
• Nothing new – only matrix formalism for
previous results
• Remember the normal error regression model

• This implies

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 11


Regression Matrices
• If we identify the following matrices

• We can write the linear regression equations


in a compact form

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 12


Regression Matrices
• Of course, in the normal regression model the
expected value of each of the ǫi’s is zero, we
can write

• This is because

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 13


Error Covariance
• Because the error terms are independent and
have constant variance σ

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 14


Matrix Normal Regression Model
• In matrix terms the normal regression model
can be written as

where

and

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 15


Least Squares Estimation
• Starting from the normal equations you have
derived

we can see that these equations are


equivalent to the following matrix operations

with

demonstrate this on board


Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 16
Estimation
• We can solve this equation

(if the inverse of X’X exists) by the following

and since

we have
Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 17
Least Squares Solution
• The matrix normal equations can be derived
directly from the minimization of

w.r.t. to β

Do this on board.

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 18


Fitted Values and Residuals
• Let the vector of the fitted values be

in matrix notation we then have

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 19


Hat Matrix – Puts hat on Y
• We can also directly express the fitted values
in terms of only the X and Y matrices

and we can further define H, the “hat matrix”

• The hat matrix plans an important role in


diagnostics for regression analysis.
write H on board

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 20


Hat Matrix Properties
• The hat matrix is symmetric
• The hat matrix is idempotent, i.e.

demonstrate on board

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 21


Residuals
• The residuals, like the fitted values of
\hat{Y_i} can be expressed as linear
combinations of the response variable
observations Yi

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 22


Covariance of Residuals
• Starting with
we see that
but
which means that

and since I-H is idempotent (check) we have

we can plug in MSE for σ as an estimate

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 23


ANOVA
• We can express the ANOVA results in matrix
form as well, starting with

where

J is matrix of all ones, do 3x3 example


leaving

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 24


SSE
• Remember

• We have

derive this on board

and this

• Simplified

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 25


SSR
• It can be shown that
– for instance, remember SSR = SSTO-SSE

write these on board

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 26


Tests and Inference
• The ANOVA tests and inferences we can
perform are the same as before
• Only the algebraic method of getting the
quantities changes
• Matrix notation is a writing short-cut, not a
computational shortcut

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 27


Quadratic Forms
• The ANOVA sums of squares can be shown
to be quadratic forms. An example of a
quadratic form is given by

• Note that this can be expressed in matrix


notation as (where A is a symmetric matrix)

do on board
Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 28
Quadratic Forms
• In general, a quadratic form is defined by

A is the matrix of the quadratic form.


• The ANOVA sums SSTO, SSE, and SSR are
all quadratic forms.

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 29


ANOVA quadratic forms
• Consider the following rexpression of b’X’

• With this it is easy to see that

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 30


Inference
• We can derive the sampling variance of the β
vector estimator by remembering that

where A is a constant matrix

which yields

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 31


Variance of b
• Since is symmetric we can write

and thus

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 32


Variance of b
• Of course this assumes that we know σ. If
we don’t, we, as usual, replace it with the
MSE.

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 33


Mean Response
• To estimate the mean response we can
create the following matrix

• The fit (or prediction) is then

since

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 34


Variance of Mean Response
• Is given by

and is arrived at in the same way as for the


variance of \beta
• Similarly the estimated variance in matrix
notation is given by

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 35


Wrap-Up
• Expectation and variance of random vector
and matrices
• Simple linear regression in matrix form

• Next: multiple regression

Frank Wood, [email protected] Linear Regression Models Lecture 11, Slide 36

You might also like