SEM Essentials
SEM Essentials
(SEM) Essentials
by
Jim Grace
1
Outline
I. Essential Points about SEM
II. Structural Equation Models: Form and Function
I. SEM Essentials:
1. SEM is a form of graphical modeling, and therefore, a system
in which relationships can be represented in either graphical
or equational form.
equational
form
graphical
form
y1 = 11x1 + 1
x1
11
1
y1
Complex
Hypothesis
x1
y3
3
y1
Corresponding
Equations
e.g.
y2
1
y1 = 11x1 + 1
y2 = 21y1 + 21x1 + 2
y3 = 32y2 + 31x1 + 3
Spearman
(1904)
Pearson
(1890s)
r, ch
path analysis
r
facto
are
u
q
s
-
Joreskog
(1973)
s is
a na ly
oo
h
i
el
lik
Conventional
Statistics
Fisher
(1922)
Neyman & E. Pearson
(1934)
SEM
Lee
(2007)
tes
a lt . t in g
mod
els
Wright
(1918)
Raftery
(1993)
Bayesian
Analysis
The Scientific
Enterprise
8
structural
equation
modeling
univariate
descriptive statistics
univariate data
modeling
Data
exploration,
methodology and
theory development
simplistic
models
Understanding of Processes
modified from Starfield and Bleloch (1991)
realistic
predictive models
detailed process
models
10
one aim of
SEM
confirmatory/
hypothesis-testing
applications
11
12
Multivariate Models
x1
x1
y1
x2
x2
y2
x3
y1
x3
y3
x4
x4
y4
x5
x5
y5
15
16
Some Terminology
direct effect of x1 on y2
path
coefficients
x1
exogenous
variable
21
11
21
y1
1
indirect effect of x1 on y2
is 11 times
21
y2
2
endogenous
variables
17
A
x1
y1
B
x1
y2
y1
1
C
y2
1
D
x1
y2
x1
y2
2
x2
y1
2
x2
nonrecursive
y1
1
18
x1
y1
.40
x2
x2
y1
----------------------------x1
1.0
x2
0.40
1.0
y1
0.50
0.60
1.0
19
x1
11 = .50
y1
21 = .60
y2
x1
y1
y2
------------------------------------------------x1
1.0
y1
0.50
1.0
y2
0.30
0.60
1.0
(gamma) used to represent effect of exogenous on endogenous.
(beta) used to represent effect of endogenous on endogenous.20
x1
.50
y1
.60
y2
y1
y2
------------------------------------------------x1
1.0
y1
0.55
1.0
y2
0.50
0.60
x1
r = .55
y1
1.0
r = .60
y2
22
x1
additional process
y2
y1
Fourth Rule of Path Coefficients: when variables are
connected by more than one causal pathway, the path
coefficients are "partial" regression coefficients.
Which pairs of variables are connected by two causal paths?
answer: x1 and y2 (obvious one), but also y1 and y2, which are connected by
the joint influence of x1 on both of them.
23
x1
y1
x2
A case of shared causal influence: the unanalyzed relation
between x1 and x2 represents the effects of an unspecified
joint causal process. Therefore, x1 and y1 connected by two
causal paths. x2 and y1 likewise.
24
x1
.40
.31
y1
y2
.48
from Sourceforge.net
26
x1
.40
equation for path
from error variable
1 R
2
yi
.31
R2 = 0.16
y1
.92
R2 = 0.44
y2
.48
1
.84
.73
2
.56
alternative is to
show values for zetas,
which = 1-R2
27
R2 = 0.25
.50
y2
.40
y1
x1
x1
y1
y2
------------------------------x1
1.0
y1
0.40
1.0
y2
0.50
0.60
R2 = 0.16
1.0
28
R2 = 0.25
.50
y2
2
.40
x1
.40
y1
R2 = 0.16
.15
.64
.80
x2
-.11
Total Effects:
x1
x2
y2
.27
y1
1
y1
------------------------------y1
0.64 -0.11 --y2
0.32
-0.03 0.27
.15
.64
.80
x2
-.11
y2
.27
y1
x1
x2
y1
y2
----------------------------------------------x1
1.0
x2
0.80
1.0
y1
0.55
0.40
1.0
y2
0.30
0.23
0.35
1.0
32
Latent Variables
Latent variables are those whose presence we suspect or
theorize, but for which we have no direct measures.
fixed loading*
Intelligence
latent variable
1.0
IQ score
1.0
observed indicator
error
variable
33
34
Range of Examples
single-indicator
estimate
from map
multi-method
Elevation
soil C
loss on
ignition
repeated measures
singing range, t1
singing range, t2
singing range, t3
repeatability
observer 1
Territory
Size
Soil
Organic
Caribou
Counts
observer 2
35
0.60
y
R2 = 0.30
36
Example
Imagine that some of the observed variance in x is
due to error of measurement.
calibration data set based
on repeated measurement trials
plot
x-trial1 x-trial2 x-trial3
1
1.272 1.206 1.281
2
1.604 1.577 1.671
3
2.177 2.192 2.104
4
1.983 2.080 1.999
.
........
........
.......
n
2.460 2.266 2.418
imagine in this case VARx = 3.14, so error variance = 0.19 x 3.14 = 0.60
.60
.90
LV1
.65
LV2
1.0
R2 = .42
y
37
38
y = + y + x +
= p x 1 vector of intercepts
= p x p coefficient matrix of ys on ys
= cov () = q x q matrix of
y = p x 1 vector of responses
= p x q coefficient matrix of ys on xs
39
= + + +
The LISREL
Equations
Jreskg 1973
where:
is a vector of latent responses,
is a vector of latent predictors,
and are matrices of coefficients,
is a vector of errors for , and
is a vector of intercepts for
x = x +
y = y +
where:
x is a vector of loadings that link observed x
variables to latent predictors,
y is a vector of loadings that link observed y
variables to latent responses, and
and are vectors are errors
40
2. Estimation Methods
(a) decomposition of correlations (original path analysis)
(b) least-squares procedures (historic or in special cases)
(c) maximum likelihood (standard method)
(d) Markov chain Monte Carlo (MCMC) methods
(including Bayesian applications)
41
Bayesian References:
Bayesian SEM:
Lee, SY (2007) Structural Equation Modeling: A Bayesian
Approach. Wiley & Sons.
Bayesian Networks:
Neopolitan, R.E. (2004). Learning Bayesian Networks. Upper
Saddle River, NJ, Prentice Hall Publs.
42
r = 0.86
r = 0.50
compare
Observed Correlations*
{ }
1.0
.24 1.0
.01 .70 1.0
Model-Implied Correlations
{ }
11
12 22
13 23 33
3. Evaluation
Observed Covariance Matrix
Hypothesized Model
x1
y2
y1
(e.g
Parameter
Estimates
S=
)
tion lihood
a
m
esti m like
mu
i
x
a
., m
Model Fit
Evaluations
{ }
1.3
.24 .41
.01 9.7 12.3
compare
{ }
11
=
12 22
13 23 33
Implied Covariance Matrix
45
Fitting Functions
The most commonly used fitting function in maximum likelihood
estimation of structural equation models is based on the log
likelihood ratio, which compares the likelihood for a given model to
the likelihood of a model with perfect fit.
FML
log tr S log S p q
Note that when sample matrix and implied matrix are equal, terms 1 and
3 = 0 and terms 2 and 4 = 0. Thus, perfect model fit yields a value of FML
of 0.
47
The 2 Test
One of the most commonly used approaches to performing such tests (the
model 2 test) utilizes the fact that the maximum likelihood fitting function
FML follows a X2 (chi-square) distribution.
X2 = n-1(FML)
Here, n refers to the sample size, thus X2 is a direct function of sample
size.
49
0.40
y1
y2
0.50
1.0
0.4 1.0
0.35 0.5
1.0
51
y1
y2
x
y1
y2
x
-------- -------- -------1.00
0.50
1.00
0.40
0.35
1.00
residual = 0.15
53
54