Professional Documents
Culture Documents
1 2
Y1 Y2 Y3
One variable, say Party Identification , influencing itself over 3 waves (autoregressive
model)
Two equations, one predicting PID in wave 2, one predicting PID in wave 3
Two endogenous variables (PID2 and PID3), one exogeneous variable (PID1)
It is a recursive model, since no causal feedback effects between variables
1 and 2 are the structural effects linking PID in waves 1-2 and 2-3. They are
sometimes in longitudinal models called the stability effects
1 and 2 are the disturbance or error terms, unobserved influences and
idiosyncratic errors that predict PID in waves 2 and 3
If we express all variables as deviations from their respective means, there is no need
for an intercept term in the Y2 or Y3 equations
Analysis of Panel Data, University of Gothenburg, 15-18 June 2015
1 2
1 2
Y1 Y2 Y3
1
Correlations
1 1
Implied by
the Causal
Model
11 2 1
Analysis of Panel Data, University of Gothenburg, 15-18 June 2015
Step 2: Solve for unknown structural parameters
What are unknowns? 1 and 2
Easy to solve for each:
1 = r(Y1Y2)
2 = r(Y2Y3)
1
Observed .872 1
.856 .884 1
1
Implied 1 1
11 2 1
Analysis of Panel Data, University of Gothenburg, 15-18 June 2015
Estimates and Model Fit
1 = r(Y1, Y2) = .872
2 = r(Y2,Y3) = .884
1 2
Y2 Y3
Y1
3
Y1 Y2 Y3 Y1 Y2 Y3
Y1 1 Y1 1
Y2 r(Y1Y2 ) 1 Y2 1 1
Y3 r(Y1Y3 ) r(Y2Y3 ) 1 Y3 1 2 + 3 2 + 13 1
One measure along these lines that is very popular in the SEM literature is the
Root Mean Square Error of Approximation (RMSEA),:
m2 df m
RMSEA =
(N 1)df m
which in this case is: square root of 108.76/737=.384
It is desirable for this value NOT TO EXCEED .10 for models to be acceptable!!!
Y1 Y2
1
pid2000 pid2002 1
2
rX1Y1
4
app2000
3 app2002 2
X1 3 X2
Term Multiplier) 1 1
(b) Y = ( 1 1)Y1 + ( 5 + 2 ) X 1 + 2 X + 1
In Time Series language:
Model a: Autoregressive Distributed Lag Model (ADL, 1, 1)
Model b: Error Correction Model (ECM)
BUT in a multiple causal system, including both synchronous and
lagged effects in both directions makes the model non-recursive, and
underidentified. Need more information, as we will see !!!
>
1.5 Information criteria
app2000 .16
app2002 2 .71 AIC 10452.701 Akaike's information criterion
1 1.6 BIC 10512.552 Bayesian information criterion
>
Baseline comparison
CFI 0.943 Comparative fit index
TLI 0.715 Tucker-Lewis index
>
Size of residuals
Modification indices SRMR 0.035 Standardized root mean squared residua
> l
Standard CD 0.791 Coefficient of determination
MI df P>MI EPC EPC
Structural
>
pid2002 <-
app2002 77.161 1 0.00 .2556865 .1830259
app2002 <-
pid2002 77.161 1 0.00 .4089136 .571251
1.4 >
.79
pid2000 pid2002 1 .23 Population error
1 .11 RMSEA 0.000 Root mean squared error of approximati
> on
90% CI, lower bound 0.000
.62 upper bound 0.000
.43 .32
.13 pclose 1.000 Probability RMSEA <= 0.05
>
Information criteria
1.5 AIC 10373.201 Akaike's information criterion
app2000 .16
app2002 2 .71 BIC 10437.656 Bayesian information criterion
1 1.6
>
Baseline comparison
CFI 1.000 Comparative fit index
TLI 1.000 Tucker-Lewis index
>
Size of residuals
SRMR 0.000 Standardized root mean squared residua
> l
CD 0.773 Coefficient of determination
>
3 4
4
5 6
1 2 3
4
5 6
(5)
Y1 = 1 X 1 + 2
X 1 = 2Y1 + 1
X 1 = 2 ( 1 X 1 + 2 ) + 1
X 1 = 2 1 X 1 + 2 2 + 1
pid2000 pid2002
app2000 app2002
Y2 = 3 X 2 + 1Y1 + 1
X 2 = 4Y1 + 2 X 1 + 2
Analysis of Panel Data, University of Gothenburg, 15-18 June 2015
ML Estimation of the Model
Each of the equations is just-identified: one included
endogenous variable and one excluded exogenous variable.
X1 is an instrument for X2 in the Y2 equation
Y1 is an instrument for Y2 in the X2 equation.
We can arrive at the coefficients through algebraic manipulation
multiply each endogenous equation by each of the wave 1 variables
(X1,Y1) and solve for the 4 structural effects in terms of the observed
covariances. This is tedious algebra but follows the same logic as
what we have been doing so far
ML estimation also follows the same logic as before.
10 observed variances-covariances, 9 unknowns (4 structural effects,
2 disturbances, 3 variances-covariances of wave 1 variables)
Model as a whole has one over-identifying restriction (either one of
the cross-lagged effects is NOT zero, or the disturbance covariance
between the is NOT zero).
The model 2 provides a test of the restriction.
3 4
Likelihood ratio
chi2_ms(6) 134.899 model vs. saturated 1 .23 2 .21
p > chi2 0.000
chi2_bs(14) 3458.563 baseline vs. saturated 1.4
.82
p > chi2 0.000 pid2000 .8 pid2002 pid2004
1 .11 -.093
Population error .1
.62 .12
RMSEA 0.171 Root mean squared error of approximation -.0092 .52 .0089 -.008 .48 .0098
90% CI, lower bound 0.146
upper bound 0.196 1.5
.1
pclose 0.000 Probability RMSEA <= 0.05 app2000 app2002 .43 app2004
1 1.5 .16
Information criteria
AIC 14813.960 Akaike's information criterion 3 .65 4 .36
BIC 14910.643 Bayesian information criterion
Baseline comparison
CFI 0.963 Comparative fit index
TLI 0.913 Tucker-Lewis index
Size of residuals
SRMR 0.029 Standardized root mean squared residual
CD 0.777 Coefficient of determination
Modification indices
Modification indices shows some unexplained
covariation between PID2000 and PID2004
Standard
MI df P>MI EPC EPC
(Presidential Year Effect). Options: add
Structural
pid2002 <- either structural disturbance covariance, or a
pid2004 102.723 1 0.00 -.398454 -.4252892
app2004 15.564 1 0.00 -.1289866 -.1068166
direct causal effect from PID2000->PID2004
pid2004 <-
pid2000 105.332 1 0.00 .3849349 .3561638
app2000 24.988 1 0.00 .1498549 .1073146
app2004 <-
app2000 6.171 1 0.01 .0705983 .0651619
Likelihood ratio
chi2_ms(5) 21.595 model vs. saturated
1 .23
p > chi2 0.001
chi2_bs(14) 3458.563 baseline vs. saturated
2 .18
p > chi2 0.000
pid2002
.13 .52
Population error 1.4 .8
RMSEA 0.067 Root mean squared error of approximation pid2000 pid2004
1 .36 -.14
90% CI, lower bound 0.040
upper bound 0.097 .12
.57 -.011 -.041
.11
pclose 0.141 Probability RMSEA <= 0.05 .62
-.048 .52 -.012
Information criteria
AIC 14702.656 Akaike's information criterion 1.5
.1
BIC 14803.943 Bayesian information criterion app2000 app2002 .42 app2004
1 1.5 .16
Baseline comparison
CFI 0.995 Comparative fit index 3 .65 4 .36
TLI 0.987 Tucker-Lewis index
Size of residuals
SRMR 0.016 Standardized root mean squared residual
CD 0.801 Coefficient of determination
Yi = 1 X i + i
(3) and x *i = X i + vi
then Yi = 1 (x vi ) + i *
i
and Yi = 1x *i + ( i 1vi )
xi* is related to the error term (since vi and xi* are related). So
fallible xi* is endogenous and we get inconsistent estimates of the
causal effect 1
1
Y1 Y2
y1* y2*
w1 w2
Subtract equation 6(a) from 6(b) to arrive at an expression for the change in y:
y* = Y + (w2 w1 )
Multiply equation by y*1 in the first step, substitute (Y1+w1) for y*1 in one term on
the right side of the equation in the second step, and take expectations to yield:
So the more measurement error in y1, the more we overestimate the amount of
negative covariation between the variables initial level and change if we dont
take the measurement error into account. That means that, with measurement
error in the indicators, we think that there is more regression to the mean than
there really is and more speed toward equilibrium than there really is.
In terms of the level of the dependent variable, we think there is less stability
from one time point to the next when we analyze the fallible indicators than
there really is in the true score latent variables. And to the extent that
exogenous variables X are related to the initial level of y,* we will consequently
wrongly estimate their impact on y2* as well.
1 2
Y1 Y2 Y3
w1 w2 w3
4.68
S = 4.11 4.75
4.30 4.47 5.42
Results
Notes:
1 .26
2 .27
Zero df, chi-square=0.0
Estimated error variance of single
1
indicator of PID=.48, as we
.98
PID2000 PID2002 PID2004
4.2
on this!)
6 .69
7 .38
1.5
app2000 app2002 .44 app2004
1 .1 2.2 .73
.65 .0068
.043
.48 .43 .46
.47
.94 .97
PID2000 PID2002 PID2004
1
3 .1 4 .099 5 .087
Eval1 Eval2
1 1
2 3 4 5
Baseline comparison
CFI 0.970 Comparative fit index
TLI 0.818 Tucker-Lewis index
Size of residuals
SRMR 0.025 Standardized root mean squared residual
CD 0.813 Coefficient of determination
1 .24
unstandardized
.87
Eval1 Eval2
1
educ
Eval1 Eval2
1 1
2 3 4 5
.12 .9
male
1 1 .46
2.9
.048
educ -.12
1
Notes: .13
.13
.22
.016 Eval1
1
.62
Eval2 from PID onto
.6
Eval (.25),
.65 .5 .7
weaker but still
Moral2000 InTouch2000 Moral2004
2.9
InTouch2004
2.1
significant effect
from Eval to
3.6 3
4 5
PID (.15)!
.51 .64
2 .58 3 .75
.13
.11
Or, in a standardized solution (where y and are both standardized), the reliability
of yi will be 2.