Data Analysis With Mplus
Data Analysis With Mplus
This series provides applied researchers and students with analysis and research
design books that emphasize the use of methods to answer research questions.
Rather than emphasizing statistical theory, each volume in the series illustrates when
a technique should (and should not) be used and how the output from available
software programs should (and should not) be interpreted. Common pitfalls as well
as areas of further development are clearly articulated.
R e c en t vol u m e s
Christian Geiser
I’m very pleased to introduce Christian Geiser’s book to you. This book
was originally published in German by Springer Verlag. I felt it was so
important to bring the book to a broader readership that we secured the
rights to produce it in English as part of The Guilford Press series Meth-
odology in the Social Sciences. Given that Christian is a very talented
multilingual scholar, I knew that rendering this excellent work in English
would be both easy for him and an important contribution. Special thanks
go to the legal and contracts team at The Guilford Press who worked dili-
gently to secure the English-language rights to Christian’s book and to
Whitney Moore, whose assistance in bringing this book to fruition was
above the call of duty.
Why am I so pleased to bring this book to you? Simply, it is the most
thorough and accessible book available on how to use Mplus for all man-
ner of statistical analysis procedures. Mplus is a popular software package
that is capable of estimating a wide variety of useful analyses. Figuring
out how to get Mplus to do most of the analyses, however, is both chal-
lenging and frustrating if you rely on the Mplus User’s Guide. Christian’s
book is clearly organized, covers a wide variety of analyses that Mplus can
estimate, and is easy to follow.
In other words, Data Analysis with Mplus fills an important gap. It is
not focused simply on a cookbook model of how to run the various analy-
ses in Mplus. Because it is written for the applied researcher and not the
quantitative analyst, the models are accessible. Christian explains each
v
vi Series Editor’s Note
analysis model step-by-step. He does not presume that you will have used
Mplus before. In fact, he actually begins with how to prepare and bring in
data. This feature is a rarity and an important one because it is an essen-
tial step to master before you can do any of the fun analyses!
A wide variety of different analyses are described in this book. In a
clear and methodical manner, Christian explains what the syntax means,
teaches you very useful Mplus tricks, and actually alerts you to the defaults
that exist (which can be a mystery sometimes). Because of his skillful syn-
tax pedagogy, new users will feel comfortable knowing what the syntax
means and not be intimidated about using it, which is important for new
researchers who do not have much exposure to the world of writing syn-
tax for data analyses.
Christian’s focus on different types of analyses will help build your
knowledge and understanding of Mplus syntax. For example, some analy-
ses are presented first as manifest variable techniques before the complex-
ity of a measurement model is added. This piecewise exposure helps avoid
the overwhelming effect of presenting both manifest and latent variables
in the same model and allows for highlighting important differences in
the required syntax. Christian does a terrific job of explaining Mplus
syntax in bite-size pieces so that a reader without prior knowledge can
understand it. This aspect of his book is essential, I think, because often it
is the failure of nonquantitative users to grasp a simple, or basic, concept
that results in their not understanding the following steps. The chapter-
specific recommendations focus on the applied researcher.
Christian’s book is an excellent resource manual for anyone wishing
to use Mplus for advanced statistical data analyses. It is a terrific comple-
ment to Rex B. Kline’s Principles and Practices of Structural Equation Model-
ing, Timothy A. Brown’s Confirmatory Factor Analysis for Applied Research,
as well as my forthcoming book, Longitudinal Structural Equation Modeling.
Todd D. Little
Berlin, Germany
Preface
The book starts with the question of how to prepare a data set for use
in Mplus. Given the great popularity of SPSS in the social sciences, I chose
to demonstrate the key steps in data preparation and data export using
this general statistics program. The basic principles can be easily trans-
ferred to other statistics programs, such as R, SAS, STATA, STATISTICA,
and Systat.
The book is meant to be more than a pure Mplus syntax guide, and
it is not a replacement for the Mplus User’s Guide (Muthén & Muthén,
1998–2012; www.statmodel.com/ugexcerpts.shtml). My goal is not only to
show the reader how to specify different types of models using Mplus
syntax, but also to provide guidelines for useful modeling strategies and
the interpretation of the results. Furthermore, the book also points out
some typical caveats in the analysis of complex multivariate data (e.g., the
question of how to assess measurement invariance in longitudinal SEMs
or how to avoid local likelihood maxima in LCA) and how to properly
address them. Specific issues and solution strategies are discussed in vari-
ous text boxes throughout.
Despite the focus on application, the theory of each of the statisti-
cal approaches is briefly reviewed at the beginning of each chapter. It is
important to note that the book is not a replacement for relevant statis-
tics textbooks or for specialized literature. Therefore, key references to
relevant literature are provided at the beginning of each chapter and for
the discussion of specific problems, such as the question of how to most
appropriately test the mediated effect in path analysis (see Section 3.5 in
Chapter 3).
The choice of statistical methods was made on the basis of my experi-
ences as a lecturer for numerous statistical workshops, focusing on what
seemed to be the needs of a large number of empirical scientists. I hope
that the selection of methods and models presented in this book will be
useful for researchers who want to use these techniques in their research.
The use of more complex statistical methods is discussed in Geiser,
Crayen, and Enders (2012).
The present book evolved from my seminar and workshop teaching
experiences. I am grateful to my students and to the participants in my
workshops for the feedback provided in these courses. In addition, I would
like to thank Martin Corth, David Cole, Mark Roosa, and the Deutsches
Zentrum für Altersfragen (German Center for Gerontology) for letting me
use data sets as examples in the book. I also want to thank Christopher
Marx for his help in preparing and analyzing the data set used to illus-
trate path modeling and mediation analysis in Section 3.5 as well as for
x Preface
Christian Geiser
Logan, Utah
Contents
xi
xii Contents
References 285
In this chapter, I explain how to prepare an SPSS raw data set for use in
Mplus. The procedure is demonstrated in the program SPSS, because it
is one of the most widely used general statistics software packages in the
social and behavioral sciences. The essential steps shown here can be
transferred easily to other common programs, such as R, SAS, STATA,
STATISTICA, SYSTAT, etc.
Mplus cannot directly read data sets in the common SPSS data for-
mat (*.sav). However, it is easy to export a data set as a simple text file (a
file with the typical file extensions *.txt or *.dat) from SPSS or another
general data management program. Such a text file can then be used for
statistical analyses in Mplus. In my experience, the text file format that
works best is a simple tab-delimited format. Tab-delimited means that each
column (typically representing a variable) in the text file is separated by
a tab symbol. Reading data in other formats is possible (see Mplus User’s
Guide; Muthén & Muthén, 1998–2012), and one may also use summary
data (e.g., covariance matrices, correlation matrices, means, and standard
deviations) instead of raw data, as described in Section 2.2.
In the following sections, a convenient procedure for preparing and
exporting a raw data set from SPSS is demonstrated, step by step, using
the sample data file KFT.sav. It should be noted that the method described
here is not the only possible way; other procedures yield the same results.
The sample data file as well as an SPSS syntax file with the relevant com-
mands can be found on the companion website.
1
2 Data Analysis with Mplus
The data set KFT.sav contains data from N = 455 German high school
students who were tested with a German intelligence test battery, Kogni-
tiver Fähigkeitstest (Cognitive Abilities Test, KFT; Heller, Gaedicke, & Wein-
läder, 1976). The KFT is frequently used to measure school achievement in
German-speaking countries. Subtests in the present data set included verbal
material (variables kft_v1 and kft_v2), numerical (“quantitative”) material
(variables kft_q1 and kft_q3), and figural (“nonverbal”) material (variables
kft_n1 and kft_n3). The data set contains the sum scores of the students for
each of the six KFT subtests.
FIGURE 1.2. To assign (or change) a numeric missing value code, all variables are
transferred to the window on the right-hand side.
4 Data Analysis with Mplus
FIGURE 1.3. The option Old and New Values is used to recode system- and user-
defined missing values into the same numeric code, here –99.
Missing values now
coded as –99
FIGURE 1.4. Modified data file with new missing value code –99.
FIGURE 1.5. SPSS does not automatically recognize the new missing value code
as “missing.” This code has to be manually assigned under “Missing” in the SPSS
variable view.
FIGURE 1.6. Entering the value –99 as a user-defined missing value in the SPSS
variable view.
FIGURE 1.7. Using the copy-and-paste function to assign the value –99 as missing
value to each of the six variables in the data set.
FIGURE 1.8. The SPSS variable view with correctly assigned missing value code
for all six variables.
Data Management in SPSS 7
After defining missing values appropriately, one can export the data file
into an ASCII format that is acceptable for Mplus. When using the point-
and-click interface in SPSS, one can simply use the option File → Save as
(as shown in Figures 1.9–1.10). Alternatively, the syntax command SAVE
TRANSLATE (see companion website) can be used for the same purpose.
The resulting tab-delimited text data file that can be processed by Mplus
is shown in Figure 1.11.
FIGURE 1.10. Deactivating the SPSS option Write Variable Names to Spread-
sheet.
FIGURE 1.11. Excerpt from the newly created text file that contains the variables
in tab-delimited format. This file can be processed by Mplus.
2
Mplus can process data in different formats. The two most relevant for-
mats are discussed here: individual data and summary data. The indi-
vidual data format is probably the most commonly used format in prac-
tice. Individual data are raw data in which the scores of all individuals
on all variables are preserved. This is the case, for example, in our KFT
data file (see Figure 1.11). The summary data format is used when one
wants to analyze data in summary format, for example, a covariance or a
correlation matrix (and potentially the means and standard deviations of
the variables). The possibility to analyze summary data can be useful, for
example, when a researcher wants to reanalyze data reported in a research
report (e.g., a journal article, grant proposal). Many scholarly publications
provide the covariance or correlation matrices that served as the basis for
the analyses (e.g., path or structural equation models [SEMs]), whereas
raw data are only rarely made available in publications. For many types of
structural equation analyses, it is not necessary to analyze individual data
(although using individual data may often be advantageous, e.g., when
the user wants to take missing data into account or use specific estima-
tion procedures, e.g., for clustered data). For many models, the covariance
matrix (sometimes in combination with the mean vector) of the variables
is sufficient as input for estimating the model. Let us first consider the
case of reading individual data, using the data example from Chapter 1.
9
10 Data Analysis with Mplus
FIGURE 2.1. Mplus syntax file for checking the correct data import using the
option type = basic. The menu option File → Save as is used here to save the
input as a *.inp file. It is most convenient to save the input file to the same directory
that also contains the data file to be analyzed. That way, no specific path has to be
included under data: file =.
FIGURE 2.2. Exporting variable names from SPSS via the menu option Utilities
→ Variables in SPSS. On the left-hand side, all variable names are highlighted. By
clicking on Paste, the variable names are added to an SPSS syntax window (see
Figure 2.3).
FIGURE 2.3. Variable names in the SPSS syntax window. The names can now be
copied and pasted into an Mplus syntax under variable: names =. Note that
the correct order of the variable names in the SPSS syntax file should be checked
before transferring the names to Mplus.
14 Data Analysis with Mplus
• In the Mplus syntax no differences are made between upper- and lower-
case letters. (In other words, Mplus is not case sensitive.)
• In addition, the order in which different commands are listed is largely
arbitrary.
• Every command line has to end with a semicolon (;).
• A single command line should not exceed 90 characters (in older Mplus
versions, only 80 characters were allowed per line). Longer lines can
occur, for example, when there is a long list of variables to read into Mplus
or when a long file path has to be specified. Lines that are too long will be
cut after 80/90 characters (depending on the program version), and Mplus
will ignore all following specifications in that line. Clearly, this action can
lead to very significant errors. Mplus will inform users about this problem
in the output by means of an error message. One should not overlook
or ignore this error message. A simple way to deal with the problem of
overly long lines is to simply break the lines using the ENTER key on the
keyboard.
• Variable names cannot be longer than eight characters.
• Each line of comment has to start with an exclamation mark. Comments
appear in green font in the input and output. Comments do not have to
end with a specific symbol; however, every new line of comment has to
start with an exclamation mark again.
INPUT INSTRUCTIONS
! This is a comment
variable: names = kft_v1 kft_v3 kft_q1 kft_q3 kft_n1 kft_n3;
missing = all(-99);
*** WARNING
Data set contains cases with missing on all variables.
These cases were not included in the analysis.
Number of cases with missing on all variables: 131
1 WARNING(S) FOUND IN THE INPUT INSTRUCTIONS
SUMMARY OF ANALYSIS
Number of groups 1
Number of observations 456
Continuous
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1 KFT_N3
Estimator ML
Information matrix OBSERVED
Maximum number of iterations 1000
Convergence criterion 0.500D-04
Maximum number of steepest descent iterations 20
Maximum number of iterations for H1 2000
Convergence criterion for H1 0.100D-03
This information is useful to check, among other things, that the cor-
rect data set was used as well as the intended variables. For Number of
observations we can see that 456 students contributed one or more
values (here, those cases that have missing data on all six variables are
already excluded). Next is a summary of the missing data patterns that
occurred in the present application:
SUMMARY OF DATA
1 2
KFT_V1 x x
KFT_V3 x x
KFT_Q1 x x
KFT_Q3 x x
KFT_N1 x x
KFT_N3 x
the pattern in which all values are missing, which is not shown). Column
1 (missing data pattern 1) contains only “x’s,” indicating that this is the
pattern in which no missing data occurred. The second column contains
“x’s” except for the last variable KFT_N3. This means that individuals
with pattern 2 have valid scores on all variables except for the variable
KFT_N3.
The section on MISSING DATA PATTERN FREQUENCIES shows
that of the 456 students, all but one student contributed complete data
(values on all six variables): 455 students showed missing data pattern 1
(frequency = 455), whereas only one student showed missing data pattern
2 (frequency = 1), that is, a missing value on KFT_N3.
The covariance coverage shows us the proportion of cases that con-
tributes values for the calculation of each variance or covariance. In our
case, for the variances and covariances, those are 100% of cases, with the
exception of the variance and covariances that are associated with vari-
able KFT_N3 (only 99.8% of cases contribute to this variable because
of the one student who has a missing value on this variable). The mini-
mum acceptable value for the covariance coverage is 10% according to
the Mplus default (“Minimum covariance coverage value 0.100”). If the
covariance coverage fell below this value, Mplus would no longer estimate
a model by default because the coverage would be seen as too weak.
Covariance Coverage
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1
________ ________ ________ ________ ________
KFT_V1 1.000
KFT_V3 1.000 1.000
KFT_Q1 1.000 1.000 1.000
KFT_Q3 1.000 1.000 1.000 1.000
KFT_N1 1.000 1.000 1.000 1.000 1.000
KFT_N3 0.998 0.998 0.998 0.998 0.998
Covariance Coverage
KFT_N3
________
KFT_N3 0.998
Note that the covariance coverage does not refer to the actual vari-
ances and covariances of the variables, but instead simply informs us about
the “completeness” of the data (the amount of data available to calculate
those statistics). The actual estimated sample variances and covariances
18 Data Analysis with Mplus
are obtained next, under the header RESULTS FOR BASIC ANALY-
SIS/ESTIMATED SAMPLE STATISTICS:
Means
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1 KFT_N3
________ ________ ________ ________ ________ ________
11.904 8.978 12.377 7.730 11.088 8.277
Covariances
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1
________ ________ ________ ________ ________
KFT_V1 21.592
KFT_V3 10.761 17.938
KFT_Q1 6.010 5.267 11.235
KFT_Q3 4.645 4.406 4.027 6.754
KFT_N1 11.072 11.644 6.870 6.111 29.826
KFT_N3 6.347 7.190 4.415 3.986 9.298
Covariances
KFT_N3
________
KFT_N3 11.782
Correlations
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1
________ ________ ________ ________ ________
KFT_V1 1.000
KFT_V3 0.547 1.000
KFT_Q1 0.386 0.371 1.000
KFT_Q3 0.385 0.400 0.462 1.000
KFT_N1 0.436 0.503 0.375 0.431 1.000
KFT_N3 0.398 0.495 0.384 0.447 0.496
Correlations
KFT_N3
________
KFT_N3 1.000
The descriptive statistics obtained for each of the six observed KFT
variables (i.e., the estimated means, variances, covariances, and product–
moment correlations of the variables) were estimated using the FIML pro-
cedure. Therefore, the statistics in this form are not directly comparable
to the corresponding statistics calculated in SPSS. The reason is that SPSS
statistics are based either on pairwise or listwise deletion of cases rather
than on FIML estimation.
In order to directly compare Mplus and SPSS descriptive statistics,
one can request the use of listwise deletion in Mplus rather than the
default (FIML). Note that we discourage the use of listwise deletion in
later analysis steps that involve actual model fitting. The procedure is
used here for technical reasons only, in order to facilitate the process
of checking the proper data entry into Mplus. In later analysis steps,
researchers should consider using more advanced missing data analytic
techniques, such as FIML, as described by Enders (2010) or by Schafer
and Graham (2002), because these techniques are usually more reason-
able than listwise deletion.
Listwise deletion of cases is obtained in Mplus by adding the subcom-
mand listwise = on; under the data command. The listwise
= on; subcommand deactivates the FIML procedure and excludes all
cases that have missing data on at least one of the variables included in
the analysis. Note that FIML has been the default since Mplus version 5.
In previous versions of the program through Mplus version 4, listwise
deletion was the default.
The extended Mplus syntax using listwise deletion is shown in Fig-
ure 2.4. The resulting descriptive statistics are shown below. For compari-
son, the corresponding SPSS statistics (with listwise deletion) are shown
in Figure 2.5. The SPSS statistics were generated using the SPSS option
Analyze → Scale → Reliability Analysis → Statistics → Descriptives
for Item/Inter-Item Covariances/Correlations. The results are identical
to the Mplus results, rounded to three decimals, which shows us that the
data set KFT.dat was apparently correctly read by Mplus. In addition,
SPSS returns the same sample size (N = 455 listwise cases). After this
check, we can start the first actual model in Mplus (see Chapter 3).
Means
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1 KFT_N3
________ ________ ________ ________ ________ ________
11.899 8.974 12.387 7.732 11.112 8.281
20 Data Analysis with Mplus
Covariances
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1
________ ________ ________ ________ ________
KFT_V1 21.677
KFT_V3 10.799 18.008
KFT_Q1 6.057 5.310 11.242
KFT_Q3 4.669 4.429 4.038 6.783
KFT_N1 11.172 11.745 6.794 6.120 29.686
KFT_N3 6.383 7.230 4.417 4.001 9.294
Covariances
KFT_N3
________
KFT_N3 11.811
Correlations
KFT_V1 KFT_V3 KFT_Q1 KFT_Q3 KFT_N1
________ ________ ________ ________ ________
KFT_V1 1.000
KFT_V3 0.547 1.000
KFT_Q1 0.388 0.373 1.000
KFT_Q3 0.385 0.401 0.462 1.000
KFT_N1 0.440 0.508 0.372 0.431 1.000
KFT_N3 0.399 0.496 0.383 0.447 0.496
Correlations
KFT_N3
________
KFT_N3 1.000
FIGURE 2.4. Modified Mplus input file for the basic analysis, now with FIML
estimation turned off. This file generates descriptive statistics based on a listwise
deletion of cases that can be more easily compared to results in SPSS.
Reading Data into Mplus 21
FIGURE 2.5. Descriptive statistics for the six KFT variables produced in SPSS
through the option Analyze → Scale → Reliability → Statistics → Descriptive
Statistics for Items/Inter-Items → Covariances/Correlations. The values match
the Mplus estimates obtained under the listwise = on; option.
using the six KFT variables as an example. The simplest way to enter
summary data is to copy and paste the covariance or correlation matrix
into a simple text file. For this purpose one can use, for example, the
simple Editor program in Windows (Start → Programs → Accessories)
or WordPad.
Figure 2.6 shows an example of a text file that contains the means
(first row), standard deviations (second row), and product moment cor-
relation matrix of the six observed KFT variables (the file is located on
the companion website and is named KFT_summary-data.txt). The
data shown in this text file can easily be used by Mplus to estimate, for
example, an SEM. The Mplus syntax to read the data from the file KFT_
summary-data.txt is shown in Figure 2.7. The data command is again
used to define the name of the data set. In addition, it has to be specified
what type of summary statistics are being read from this file. In our case
the subcommand type = means std corr; means that the data
set KFT_summary-data.txt contains the means (means) followed by the
standard deviations (std) and correlations (corr) of the variables. Note
that the order of the subcommands is important. (They have to be in line
with the order in which the summary statistics appear in the data file.) To
read a covariance matrix, one would use the subcommand cova instead
of corr.
The additional subcommand nobservations = 455; is used
to specify the sample size (number of individuals on which the summary
data set is based). In this case, the data are based on N = 455 listwise
FIGURE 2.6. Text file with summary data for the six KFT variables. The first line
contains the means of the variables. The second line contains their standard devia-
tions. The matrix contains the product–moment correlations of the variables. Figure
2.7 shows how these data can be properly processed in Mplus.
Reading Data into Mplus 23
FIGURE 2.7. Mplus input file for reading the summary data shown in Figure 2.6.
The basic option is not available for summary data. Therefore, descriptive statistics
for data checking are requested via the output: sampstat; option. Given that
we do not explicitly specify a model for the data, by default Mplus estimates a so-
called null model for all variables listed under variable: names =.
Linear SEMs with latent variables (often called covariance structure models)
are used to model complex relationships between continuous variables at
the latent level, that is, at the levels of variables that are corrected for mea-
surement error. Linear SEMs can also be viewed as multivariate regres-
sion models. In these models, structural relationships between multiple
dependent variables and multiple independent variables can be analyzed
simultaneously. Variables that serve exclusively as independent variables
are referred to as exogenous variables, because variability in these variables
is not explained by other variables in the model. For exogenous variables,
only nondirectional relationships (e.g., covariances or correlations) are
analyzed.
Variables that are regressed on one or more other variables in the
model are called endogenous variables. Endogenous variables that serve as
independent and dependent variables at the same time are called inter-
vening variables or mediator variables, because those variables mediate the
relationship between other variables. Mediated effects are frequently of
interest and often examined using manifest or latent path analysis, as
discussed in detail in Section 3.5.
Linear SEMs with latent variables can be described as consisting of
two parts (see Figure 3.1): a measurement model and a structural (or latent
variable) model. The parameters pertaining to both parts are estimated
24
Linear Structural Equation Models 25
ζ2 Structural model
β21
η1 η2
Y1 Y2 Y3 Y4
ε1 ε2 ε3 ε4
Measurement
model
FIGURE 3.1. Path diagram of a simple structural equation model with two latent
variables (h1 and h2), each of which is measured by two indicators (Y1 and Y2 are
indicators of h1, whereas Y3 and Y4 are indicators of h2). The parameters l11–l42
denote factor loadings. The variables e1–e4 are residual (error) variables. In the struc-
tural model, the latent regression of the factor h2 on h1 is estimated. In this latent
regression, b21 denotes the regression slope coefficient and z2 denotes a latent resid-
ual variable.
variables are indicated by the letter Y or with the actual variable name.
Latent variables are indicated by h or receive a substantive label that char-
acterizes their meaning. Error variables (residual variables) associated
with manifest variables are indicated by e, and latent variable residuals
are indicated by z.
Useful introductory textbooks on the theory of linear SEMs are, for
example, books by Bollen (1989), Kaplan (2009), Kline (2011), Loeh-
lin (1998), Raykov and Marcoulides (2006), as well as Schumacker and
Lomax (1996).
• Factor loading (regression slope coefficients) lij of the ith observed vari-
able Yi on the jth latent variables hj
• Intercept ai of the observed variable Yi
• Residual (error) variance Var(ei) of the observed variable Yi
• Residual covariance Cov(ei, ei′) or correlation Corr(ei, ei′)
Y = b0 + b1X + e
(9.65)
FIGURE 3.2. Path diagram of the estimated bivariate regression model for pre-
dicting the observed variable kft_q1 (quantitative ability) from kft_n1 (nonverbal/
figural ability). The numbers indicate the estimated unstandardized regression coef-
ficient (0.23) as well as the estimated variances of the variables (29.83 and 11.24) and
the estimated residual variance for kft_q1 (9.65).
FIGURE 3.3. Saving the Mplus input file from Figure 2.1 under a new name to
generate a new input file for the bivariate regression model.
30 Data Analysis with Mplus
After checking the correct data import into Mplus using the BASIC option
(see Chapter 2), one can begin with the specification of the first actual model.
In general, it is useful to start with a simple model that may represent one
part of a more complex model with many variables. This approach has the
advantage that possible errors or technical problems that may arise in a com-
plex model are easier to detect in a small submodel than in a larger model.
For example, if one wants to analyze a complex model with many variables,
it is useful to build up the model in a step-by-step fashion (Mulaik & Millsap,
2000). This means starting with only some of the manifest variables and/or
latent factors instead of including all of the variables at once. Then one can
slowly increase the complexity of the model. In this way, possible errors and
problems can be more easily traced back to a specific part of the model or
constellation of variables. As a consequence, troubleshooting will be easier
than for a complex model that already contains many variables and param-
eters. It can also be useful to first estimate the measurement model before
placing restrictions on the structural part of the model (Mulaik & Millsap,
2000).
It is important to note that only variables that are actually used in the
model should be listed under usevar, because Mplus assumes that all
variables listed under usevar are part of the model. If variables listed
under usevar are not included in the actual model specification, Mplus
will automatically assume that those variables are uncorrelated with all
other variables in the model (see discussion below as well as Appendix B),
resulting in the following error message:
FIGURE 3.4. Mplus input file for the bivariate linear regression model from Figure
3.2.
32 Data Analysis with Mplus
be obtained. Two new commands have been added to this input file. The
first one is the model command, which is used to specify the desired
model. The second one is the output command that is used to request
additional output that is not part of the Mplus default output:
This message means that Mplus has not found any kind of fatal syn-
tax errors that prevented the program from running the input file. For
example, this message indicates that Mplus had no difficulties finding
and reading the data set for the analysis. It is important to note that this
message is not a guarantee that Mplus read the data correctly or that the
model was correctly specified. It is still important to check the accuracy
of the calculated sample size and the descriptive statistics, as well as the
plausibility of the estimated model parameter estimates, even when the
user has already carried out a BASIC analysis (see Section 2.1.1). In addi-
tion, this message is not a guarantee that the actual estimated statistical
model converged or that all parameter estimates are proper and mean-
ingful. Problems related to the actual estimation of the specific statistical
model would appear further below in the Mplus output (below the sample
statistics).
Linear Structural Equation Models 33
SUMMARY OF ANALYSIS
Number of groups 1
Number of observations 456
Continuous
KFT_Q1
Estimator ML
Information matrix OBSERVED
Maximum number of iterations 1000
Convergence criterion 0.500D-04
Maximum number of steepest descent iterations 20
Mplus. In addition, these statistics are often useful for the interpretation
of the model results. In our case, we can see that the correlation between
kft_q1 and kft_n1 is moderately large (r = .375) as shown in the sub-
section Correlations.
SAMPLE STATISTICS
SAMPLE STATISTICS
Means
KFT_Q1 KFT_N1
________ ________
1 12.377 11.088
Covariances
KFT_Q1 KFT_N1
________ ________
KFT_Q1 11.235
KFT_N1 6.870 29.826
Correlations
KFT_Q1 KFT_N1
________ ________
KFT_Q1 1.000
KFT_N1 0.375 1.000
Following the descriptive statistics, the user should check once again
for possible error messages. As mentioned above, Mplus would output
additional error messages at this point if there were any kinds of specific
problems related to the actual estimation of the model. Possible reasons
for such types of error messages include (1) the model is not identified, (2)
inadmissible estimation results are encountered (e.g., negative variance or
residual variance estimates or correlations > | 1 |), or (3) when the estima-
tion procedure failed to converge on a final optimal set of parameter esti-
mates. In our example, no error messages appear. The model estimation
terminated without problems, as is expected for a simple linear regression
model with manifest variables. Therefore, we just obtain the message
should be interpreted only with great caution or not at all, because severe
estimation problems may have occurred.
Subsequently, a number of statistical measures are reported that can
be used to evaluate the overall goodness of fit of the model under MODEL
FIT INFORMATION (formerly referred to as TESTS OF MODEL FIT
in older versions of Mplus). A description of these indices can be found
in Box 3.7 on page 45 (for more detailed discussions, see Bollen & Long,
1993, or Schermelleh-Engel, Moosbrugger, & Müller, 2003). In our case
we are analyzing just a simple linear regression model with only mani-
fest variables. The simple regression model uses the entire information
in the data to estimate model parameters. Such a model is saturated so
that most of the global indices of model fit are not of great interest to us
here. Saturated models always show a perfect fit with the observed data.
Only the chi-square test for the baseline model (the so-called independence
model) is relevant. The independence model assumes that all variables
used in the model are uncorrelated. In our case, this means that kft_q1
and kft_n1 are assumed to be uncorrelated. The chi-square test for the
independence model is significant in our example, c2 = 69.235, df = 1, p
< .0001, which indicates that the independence model has to be rejected.
This shows us that the correlation between kft_q1 and kft_n1 is
significantly different from zero and that the estimation of a regression
model is this meaningful.
Loglikelihood
H0 Value -1163.957
H1 Value -1163.957
Information Criteria
Value 0.000
Degrees of Freedom 0
P-Value 0.0000
36 Data Analysis with Mplus
Estimate 0.000
90 Percent C.I. 0.000 0.000
Probability RMSEA <= .05 0.000
CFI/TLI
CFI 1.000
TLI 1.000
Value 69.235
Degrees of Freedom 1
P-Value 0.0000
Value 0.000
Following are the parameter estimates for the model, first of all in
unstandardized form (under MODEL RESULTS).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_Q1 ON
KFT_N1 0.230 0.027 8.647 0.000
Intercepts
KFT_Q1 9.823 0.329 29.833 0.000
Residual Variances
KFT_Q1 9.652 0.639 15.100 0.000
test. Here the p-value is smaller than .0001, which tells us that the slope
coefficient is significantly different from zero at an a-level of .0001. This
means that kft_n1 significantly predicts kft_q1 (or equivalently, that
the two variables are significantly correlated).
The next row gives the estimated intercept parameter, which is bˆ 0 =
9.823. The intercept is also statistically significant (z = 29.833). Using the
estimates for bˆ 1 and bˆ 0 we can formulate the regression equation for our
example:
This means that for individuals with a kft_n1 score of zero (none
of the items solved correctly), a score of 9.823 on kft_q1 is predicted.
Furthermore, according to this equation, the expected increase in the
kft_q1 score for a one-unit increase in the kft_n1 score is 0.23 points.
The third row of the MODEL RESULTS table (p. 36) provides us with
the unstandardized residual variance (estimated variance of the error
variable e). Here the error of variance equals σˆ 2e = 9.652. The value of the
error variance becomes meaningful when we relate it to the total variance
of the dependent variable (σˆ 2Y ). The total variance of the dependent vari-
able can be found under SAMPLE STATISTICS in the upper part of the
output (in the main diagonal of the covariance matrix). The estimated
variance of kft_q1 is 11.235 in our example. This shows us that the
residual variance (i.e., that part of the variance that is not explained in the
regression analysis) is relatively large compared to the total variance of the
dependent variable. In other words, only a small portion of individual dif-
ferences in kft_q1 scores can be explained by individual differences in
the kft_n1 scores. By calculating the so-called coefficient of determination
(R 2), we can see that only about 14.1% of the variability in kft_q1 scores
can be explained by this simple regression model:
ˆ2
ˆR 2 = 1 − σ e = 1 − 9.652 = .141
σˆ Y2 11.235
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_Q1 ON
KFT_N1 0.375 0.040 9.329 0.000
Intercepts
KFT_Q1 2.931 0.166 17.678 0.000
Residual Variances
KFT_Q1 0.859 0.030 28.448 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
In the data set KFT.dat there are two indicators for each ability facet (i.e., ver-
bal, quantitative, and nonverbal). Therefore, we can model both quantitative
and nonverbal ability through a latent factor. The manifest variables kft_q1
and kft_q3 are used as indicators of a quantitative ability factor (KFT_Q).
The variables kft_n1 and kft_n3 are used as indicators of a latent vari-
able representing nonverbal abilities (KFT_N). The regression of quantitative
abilities on nonverbal abilities is modeled at the level of latent variables (i.e.,
in the structural model). The resulting model can be referred to as a latent
regression analysis. It is shown as a path diagram in Figure 3.5.
Linear Structural Equation Models 41
(1.24)
KFT_N 0.48
KFT_Q
(14.22)
FIGURE 3.5. Latent regression model for predicting quantitative ability (fac-
tor KFT_Q) versus nonverbal ability (factor KFT_N). Each factor is measured by
two indicators (manifest variables = KFT subscale scores). The values indicate the
unstandardized Mplus parameter estimates (the loadings, path coefficients, and vari-
ances/residual variances). The superscript f indicates factor loadings that were fixed
to 1 to identify the metric of the latent variables.
The Mplus syntax command for specifying the model shown in Fig-
ure 3.5 is depicted in Figure 3.6. First of all, note that all manifest vari-
ables that are used in the model (and only those) have to be listed under
usevariables. The measurement models are specified in the model
command using by statements. These by statements serve to indicate
which manifest variables load onto which latent factor. The word by here
stands for “measured by.” In our example, the latent factor KFT_Q is mea-
sured by the manifest variables kft_q1 and kft_q3, whereas the factor
KFT_N is measured by the variables kft_n1 and kft_n3.
The factor loading of the first indicator listed in each by statement
will be fixed to 1 by default in Mplus (here the loadings of kft_q1 and
kft_n1). Fixing one loading per factor is necessary to assign a metric to
each factor (see also Box 3.6). The loadings of all remaining indicators are
freely estimated as the default.
The command KFT_Q on KFT_N; is used in the same way as the
on command is used in the manifest regression analysis in Section 3.2.
Here this command refers to the regression of the latent factor KFT_Q on
the factor KFT_N. It therefore defines the structural part of this simple
structural equation model. Residual variances for the manifest variables
42 Data Analysis with Mplus
This causes Mplus to set the first loading free. The variance of factor KFT_N
can be fixed to 1 by adding an additional command that lists the name of the
factor and adds “@1” after the name of the factor:
KFT_N@1;
Neither the model fit nor the standardized parameter estimates are affected
by fixing the factor variance rather than the loading in this example (see
output on the companion website).
Linear Structural Equation Models 43
FIGURE 3.6. Mplus input file for the latent regression model from Figure 3.5.
and the latent dependent variable KFT_Q are estimated by Mplus as the
default. All other commands shown in this input have already been dis-
cussed. By clicking on run, we obtain the model results (maximum likeli-
hood estimates are the default), which are shown, in part, below:
Loglikelihood
H0 Value -4696.080
H1 Value -4696.079
Information Criteria
Value 0.002
Degrees of Freedom 1
P-Value 0.9653
44 Data Analysis with Mplus
Estimate 0.000
90 Percent C.I. 0.000 0.000
Probability RMSEA <= .05 0.980
CFI/TLI
CFI 1.000
TLI 1.015
Value 406.899
Degrees of Freedom 6
P-Value 0.0000
Value 0.000
The model fit information shows that the model fits the data very
well. The chi-square value for the target model shows that the model-
implied covariance matrix and mean vector do not differ significantly
from the corresponding observed values, c2 = 0.002, df = 1, p = .9635.
The chi-square value for the baseline model is very large and highly sta-
tistically significant, c2 = 406.899, df = 6, p < .0001, which shows us that
the assumption that all four observed KFT variables are unrelated is not
tenable. The remaining fit statistics also indicate a good model fit, given
that they all show values better than the recommended cutoff scores (see
Box 3.7 for details). According to these cutoff scores, CFI and TLI should
be larger than 0.95 or 0.97, and RMSEA and SRMR should both be smaller
than 0.05 (Schermelleh-Engel et al., 2003). In the present example, CFI
and TLI are both larger than 0.97, and RMSEA and SRMR are both smaller
than 0.05. The test of close fit also leads to a positive outcome for the
model. This test can be used to test the hypothesis that RMSEA is smaller
than or equal to .05 in the population. This hypothesis is not rejected at
the .05 level in the present case given that p = .980.
In the following (beginning on page 47), the unstandardized Mplus
parameter estimates for the latent regression model from Figure 3.5 are
shown. The unstandardized parameter estimates appear under the header
MODEL RESULTS. The most important estimates are also shown in Fig-
ure 3.5.
Linear Structural Equation Models 45
This box provides an overview of the most important tests and indices avail-
able in Mplus to assess model fit of SEM as well as common cutoff values for
acceptable fit (Hu & Bentler, 1999; Schermelleh-Engel et al., 2003).
Chi-Square Test. This tests the null hypothesis that the covariance
matrix and mean vector in the population are equal to the model-implied
covariance matrix and mean vector (test of exact model fit). A significant chi-
square value leads to the rejection of the null hypothesis that the model fits
exactly in the population. The degrees of freedom of this test are calculated
as the difference between the number of pieces of available information (vari-
ances, covariances, and means of the manifest variables) minus the number
of estimated model parameters. In Mplus, the chi-square value, the corre-
sponding degrees of freedom, and the p-value are reported as the default.
Comparative Fit Index (CFI). The CFI belongs to the family of so-
called incremental fit indices that compare the fit of the target model to the fit
of a baseline model. In Mplus, the baseline model currently is the so-called
independence model, which assumes that the population covariance matrix of
the observed variables is a diagonal matrix. This means that the observed
46 Data Analysis with Mplus
variables are allowed to have different variances, but are assumed to have
zero covariances. In other words, it is assumed that there are no relation-
ships between any of the variables. The CFI indicates by how much the target
model fits better than the more parsimonious independence model. For a
good model, the CFI should be larger than 0.95 (better: larger than 0.97).
The CFI is reported as part of the Mplus standard output.
output: residual;
output: modindices;
Mplus also outputs the expected parameter change, that is, the actual value
that is expected for the additional parameter to be estimated in a less restric-
tive version of the model.
Similar to residuals, modification indices often provide useful hints as
to the cause of an insufficient global model fit. However, they can also some-
times lead to confusing results and often are not directly related to the true
cause of a bad fit. Therefore, model modification indices should never be
used in a purely data-driven way. In addition, users should always imple-
ment just one (theoretically meaningful) modification at a time. Model modi-
fications should be theoretically justified rather than purely data-driven, and
they should be validated with independent data.
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_Q BY
KFT_Q1 1.000 0.000 999.000 999.000
KFT_Q3 0.903 0.094 9.588 0.000
KFT_N BY
KFT_N1 1.000 0.000 999.000 999.000
KFT_N3 0.652 0.064 10.222 0.000
48 Data Analysis with Mplus
KFT_Q ON
KFT_N 0.476 0.056 8.467 0.000
Intercepts
KFT_Q1 12.387 0.157 78.890 0.000
KFT_Q3 7.732 0.122 63.398 0.000
KFT_N1 11.112 0.255 43.552 0.000
KFT_N3 8.281 0.161 51.457 0.000
Variances
KFT_N 14.215 2.074 6.854 0.000
Residual Variances
KFT_Q1 6.758 0.608 11.107 0.000
KFT_Q3 3.128 0.395 7.921 0.000
KFT_N1 15.406 1.591 9.683 0.000
KFT_N3 5.734 0.644 8.911 0.000
KFT_Q 1.238 0.425 2.916 0.004
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_Q BY
KFT_Q1 0.631 0.041 15.548 0.000
KFT_Q3 0.733 0.040 18.247 0.000
KFT_N BY
KFT_N1 0.693 0.038 18.019 0.000
KFT_N3 0.717 0.038 18.717 0.000
KFT_Q ON
KFT_N 0.850 0.049 17.245 0.000
Intercepts
KFT_Q1 3.698 0.131 28.177 0.000
KFT_Q3 2.972 0.109 27.240 0.000
KFT_N1 2.042 0.082 24.798 0.000
KFT_N3 2.412 0.093 26.024 0.000
Variances
KFT_N 1.000 0.000 999.000 999.000
Residual Variances
KFT_Q1 0.602 0.051 11.781 0.000
KFT_Q3 0.462 0.059 7.840 0.000
KFT_N1 0.520 0.053 9.765 0.000
KFT_N3 0.487 0.055 8.869 0.000
KFT_Q 0.278 0.084 3.315 0.001
50 Data Analysis with Mplus
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
KFT_Q1 0.398 0.051 7.774 0.000
KFT_Q3 0.538 0.059 9.124 0.000
KFT_N1 0.480 0.053 9.010 0.000
KFT_N3 0.513 0.055 9.359 0.000
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
that indicator-specific variance was separated out and became part of the
error variance. This likely contributed to the fairly large difference in the
amount of explained variance between the manifest and the latent model.
As an example, consider the six KFT variables in the data set KFT.dat. The
following research question may be of interest: Do the six KFT scales kft_v1,
kft_v2, kft_q1, kft_q3, kft_n1, and kft_n3 measure three distinct dimen-
sions (verbal, quantitative, and nonverbal/figural abilities) or are they merely
indicators of the single latent ability dimension, for example, a general factor
of intelligence? We examine this question by means of two different first-
order CFA models in Mplus. Subsequently, we also discuss the specification
of a second-order CFA model.
52 Data Analysis with Mplus
KFT-g
(9.41)
FIGURE 3.7. Single-factor (g-factor) model for the six KFT variables with unstan-
dardized Mplus parameter estimates. The superscript f indicates that the factor load-
ing of the first indicator (KFT_V1) was fixed to 1 to identify the metric of the latent
variable.
The second model (see Figure 3.8) contains three latent factors sug-
gesting that the six KFT scales measure three distinct dimensions of men-
tal ability (kft_v1 and kft_v3 as indicators of a verbal ability factor, kft_q1
and kft_q3 as indicators of a quantitative/numerical ability factor, and
kft_n1 and kft_n3 as indicators of a nonverbal/figural ability factor). A
statistical comparison of the two models allows us to find out whether a
one- or a three-factor model is more appropriate to represent what is mea-
sured by the KFT scales.
The Mplus input files for the one- and three-factor models are shown
in Figure 3.9 and Figure 3.10, respectively. Note that as an exception, the
usevariables command is not needed in this example, because all
variables in the data set (all variables that are listed under names =) are
actually used in the model statement. In all other cases, the usevari-
ables command needs to be included (compare Section 3.2).
Further note that the covariances between the three latent variables
in the three-factor model do not have to be explicitly specified in the
Mplus input file. By default Mplus estimates all covariances between exog-
enous factors automatically (because none of the three factors is predicted
by another variable in the model, they are all considered exogenous vari-
ables).
Sometimes a user may be interested in explicitly setting the covari-
ance/correlation between two factors to zero (e.g., see Section 4.1.3). This
Linear Structural Equation Models 53
5.25 (r = .75)
FIGURE 3.8. Three-factor model for the six KFT variables with unstandardized
Mplus parameter estimates (latent correlations are given in parentheses). The super-
script f indicates factor loadings that were fixed to 1 to identify the metric of the
latent variables.
FIGURE 3.9. Mplus input file for the single-factor model from Figure 3.7.
54 Data Analysis with Mplus
FIGURE 3.10. Mplus input file for the three-factor model from Figure 3.8.
CFI, TLI, and SRMR values, both the one- and the three-factor model
would be acceptable. The AIC value for the one-factor model is somewhat
higher than the AIC value for the three-factor model, so that according to
this criterion, the three-factor model would again be preferred. In sum-
mary, the fit statistics seem to point more to the three-factor solution than
to the one-factor solution.
The Mplus parameter estimates for the three-factor model are shown
in the following. We again begin with the MODEL RESULTS table, which
provides the unstandardized solution. It turns out that all loadings (listed
under BY) that were freely estimated are significantly different from zero
(the first loadings of the first indicator is again set to 1 for each factor
to identify the metric of each factor). All three estimated factor covari-
ances (shown under WITH) are also statistically significant. The size of
the relationship between the three factors is difficult to evaluate based on
the covariances, however, because covariances are unstandardized mea-
sures of association. Standardized covariances (correlations r) are easier to
interpret because they are restricted to a standardized range (–1 ≤ r ≤ 1).
The correlations of the latent factors are provided in the standardized
solution and are discussed below.
The unstandardized solution furthermore contains the unstandard-
ized intercepts of the manifest variables. In this case, the intercepts are
identical to the observed means of the manifest variables because there
are no restrictions on the mean structure in this model. Also given are
the estimated factor variances and the residual variances for the manifest
variables.
56 Data Analysis with Mplus
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_V BY
KFT_V1 1.000 0.000 999.000 999.000
KFT_V3 1.034 0.083 12.434 0.000
KFT_Q BY
KFT_Q1 1.000 0.000 999.000 999.000
KFT_Q3 0.865 0.084 10.337 0.000
KFT_N BY
KFT_N1 1.000 0.000 999.000 999.000
KFT_N3 0.620 0.050 12.394 0.000
KFT_Q WITH
KFT_V 5.247 0.697 7.533 0.000
KFT_N WITH
KFT_V 11.088 1.222 9.074 0.000
KFT_Q 7.123 0.839 8.488 0.000
Intercepts
KFT_V1 11.899 0.218 54.574 0.000
KFT_V3 8.974 0.199 45.155 0.000
KFT_Q1 12.387 0.157 78.888 0.000
KFT_Q3 7.732 0.122 63.396 0.000
KFT_N1 11.112 0.255 43.552 0.000
KFT_N3 8.281 0.161 51.457 0.000
Variances
KFT_V 10.424 1.401 7.442 0.000
KFT_Q 4.659 0.720 6.471 0.000
KFT_N 14.953 1.968 7.596 0.000
Residual Variances
KFT_V1 11.205 1.004 11.156 0.000
KFT_V3 6.830 0.853 8.011 0.000
KFT_Q1 6.559 0.586 11.192 0.000
KFT_Q3 3.283 0.366 8.980 0.000
KFT_N1 14.667 1.382 10.615 0.000
KFT_N3 6.033 0.550 10.967 0.000
correlations under WITH. It can be seen that all three KFT factors are
strongly positively correlated (.75 ≤ r ≤ .89), which shows that there is
rather little discriminant validity among the three KFT subscales.
Despite the rather large correlations among the three factors, the
fit statistics indicated that more than one factor is needed to sufficiently
account for the observed variances and covariances of the six KFT sub-
scales and that a g-factor solution (which implies perfectly correlated
factors) had to be rejected for the present data. Therefore, we can con-
clude that even though the three subscales appear to share a substantial
amount of common variance, there also is a nontrivial portion of system-
atic domain-specific (verbal, numerical, and figural) ability variance that
needs to be taken into account and that explains why the correlations
among the three factors differ from 1.
Two-Tailed
Estimate S.E. Est./S.E. P-Value
KFT_V BY
KFT_V1 0.694 0.033 20.963 0.000
KFT_V3 0.787 0.031 25.290 0.000
KFT_Q BY
KFT_Q1 0.644 0.039 16.698 0.000
KFT_Q3 0.718 0.038 18.959 0.000
KFT_N BY
KFT_N1 0.711 0.033 21.591 0.000
KFT_N3 0.699 0.033 21.046 0.000
KFT_Q WITH
KFT_V 0.753 0.051 14.764 0.000
KFT_N WITH
KFT_V 0.888 0.041 21.589 0.000
KFT_Q 0.853 0.049 17.458 0.000
Intercepts
KFT_V1 2.558 0.097 26.401 0.000
KFT_V3 2.117 0.084 25.083 0.000
KFT_Q1 3.698 0.131 28.176 0.000
KFT_Q3 2.972 0.109 27.239 0.000
KFT_N1 2.042 0.082 24.798 0.000
KFT_N3 2.412 0.093 26.024 0.000
58 Data Analysis with Mplus
Variances
KFT_V 1.000 0.000 999.000 999.000
KFT_Q 1.000 0.000 999.000 999.000
KFT_N 1.000 0.000 999.000 999.000
Residual Variances
KFT_V1 0.518 0.046 11.267 0.000
KFT_V3 0.380 0.049 7.753 0.000
KFT_Q1 0.585 0.050 11.753 0.000
KFT_Q3 0.485 0.054 8.933 0.000
KFT_N1 0.495 0.047 10.589 0.000
KFT_N3 0.512 0.046 11.037 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
KFT_G
(8.17)
1.00f 0.64
1.36
FIGURE 3.11. Second-order factor model for the six KFT variables with unstan-
dardized Mplus parameter estimates. In this model, the three first-order KFT factors
KFT_V, KFT_N, and KFT_Q load onto a second-order factor (KFT_G). The super-
script f indicates factor loadings that were fixed to 1 to identify the metric of the
latent variables. Variances and residual variances are given in parentheses. Note the
estimated negative residual variance (“Heywood case”) for the factor KFT_N (–0.10).
FIGURE 3.12. Mplus input file for the second-order factor model from Figure
3.11.
MODEL RESULTS
Residual Variances
KFT_V1 11.206 1.004 11.156 0.000
KFT_V3 6.829 0.853 8.010 0.000
KFT_Q1 6.559 0.586 11.192 0.000
KFT_Q3 3.283 0.366 8.979 0.000
KFT_N1 14.667 1.382 10.614 0.000
KFT_N3 6.033 0.550 10.967 0.000
KFT_V 2.256 0.754 2.992 0.003
KFT_Q 1.289 0.402 3.209 0.001
KFT_N -0.099 1.206 -0.082 0.935
β3
ε1 ε2
X β1 Y1 β2 Y2
Physical Functional Subjective
Health Health Health
(SICK) (FH) (SHP)
FIGURE 3.13. Manifest path model for the relationship between physical health
(X), functional health (Y1), and subjective health (Y2) proposed by Whitelaw and
Liang (1991). The constants b1, b2, and b3 indicate regression (path) coefficients.
The variables e1 and e2 are residual variables. The model contains one indirect effect,
which can be expressed as the product b1 · b2. The variable names used in the Mplus
analyses are given in parentheses.
Linear Structural Equation Models 63
*I would like to thank Susanne Wurm from the DZA for providing the data for this
example as well as Christopher Marx for running some of the analyses.
One reason why path analysis is very popular in the social sciences is
that many social science theories include hypotheses with regard to direct
and indirect effects of variables on each other (e.g., see MacKinnon, 2008).
Indirect effects are effects that are mediated through other variables and are
therefore also referred to as mediated effects (Baron & Kenny, 1986). Medi-
ated effects are present in a path model when the model contains one or
more variables that are dependent and independent variables at the same
time. Those types of variables are often referred to as intermittent or media-
tor variables. Mediated effects can be analyzed by means of either manifest
or latent path analyses.
In the model in Figure 3.13, the variable Y1 (functional health, FH)
is a mediator variable, because this variable is a dependent variable with
regard to the variable X (physical health, SICK) but an independent
variable with regard to the variable Y2 (subjective health, SHP). In other
64 Data Analysis with Mplus
words, Y1 receives an effect from X (path b1) while at the same time send-
ing an effect out to Y2 (path b2). Hence, Y1 partly mediates the effect of X
on Y2, which explains the term mediator variable. Formally, the indirect
(or mediated) effect of X on Y2 can be quantified as the product of b1 times
b2 (MacKinnon, 2008). Variable X, in this example, also has a direct effect
on Y2, which is expressed through the path coefficient b3.
In substantive terms, all this means that, in this model, it is assumed
that the objective health status (physical health) of a person has both
a direct (nonmediated) and an indirect (mediated) effect (via functional
health) on subjective health. The mediated effect is caused by the fact that
objective physical problems often result in a loss of functional competen-
cies that, in turn, cause a decrease in subjective health. By fitting the path
model to our data, we can test whether physical health indeed has both a
direct and an indirect effect on subjective health, or whether the effect of
physical health on subjective health is fully mediated through functional
health.
In path analyses that contain one or more indirect effects, an effect
decomposition is often of interest. In this decomposition, the total effect is
split up into the sum of all indirect effects plus the direct effect (if a direct
effect is assumed). In other words, the total effect is equal to the sum of
the direct effect plus all indirect effects. In our example, the direct effect
of X on Y2 is measured by the coefficient b3, whereas the indirect effect is
measured by the product b1 · b2. Therefore, the total effect can be calcu-
lated as the sum b3 + (b1 · b2).
Mediator effects should not be confused with moderator effects (see also
Baron & Kenny, 1986). In contrast to a mediator effect, a moderator effect
is present when there is an interaction between two variables. A variable is
referred to as a moderator variable when the strength of the direct influence
of one variable (say, X) on another variable (say, Y) depends on the values of
a third variable (say, Z). If, for example, the regression coefficient b3 (which
characterizes the direct effect of X on Y2 in Figure 3.13) varied across dif-
ferent levels of Y1 (i.e., if the direct influence of X on Y2 was moderated by
Y1), then Y1 could be seen as a moderator variable. The analysis of moderator
(interaction) effects is not discussed in this book. Aiken and West (1991;
as well as Cohen et al., 2003, their Chapters 7–9) discuss the analysis of
moderated effects in detail in the context of manifest regression models (so-
called moderated regression). Latent moderator models (SEMs with interaction
effects) are discussed, for example, in Marsh, Wen, and Hau (2006).
Linear Structural Equation Models 65
FH on SICK;
SHP on SICK FH;
66 Data Analysis with Mplus
FIGURE 3.14. Mplus input file for the manifest path model from Figure 3.13.
The keyword ind between the two variable names causes Mplus to out-
put all indirect effects that the variable SICK has on the variable SHP (in
this case, there is only one indirect effect via FH), along with the relevant
standard errors and tests of significance. The model indirect sub-
command is particularly useful when a researcher wants to test specific
indirect effects for significance (see below).
In the Mplus output, we obtain sample statistics with the means,
covariances, and correlations of the three manifest variables (only the
correlation table is shown here). As expected, subjective health corre-
lates positively with functional health and negatively with the number
of health problems in the last 12 months. In addition, the frequency of
health problems is negatively correlated with functional health.
Linear Structural Equation Models 67
Correlations
SHP FH SICK
________ ________ ________
SHP 1.000
FH 0.594 1.000
SICK -0.458 -0.447 1.000
Under the rubric MODEL FIT INFORMATION we can see that the
estimated manifest path model in our example has zero degrees of free-
dom. The reason is that all means, variances, and covariances of the three
manifest variables are used up in the model to estimate model parameters.
One therefore speaks of a so-called saturated or just identified model in this
case (this was the same issue for the manifest regression model discussed
in Section 3.2). One consequence is that the model does not contain any
testable restrictions with respect to the mean and covariance structure of
the observed variables and therefore fits perfectly (for details see Bollen,
1989). This can be seen from the chi-square model test:
Value 0.000
Degrees of Freedom 0
P-Value 0.0000
The chi-square value is zero with zero degrees of freedom. The test
is therefore not useful to evaluate the model fit in this case (the p-value is
erroneously reported to be 0 by Mplus; it should in fact be printed as p =
1 in this case, given that the model has a perfect fit to the data). In gen-
eral, the judgment of the model fit of a saturated model focuses more on
the estimated model parameters (in particular, the estimated path coef-
ficients) and the obtained proportion of the explained variability in the
endogenous variables (as measured by the R 2 value) rather than the global
fit statistics output by Mplus.
The Mplus MODEL RESULTS table provides us with the estimates
of the unstandardized path coefficients (here for the saturated path model
that includes the direct effect from SICK to SHP). Fully standardized
path coefficients can again be found under STANDARDIZED MODEL
RESULTS (STDYX Standardization). It can be seen that there is
a significant negative direct effect of the frequency of health issues on sub-
jective health (direct effect from SICK to SHP, bˆ 3 = –0.094, z = –11.258,
p < .001, bˆ standardized
3 = –0.24). For every additional health issue encoun-
tered in the last 12 months, the expected decrease in subjective health is
68 Data Analysis with Mplus
Note that not every manifest path model is necessarily saturated. If, for
example, in the present model one deletes the direct path from SICK to SHP
(which is equivalent to fixing the corresponding path coefficient b3 to zero),
the model would no longer be saturated. Instead, it would have one degree
of freedom, implying one testable restriction. Using tests or indices of model
fit, the assumption that the effect of physical health on subjective health is
fully mediated by functional health could then be tested. In other words,
this restricted model would be in line with the idea that physical health only
indirectly affects subjective health.
The Mplus model specification would have to be changed, as follows,
to obtain a nonsaturated model as described above (the full input for this
alternative model can be found on the companion website):
FH on SICK;
SHP on FH;
The resulting more restricted model, without the direct effect from SICK to
SHP, shows a highly significant chi-square value of 122.11 with one degree
of freedom (p < .0001). The remaining fit indices (especially the RMSEA
index) also point to a rather bad fit of this model (RMSEA = .27, CFI = .90,
SRMR = .06). Therefore, the hypothesis of complete mediation has to be
rejected (there is also a significant direct effect from physical health to sub-
jective health).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
FH ON
SICK -0.096 0.005 -20.273 0.000
SHP ON
SICK -0.094 0.008 -11.258 0.000
FH 0.885 0.039 22.765 0.000
Intercepts
SHP 1.408 0.117 12.034 0.000
FH 2.936 0.016 179.357 0.000
Residual Variances
SHP 0.432 0.015 28.705 0.000
FH 0.174 0.006 28.705 0.000
STDYX Standardization
Two-Tailed
Estimate S.E. Est./S.E. P-Value
FH ON
SICK -0.447 0.020 -22.661 0.000
SHP ON
SICK -0.240 0.021 -11.441 0.000
FH 0.486 0.019 25.057 0.000
Intercepts
SHP 1.661 0.151 10.973 0.000
FH 6.304 0.099 63.378 0.000
Residual Variances
SHP 0.601 0.019 32.153 0.000
FH 0.800 0.018 45.431 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Specific indirect
SHP
FH
SICK -0.085 0.006 -15.140 0.000
Direct
SHP
SICK -0.094 0.008 -11.258 0.000
The estimated total effect is the sum of the estimated direct effect bˆ 3
and the estimated indirect effect bˆ 1 · bˆ 2 :
In this example, we have to deal with only one indirect effect from
physical health to subjective health (mediated by functional health), such
that the total indirect effect is equal to the specific indirect effect (given
under Specific indirect). The default test used by Mplus to test
the indirect effect for significance is the so-called Sobel test (Sobel, 1982).
Linear Structural Equation Models 71
The standard error for the indirect effect provided by Mplus is 0.006
(z = –15.14). Therefore, the indirect effect is significant (p < .001) accord-
ing to this test.
Mplus also outputs the estimated direct effect bˆ 3 again, which equals
–0.094. The standardized solution lists the same effects again now in fully
standardized form:
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Specific indirect
SHP
FH
SICK -0.217 0.013 -16.600 0.000
Direct
SHP
SICK -0.240 0.021 -11.441 0.000
output: cinterval;
Specific indirect
SHP
FH
SICK -0.102 -0.098 -0.085 -0.072 -0.068
Direct
SHP
SICK -0.117 -0.111 -0.094 -0.076 -0.071
indirect effect. The relevant output lines are shown in boldface, respec-
tively:
Specific indirect
SHP
FH
SICK -0.257 -0.247 -0.217 -0.187 -0.178
Direct
SHP
SICK -0.299 -0.285 -0.240 -0.196 -0.182
β 31
ζ2 ζ3
η1 β 21 η2 β 32 η3
Phys. Health Funct. Health Subj. Health
(PH) (FH) (SH)
λ11 λ21 λ12 λ22 λ13 λ23
FIGURE 3.15. Latent path model with three latent variables h1, h2, and h3, each
of which is measured with two manifest variables Yij. The variable names used in
the Mplus analyses appear in parentheses. The constants lij indicate factor loadings.
The variables eij are measurement error variables. The constants bjj′ indicate latent
regression (path) coefficients. The zj variables are latent residual variables. At the
level of the latent variables, the model contains the indirect effect b21 · b32.
FIGURE 3.16. Mplus input file for the latent path model from Figure 3.15.
In the latent path model, we first specify the measurement model for
the latent factors as in a CFA:
SH by SHP SHC;
PH by SICK CONSULT;
FH by FH1 FH2;
FH on PH;
SH on PH FH;
Note that SH, PH, and FH are now latent variables that are not part of
our data set and that therefore do not have to be listed under variable:
names. The subcommand for obtaining a summary of all direct, indirect,
76 Data Analysis with Mplus
and total effect remains the same as in the manifest path model (but now
refers to the latent variables):
Value 5.509
Degrees of Freedom 6
P-Value 0.4804
CFI/TLI
CFI 1.000
TLI 1.000
Estimate 0.000
90 Percent C.I. 0.000 0.031
Probability RMSEA <= .05 1.000
Value 0.007
(lˆ 12
standardized
= 0.972; lˆ 22
standardized
= 0.955). This finding shows that
the indicators of this factor are very homogeneous (the two test halves
seem to measure very similar aspects of functional health) and are highly
reliable.
The factor for physical health (PH) has the smallest standardized
loadings (lˆ 11standardized
= 0.623, lˆ 21
standardized
= 0.494). This finding can be
explained by the fact that even though the frequency of health problems
is positively correlated with the number of physician consultations, this
correlation is rather moderate in size in the present data set (r = .308,
p < .001; see full output on the companion website). Therefore, the two
indicators of physical health can be seen as rather heterogeneous (they
measure different facets of the construct physical health). The variable
SICK shows a somewhat larger standardized loading on the factor than
the number of physician consultations and might therefore be seen as the
marker variable for this factor. (Marker variables are indicators that show
high loadings on the intended factor and are therefore seen as most rel-
evant for the interpretation of this factor.)
With regard to the structural model, the relationships between the
three constructs go in the same direction: physical health problems neg-
atively impact both functional health ( bˆ 21 = –0.246, z = –13.189, p <
.001, bˆ 21
standardized
= –0.722) and subjective health ( bˆ 31 = –0.379, z =
–6.944, p < .001, bˆ standardized
31 = –0.709), whereas functional health is
positively related to subjective health ( bˆ 32 = 0.314, z = 2.631, p = .009,
bˆ standardized
32 = –0.200).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
SH BY
SHP 1.000 0.000 999.000 999.000
SHC 0.710 0.031 23.145 0.000
PH BY
SICK 1.000 0.000 999.000 999.000
CONSULT 5.547 0.363 15.265 0.000
FH BY
FH1 1.000 0.000 999.000 999.000
FH2 0.981 0.014 68.107 0.000
FH ON
PH -0.246 0.019 -13.190 0.000
78 Data Analysis with Mplus
SH ON
PH -0.379 0.055 -6.945 0.000
FH 0.314 0.119 2.632 0.008
Intercepts
SHP 3.525 0.021 168.751 0.000
SHC 2.521 0.020 128.763 0.000
SICK 2.706 0.054 50.435 0.000
CONSULT 13.082 0.375 34.892 0.000
FH1 2.646 0.012 226.225 0.000
FH2 2.709 0.012 232.059 0.000
Variances
PH 1.839 0.179 10.277 0.000
Residual Variances
SHP 0.194 0.019 10.359 0.000
SHC 0.367 0.016 23.627 0.000
SICK 2.906 0.159 18.304 0.000
CONSULT 175.061 7.166 24.429 0.000
FH1 0.012 0.002 5.135 0.000
FH2 0.020 0.002 8.231 0.000
SH 0.132 0.026 5.115 0.000
FH 0.102 0.009 11.015 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
SH BY
SHP 0.854 0.016 54.333 0.000
SHC 0.647 0.018 36.103 0.000
PH BY
SICK 0.623 0.026 24.309 0.000
CONSULT 0.494 0.025 19.907 0.000
FH BY
FH1 0.972 0.006 173.398 0.000
FH2 0.955 0.006 165.700 0.000
FH ON
PH -0.722 0.030 -24.379 0.000
SH ON
PH -0.709 0.080 -8.836 0.000
FH 0.200 0.076 2.631 0.009
Intercepts
SHP 4.157 0.076 54.352 0.000
Linear Structural Equation Models 79
Variances
PH 1.000 0.000 999.000 999.000
Residual Variances
SHP 0.270 0.027 10.062 0.000
SHC 0.581 0.023 25.043 0.000
SICK 0.612 0.032 19.203 0.000
CONSULT 0.756 0.025 30.786 0.000
FH1 0.055 0.011 5.063 0.000
FH2 0.088 0.011 7.966 0.000
SH 0.251 0.045 5.549 0.000
FH 0.479 0.043 11.219 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
4.1 LS Analysis
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
FIGURE 4.1. Latent state (LS) model for depression measured on three occasions
of measurement (T1–T3). Yik = observed variable i measured at time point k; eik =
measurement error variable; lik = factor loading. The variable names of the indicators
used in Mplus appear in parentheses.
We use data from Proyecto: La Familia (The Family Project; Roosa et al., 2008),
which assessed the health and adjustment of Mexican American children
and their parents every 2 years, beginning in the fifth grade for the target
children.* In the applications presented in this chapter, we fit longitudinal
SEMs to three waves of fathers’ (N = 323 complete cases) ratings of their own
depressive symptoms, using the 20-item Center for Epidemiological Stud-
ies Depression scale (CES-D; Radloff, 1977). Sample items include “You felt
depressed” and “You felt that everything you did was an effort.” Items were
answered on a 4-point scale ranging from “Rarely or none of the time” to
“Most or all of the time.” For the purpose of the present illustration, the items
of this questionnaire were aggregated to two test halves, so that there would
be two indicators of depression at each time point. The data set depression.
dat contains the summary data (means, standard deviations, and correla-
tions) for the two test halves measured on three occasions of measurement
(variables d11, d21, d12, d22, d13, and d23).
*I would like to thank Mark Roosa for providing the data set for the analyses.
84 Data Analysis with Mplus
intercept term aik, an occasion-specific latent state variable (State k), and
an occasion-specific measurement error (residual) variable eik:
where lik denotes a factor loading. Figure 4.1 shows an LS model for
depression, measured with two variables Y1k and Y2k on three occasions
of measurement (k = 1, 2, 3).
Figure 4.1 shows that all latent state variables can be correlated. These
correlations reflect the “true” stability of individual differences across time
(i.e., stability corrected for measurement error). The correlations between
the state variables allow us to examine whether a construct is better
described as a temporally stable disposition (“trait”) or as an attribute that
is prone to substantial occasion-specific fluctuations (a more state-like
construct). Correlations between state factors that approach 1 indicate
high stability of individual differences and therefore the construct can be
seen as rather trait-like (e.g., intelligence). In contrast, moderate to small
correlations between LS factors indicate that a construct depends strongly
on the situation (i.e., is more state-like; e.g., anger, mood states, cortisol
levels) or that some individuals changed more than others over time.
In addition, the correlations between the state factors also reveal
whether an autoregressive process is present in the data. This is the case
if the state factor correlations decline with increasing temporal distance
between the measurements. Latent autoregressive models, which can be
seen as special cases of LS models, are discussed in Section 4.3.
LS analyses typically do not involve any specific assumptions as to
the structure of the correlations between the state factors (as is the case,
e.g., in LST and latent autoregressive models; see below). This means that
all possible correlations between LS variables are usually estimated as
free, unconstrained parameters. This implies that the structural model of
LS models is typically saturated, meaning that no testable restrictions can
be derived from the structural model and therefore no degrees of freedom
are obtained from this part of the model. This aspect makes LS models
rather unrestrictive, which is one reason why they are often useful as a
starting point for a longitudinal statistical analysis. Causes of potential
misfit that may occur in more complex longitudinal SEMs are more easily
identified in the LS model, because in an LS model misfit can be caused
only by a misspecified measurement model (the structural model is satu-
rated and hence does not contribute to any global misfit of the model).
Structural Equation Models for Measuring Variability and Change 85
FIGURE 4.2. Mplus input file for the specification of a simple LS model for three
measurement occasions without measurement invariance restrictions.
Structural Equation Models for Measuring Variability and Change 87
state1 by d11
d21;
state2 by d12
d22;
state3 by d13
d23;
Value 18.474
Degrees of Freedom 6
P-Value 0.0052
In our basic LS model, we do not include latent means for the state factors
and instead make use of the Mplus default settings, according to which all
latent variable means are set to zero and all observed variable intercepts
(aik) are freely estimated. To the degree that no restrictions are imposed on
the intercepts of the observed variables, the intercepts will be identical to
the observed variable means [expected values E(Yik)]. This follows from the
decomposition of the observed variable means in the LS model, because the
LS means are set to zero by default in Mplus and the residual variables eik
have means of zero by definition (see Steyer et al., 1992):
In Section 4.1.4, I demonstrate how the means of the LS variables (that are
often of interest in longitudinal studies) can be identified and estimated in
Mplus.
88 Data Analysis with Mplus
Information Criteria
Estimate 0.080
90 Percent C.I. 0.040 0.123
Probability RMSEA <= .05 0.098
CFI/TLI
CFI 0.990
TLI 0.976
Value 0.012
It can be seen that the model shows a somewhat unsatisfactory fit, which
is indicated, in particular, by the significant chi-square test (c2 = 18.474,
df = 6, p = .0052) and the high RMSEA coefficient (.08), whereas the CFI
(0.99) and SRMR (0.012) coefficients show acceptable values, according
to commonly used criteria (see Box 3.7). A bad model fit in longitudinal
SEMs with multiple indicators can often be explained by the presence of
indicator-specific effects, which are not sufficiently accounted for in the
simple LS model.
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
FIGURE 4.3. LS model with autocorrelated error variables for modeling indicator-
specific effects. This model is underidentified in the present application because
there are only two indicators per time point.
correlations between adjacent time points are included (see Figure 4.4).
This model version is identified for the present data and the model param-
eters can thus be estimated. The relevant Mplus syntax to specify this
model is shown in Figure 4.5.
The model fit statistics for this model are slightly better compared
to the LS model without correlated residual variables, c2 = 8.536, df =
2, p = .014; RMSEA = .101, CFI = 0.995, SRMR = 0.008, except for the
higher RMSEA value. Nonetheless, the model with autocorrelated residu-
als between adjacent time points also shows a significant chi-square value,
indicating a significant discrepancy between the model and the data. In
addition, the RMSEA value is even higher than for the original model
(presumably due to the loss in degrees of freedom through the estimation
of four additional parameters). These results might suggest that indicator-
specific effects generalize beyond adjacent time points, so that the current
model (which only allows for correlated errors of adjacent time points)
does not sufficiently account for these effects.
Eid, Schneider, and Schwenkmezger (1999; see also Eid, 2000)
proposed an approach for modeling indicator-specific effects that does
not require the estimation of correlated error variables and thus allows
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
FIGURE 4.4. LS model with autocorrelated error variables for modeling indicator-
specific effects. In contrast to the model in Figure 4.3, error correlations are admit-
ted only between adjacent measurement occasions. This model is identified in the
present application.
Structural Equation Models for Measuring Variability and Change 91
FIGURE 4.5. Mplus model command for the specification of the LS model with
autocorrelated error variables between adjacent measurement occasions shown in
Figure 4.4.
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
λIS22 λIS23
λIS21
IS2
FIGURE 4.6. LS model with an indicator-specific (method) factor (IS2) for the sec-
ond indicator. The parameters lIS2k indicate factor loadings on the indicator-specific
factor. The indicator-specific factor is defined as a residual factor with respect to the
LS factors and is therefore by definition uncorrelated with all state factors that per-
tain to the same construct.
FIGURE 4.7. Mplus model command for the specification of the LS model with
one indicator-specific factor (is2) for the second indicator shown in Figure 4.6.
94 Data Analysis with Mplus
this issue, see Geiser, Eid, & Nussbeck, 2008). In addition, researchers
should carry out careful sensitivity analyses to examine in which way the
choice of different reference indicators does, or does not, affect the results
obtained for the structural model (e.g., intercorrelation of the state factors,
relationships of state factors to external variables, LS factor means and
variances).
In our example, indicator-specific effects are minor, as indicated by
the small standardized loadings on the is2 factor, which, as mentioned
above, range from .157 to .278. Therefore, at most .2782 = 7.7% of the
observed variability is due to indicator-specific effects, which can be seen
as unproblematic. From a substantive point of view this suggests that the
indicators are rather homogeneous (i.e., measure, to a large extent, the
same true score) and have only small indicator-specific variance compo-
nents.
In addition, we see that all covariances between the indicator-specific
factor and the three state factors were set to zero as requested. Therefore,
the is2 factor can be interpreted as a residual factor. In contrast, the
covariances among the state factors were freely estimated. All state covari-
ances are significantly different from zero (8.354 ≤ z ≤ 7.586). Hence,
a significant portion of individual differences with regard to depression
remained stable across time. The exact amount of stability, however, is dif-
ficult to assess based on the covariances, because covariances are unstan-
dardized measures of association. For this reason, researchers typically
report the correlations between the state factors as measures of stability
because correlations are standardized measures of association that do not
depend on the metric of the variables and are therefore easier to interpret
(see below).
In the unstandardized solution, we furthermore obtain the estimated
intercepts, factor variances, and residual variances. The unstandardized
intercepts are, in this case, identical to the observed variable means,
because the present model does not impose any restrictions on the mean
structure of the variables (cf. Box 4.2).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 1.000 0.000 999.000 999.000
D21 0.700 0.049 14.221 0.000
96 Data Analysis with Mplus
STATE2 BY
D12 1.000 0.000 999.000 999.000
D22 0.749 0.043 17.342 0.000
STATE3 BY
D13 1.000 0.000 999.000 999.000
D23 0.755 0.047 16.192 0.000
IS2 BY
D21 1.000 0.000 999.000 999.000
D22 0.623 0.378 1.650 0.099
D23 0.793 0.507 1.564 0.118
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.120 0.015 7.917 0.000
STATE3 WITH
STATE1 0.110 0.014 7.586 0.000
STATE2 0.128 0.015 8.354 0.000
Intercepts
D11 1.584 0.027 58.307 0.000
D21 1.354 0.021 65.148 0.000
D12 1.642 0.028 58.527 0.000
D22 1.389 0.023 60.772 0.000
D13 1.631 0.027 60.284 0.000
D23 1.402 0.023 62.159 0.000
Variances
STATE1 0.205 0.022 9.287 0.000
STATE2 0.230 0.023 10.185 0.000
STATE3 0.214 0.021 10.012 0.000
IS2 0.011 0.008 1.402 0.161
Residual Variances
D11 0.033 0.012 2.717 0.007
D21 0.028 0.009 3.134 0.002
D12 0.025 0.011 2.284 0.022
D22 0.036 0.007 5.326 0.000
D13 0.022 0.011 2.028 0.043
D23 0.035 0.008 4.692 0.000
observed and all latent variables are standardized, we can see that the
observed variables have rather high loadings on the state factors (≥ .849)
and rather small loadings on the indicator-specific factor (largest loading
= .278; see discussion above). The standardized loadings can again be
interpreted as the estimated correlations between each indicator and the
corresponding latent variable, because the state factors are uncorrelated
with the indicator-specific factor. Based on the indicators’ high standard-
ized loadings on the state factors and the low loadings on the is2 factor,
we can conclude that the indicators are highly reliable and rather homo-
geneous.
The standardized solution also provides us with the estimated latent
correlations between the state factors. In this example, the states show
moderately high correlations across time (.523 ≤ r ≤ .578), which indicates
a moderate degree of stability of individual differences over time. Note
that adjacent state factors (State 1 and State 2 as well as State 2 and State
3) are slightly more highly correlated than the more “distant” state factors
State 1 and State 3. This finding indicates the presence of a small autore-
gressive effect (cf. Section 4.3).
Of interest are also the estimated R 2 values for the observed variables,
which are reported at the end of the standardized solution. These values
can be used as estimates of the reliabilities of the indicators. As we already
saw based on the high standardized loadings, all six variables show high
reliabilities (Rel), which range between .785 and .907. This shows that the
six indicators of depression (here: questionnaire scales) allow for a high
precision of measurement (small error).
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 0.928 0.028 32.992 0.000
D21 0.849 0.029 28.934 0.000
STATE2 BY
D12 0.950 0.023 42.147 0.000
D22 0.874 0.024 36.260 0.000
STATE3 BY
D13 0.952 0.024 39.012 0.000
D23 0.862 0.026 33.195 0.000
98 Data Analysis with Mplus
IS2 BY
D21 0.278 0.099 2.810 0.005
D22 0.157 0.062 2.551 0.011
D23 0.203 0.076 2.687 0.007
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.552 0.045 12.293 0.000
STATE3 WITH
STATE1 0.523 0.047 11.194 0.000
STATE2 0.578 0.043 13.524 0.000
Intercepts
D11 3.244 0.139 23.299 0.000
D21 3.625 0.153 23.727 0.000
D12 3.257 0.140 23.313 0.000
D22 3.381 0.144 23.463 0.000
D13 3.354 0.143 23.420 0.000
D23 3.459 0.147 23.556 0.000
Variances
STATE1 1.000 0.000 999.000 999.000
STATE2 1.000 0.000 999.000 999.000
STATE3 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Residual Variances
D11 0.139 0.052 2.673 0.008
D21 0.202 0.066 3.063 0.002
D12 0.097 0.043 2.255 0.024
D22 0.212 0.042 5.063 0.000
D13 0.093 0.046 2.007 0.045
D23 0.215 0.048 4.501 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Widaman and Reise (1997; see also Meredith, 1993; Meredith & Horn, 2001)
distinguish between the following levels of measurement invariance:
Additional, even stricter, forms of parameter invariance such as, for example,
the assumption of equal LS factor variances and/or covariances across time,
can also be tested. In addition to the stability of the latent variance/covari-
ance structure, latent mean stability can be tested, which is often of inter-
est in longitudinal studies. A prerequisite for a meaningful interpretation
of latent mean change over time is that at least strong factorial invariance
can be assumed for the measurement model. Sometimes the assumption of
invariance in parameters across time can be made for only a subset of the
indicators, whereas other indicators show noninvariant parameters. In this
case, one speaks of partial measurement invariance. Byrne, Shavelson, and
Muthén (1989) discuss issues related to partial measurement invariance in
detail.
Structural Equation Models for Measuring Variability and Change 101
state1 by d11
d21 (1);
102 Data Analysis with Mplus
state2 by d12
d22 (1);
state3 by d13
d23 (1);
It is important to note that the specification (1) does not mean that
the parameter is supposed to be estimated to a value that is equal (or even
close) to 1. The number (1) merely causes Mplus to estimate the same
value for each of the loadings l21, l22, and l23. This specification does not
tell Mplus which specific value it is. The optimal value in terms of the ML
criterion is determined by Mplus when the parameters of the model are
estimated. (In order to fix parameters to specific values, the @ command
is used in Mplus.)
For the indicator-specific factor (is2), we also specify equal loadings
over time. Although equality of indicator-specific factor loadings is not
strictly necessary for most purposes, this restriction is often plausible and
meaningful (because the indicator remains the same). In addition, it con-
tributes to the parsimony of the model. The equality restrictions for this
factor are slightly different, because the parameters that are subject to the
equality constraint here are loadings on a time-unspecific factor. Given
that the first loading on this factor is again fixed to 1 by default in Mplus
(i.e., lIS21 = 1), we have to fix the remaining two indicator-specific factor
loadings (lIS22 and lIS23) to 1 as well to obtain equal loadings on this
factor. (This was not necessary for the three state factor loadings of the
second indicator because none of the loadings was fixed to any specific
value as the default.) The loadings lIS22 and lIS23 are therefore fixed to 1
in the model command using the @ symbol:
The complete Mplus model command for the model with weak
factorial invariance is shown in Figure 4.8. The fit indices reported by
Mplus show that—similar to the configural invariance model (see Sec-
tion 4.1.3)—the weak factorial invariance model shows a good fit to the
data, c2 = 5.094, df = 7, p = .6484, RMSEA = .00, p(RMSEA ≤ .05) = .925,
SRMR = .012, CFI = 1.00. To test whether the more restrictive model with
weak factorial invariance fits the data significantly worse than the con-
figural model with unrestricted factor loadings for the second indicator,
Structural Equation Models for Measuring Variability and Change 103
FIGURE 4.8. Mplus model command for the specification of the LS model with
one indicator-specific factor (is2) for the second indicator and time-invariant load-
ings (model of weak factorial invariance).
a chi-square difference test can be used, because the models are hierar-
chically nested (cf. Box 3.7). Nested means that the model with invariant
loadings can be seen as a special case of the baseline model with unre-
stricted loadings: The model with invariant loadings directly follows from
the configural model by constraining certain parameters.
Chi-square difference tests can be easily calculated by hand or using
specialized software. In Box 4.4, I provide a description of how chi-square
difference tests can be conducted automatically, based on Mplus outputs,
using Crayen’s (2010) software Chi-Square Difference Calculator (CDC). The
CDC program (version 3) as well as a manual can be found on the com-
panion website.
In our example, the comparison of the weak factorial invariance
model with the configural invariance model yields a chi-square differ-
ence value of c2∆ = 5.094 – 3.569 = 1.525. A chi-square value of 1.525
with df∆ = 7 – 3 = 4 is not significant at the .05 level (p = .82; see also
Box 4.4). This means that the more restrictive model with equal loadings
across time does not fit significantly worse than the configural model with
unrestricted loadings. Hence, the assumption of weak factorial invari-
ance is not rejected for these data. The AIC also favors the more restricted
model (AIC = 1072.933) over the configural model (AIC = 1079.408), even
though the difference in the values is not very large.
By looking at the unstandardized parameter estimates (rubric MODEL
RESULTS; see below), we can see that the desired invariance restric-
tions on the factor loadings were correctly implemented by Mplus. The
104 Data Analysis with Mplus
We then have to specify the location of the output file for (1) the more restric-
tive (H0) model and (2) the less restrictive (H1) model. Either the corre-
sponding paths can be typed directly into the respective fields or the Select
H0/H1 File buttons can be used to browse the computer for these files. By
clicking on Calculate, the program determines the chi-square difference
value, the differences in df, and the p-value for the chi-square difference
under the null hypothesis that the more restrictive model fits the data. In
addition, the program reports which model would be favored based on AIC,
Structural Equation Models for Measuring Variability and Change 105
BIC, and sample-size-adjusted BIC (see Box 3.7). In this example, the more
restrictive model with equal loadings is preferred according to all criteria
(see discussion in the text).
It is important to note that the CDC program does not check whether
the assumptions required for proper chi-square difference testing (see Box
3.7) are met. Specifically, the program does not check whether the two mod-
els chosen by the user are, in fact, nested. This has to be determined by the
user prior to the analysis. For certain non-nested models, statistical compari-
sons via the AIC, BIC, and/or sample-size-adjusted BIC may still be mean-
ingful, but chi-square difference testing would not be appropriate for these
models.
unstandardized state factor loadings of the second indicator were all esti-
mated to the same value ( lˆ 21 = lˆ 22 = lˆ 23 = 0.737 as intended). In addi-
tion, all three loadings show the same standard error (SE = 0.03), test
statistic (z = 24.857), and p-value (p < .001). All loadings on the indicator-
specific factor is2 were fixed to 1, as desired. The fact that no standard
errors (and consequently no test statistic) were estimated for any of the
is2 factor loadings shows that they are fixed parameters in this model—
as was intended.
The Mplus output for all models that include specific equality or other
constraints should be carefully screened for correct implementation of the
desired restrictions. The reason is that, especially with complex models,
errors can easily be made in the input specification. One should therefore
always check carefully whether the intended parameters were set equal (or
not equal). Sometimes an implausibly bad model fit can be a sign that one
or more parameter constraints were incorrectly specified. For example, it
can occur that erroneously, the same number is used twice for setting differ-
ent parameter groups equal. This may lead to the implementation of overly
restrictive and/or meaningless model constraints (e.g., the constraint that
loadings are set equal to intercepts). Such errors can lead to a significant
misfit of the model and to incorrect conclusions. Therefore, a careful inspec-
tion of the correct implementation of all restrictions should always be car-
ried out before interpreting key model results, conducting statistical model
comparisons, etc.
106 Data Analysis with Mplus
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 1.000 0.000 999.000 999.000
D21 0.737 0.030 24.857 0.000
STATE2 BY
D12 1.000 0.000 999.000 999.000
D22 0.737 0.030 24.857 0.000
STATE3 BY
D13 1.000 0.000 999.000 999.000
D23 0.737 0.030 24.857 0.000
IS2 BY
D21 1.000 0.000 999.000 999.000
D22 1.000 0.000 999.000 999.000
D23 1.000 0.000 999.000 999.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.118 0.015 7.963 0.000
STATE3 WITH
STATE1 0.108 0.014 7.622 0.000
STATE2 0.129 0.015 8.419 0.000
Intercepts
D11 1.584 0.027 58.666 0.000
D21 1.354 0.021 64.730 0.000
D12 1.642 0.028 58.461 0.000
D22 1.389 0.023 60.797 0.000
D13 1.631 0.027 60.172 0.000
D23 1.402 0.022 62.502 0.000
Variances
STATE1 0.195 0.019 10.505 0.000
STATE2 0.233 0.021 11.021 0.000
STATE3 0.218 0.020 10.982 0.000
IS2 0.007 0.002 3.456 0.001
Residual Variances
D11 0.040 0.009 4.462 0.000
D21 0.028 0.005 5.422 0.000
Structural Equation Models for Measuring Variability and Change 107
Notice that the standardized loadings (shown in the following) are not
invariant, but can vary across time in this model. The reason is that the
weak factorial invariance model does not involve invariance constraints
on the state factor variances and the residual variances. The standardized
loadings would be the same across time only if the state factor variances
and the residual variances were also held equal across time.
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 0.911 0.021 43.928 0.000
D21 0.867 0.022 39.796 0.000
STATE2 BY
D12 0.957 0.018 52.262 0.000
D22 0.867 0.020 43.413 0.000
STATE3 BY
D13 0.959 0.019 49.979 0.000
D23 0.855 0.021 41.265 0.000
IS2 BY
D21 0.221 0.033 6.773 0.000
D22 0.202 0.030 6.761 0.000
D23 0.206 0.030 6.774 0.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.552 0.045 12.346 0.000
STATE3 WITH
STATE1 0.521 0.047 11.179 0.000
STATE2 0.574 0.043 13.455 0.000
Intercepts
D11 3.264 0.138 23.632 0.000
D21 3.602 0.149 24.153 0.000
108 Data Analysis with Mplus
Variances
STATE1 1.000 0.000 999.000 999.000
STATE2 1.000 0.000 999.000 999.000
STATE3 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Residual Variances
D11 0.171 0.038 4.517 0.000
D21 0.200 0.036 5.529 0.000
D12 0.085 0.035 2.429 0.015
D22 0.207 0.033 6.316 0.000
D13 0.080 0.037 2.172 0.030
D23 0.227 0.034 6.744 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
This formula shows that the means of the LS variables, E(State k), can
be identified in a relatively straightforward way by setting the intercepts,
a1k, pertaining to the marker indicators, Y1k, to zero. If we set a1k = 0 for
all k, then all LS factor means become identified through the means of the
marker indicators:
In this way, the latent variable means are (1) identified and esti-
mable, and (2) they can potentially take on different values at each time
point, reflecting latent mean change over time. Furthermore, this speci-
fication ensures that the intercepts of the marker indicators are time-
invariant (each one takes on the same value of 0), reflecting measurement
invariance across time for these indicators. The intercepts of the remain-
ing (non-marker) indicators are set equal across time, following the same
principle as shown for the factor loadings in the weak factorial invariance
model.
The full Mplus model specification for the strong factorial invariance
model is shown in Figure 4.9. Note that commands related to the mean
structure (observed and latent variable means and intercept parameters)
involve variable names in brackets []. Depending on whether a variable
is exogenous or endogenous in the model (as determined automatically
by Mplus, based on the model specification), a variable name in brackets
refers to either the mean (for exogenous variables) or the intercept (for
endogenous variables) of that variable.
Figure 4.9 shows that the intercepts of the marker indicators (d11,
d12, and d13) are set to zero to identify the LS factor means:
FIGURE 4.9. Mplus model command for the specification of the LS model with
one indicator-specific factor (is2) for the second indicator, time-invariant loadings,
and time-invariant intercepts (model of strong factorial invariance).
Finally, the intercepts of the indicators d21, d22, and d23 are set
equal across time. Note that we use the number (2) to set these parameters
equal, which is a different number than the one used to set the loadings
equal. [The number (1) is still used to set the loadings of the indicators
d21, d22, and d23 equal across time.] The mean of the indicator-specific
factor does not have to be specified, because it is set to zero by default (as
is desired in this model).
The Mplus output shows that the model with strong factorial invari-
ance also fits the data well, c2 = 6.656, df = 9, p = .6729, RMSEA = .00,
p(RMSEA) ≤ .05 = .95, SRMR = .014, CFI = 1.00. In addition, the chi-
square difference test against the model with weak factorial invariance
is not significant, c2∆ = 1.562, df∆ = 2, p = .46, indicating that the more
restrictive, strong factorial invariance model is preferred. A comparison of
the AIC values also indicates that the more constrained model should be
chosen (strong factorial invariance model: AIC = 1070.494; weak factorial
invariance model: AIC = 1072.933). We can conclude that the assumption
of strong factorial invariance is not rejected in this example.
Based on the unstandardized parameter estimates (MODEL RESULTS,
see below), we can see that our equality constraints were correctly
Structural Equation Models for Measuring Variability and Change 111
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 1.000 0.000 999.000 999.000
D21 0.738 0.030 24.997 0.000
STATE2 BY
D12 1.000 0.000 999.000 999.000
D22 0.738 0.030 24.997 0.000
STATE3 BY
D13 1.000 0.000 999.000 999.000
D23 0.738 0.030 24.997 0.000
IS2 BY
D21 1.000 0.000 999.000 999.000
D22 1.000 0.000 999.000 999.000
D23 1.000 0.000 999.000 999.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.118 0.015 7.964 0.000
STATE3 WITH
STATE1 0.108 0.014 7.623 0.000
STATE2 0.129 0.015 8.419 0.000
Means
STATE1 1.583 0.026 60.157 0.000
STATE2 1.639 0.028 58.762 0.000
STATE3 1.634 0.027 60.652 0.000
112 Data Analysis with Mplus
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.187 0.049 3.861 0.000
D12 0.000 0.000 999.000 999.000
D22 0.187 0.049 3.861 0.000
D13 0.000 0.000 999.000 999.000
D23 0.187 0.049 3.861 0.000
Variances
STATE1 0.195 0.019 10.510 0.000
STATE2 0.233 0.021 11.023 0.000
STATE3 0.218 0.020 10.983 0.000
IS2 0.007 0.002 3.433 0.001
Residual Variances
D11 0.040 0.009 4.477 0.000
D21 0.028 0.005 5.425 0.000
D12 0.022 0.009 2.453 0.014
D22 0.035 0.006 6.299 0.000
D13 0.019 0.009 2.192 0.028
D23 0.037 0.006 6.608 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 0.911 0.021 43.988 0.000
D21 0.867 0.022 39.895 0.000
STATE2 BY
D12 0.956 0.018 52.311 0.000
D22 0.867 0.020 43.500 0.000
STATE3 BY
D13 0.959 0.019 50.014 0.000
D23 0.854 0.021 41.331 0.000
IS2 BY
D21 0.220 0.033 6.732 0.000
D22 0.202 0.030 6.721 0.000
D23 0.205 0.031 6.733 0.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
STATE2 WITH
STATE1 0.553 0.045 12.350 0.000
Structural Equation Models for Measuring Variability and Change 113
STATE3 WITH
STATE1 0.521 0.047 11.184 0.000
STATE2 0.574 0.043 13.463 0.000
Means
STATE1 3.583 0.181 19.792 0.000
STATE2 3.395 0.164 20.655 0.000
STATE3 3.499 0.169 20.664 0.000
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.498 0.134 3.707 0.000
D12 0.000 0.000 999.000 999.000
D22 0.456 0.124 3.677 0.000
D13 0.000 0.000 999.000 999.000
D23 0.465 0.126 3.686 0.000
Variances
STATE1 1.000 0.000 999.000 999.000
STATE2 1.000 0.000 999.000 999.000
STATE3 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Residual Variances
D11 0.171 0.038 4.532 0.000
D21 0.200 0.036 5.535 0.000
D12 0.086 0.035 2.448 0.014
D22 0.208 0.033 6.332 0.000
D13 0.081 0.037 2.191 0.028
D23 0.228 0.034 6.767 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Here the label (mean) is used to set the state means equal across time. (We
could have equally well used a number again.) In our example, the model
with equal means does not fit significantly worse than the strong factorial
invariance model with unequal means, according to the chi-square differ-
ence test, c2∆ = 5.181, df∆ = 2, p = .07, indicating that the observed mean
differences are not statistically significant at the .05 level. On the other hand,
according to the AIC index, the less constrained model would be preferred
(strong factorial invariance model with equal means: AIC = 1071.675; strong
factorial invariance model with unequal means: AIC = 1070.494), although
the difference in AIC values is rather small. In practice, a researcher would
have to take substantive considerations as well as the actual size of the mean
differences into account before making a final decision as to whether the
model with equal or unequal means should be preferred.
Structural Equation Models for Measuring Variability and Change 115
FIGURE 4.10. Mplus model command for the specification of the LS model with
one indicator-specific factor (is2) for the second indicator, time-invariant loadings,
time-invariant intercepts, and time-invariant error variances (model of strict facto-
rial invariance).
As mentioned above, the LS model is a useful baseline model for the anal-
ysis of more complex longitudinal SEMs. In the following material, we
discuss its extension to an LST model, using the same data example.
where lik again indicates the factor loading of Yik on the LS factor.
In contrast to the LS model, the LST model further decomposes the
LS factors into a latent trait factor (which reflects that part of the LS fac-
tors that is due to stable individual differences or trait effects) and an LS
residual variable zk that characterizes occasion-specific deviations of the
latent state from the latent trait scores (cf. Steyer et al., 1992, 1999):
Structural Equation Models for Measuring Variability and Change 117
Trait
Depression
γ1 γ3
γ2
ζ2
ζ1 ζ3
State 1: State 2: State 3:
Depression T1 Depression T2 Depression T3
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
FIGURE 4.11. Latent state–trait (LST) model for depression measured on three
occasions of measurement (T1–T3). Yik = observed variable i measured at time point
k; eik = measurement error variable; lik = first-order factor loading; gk = second-order
factor loading; zk = LS residual variable. The variable names of the indicators used in
Mplus appear in parentheses.
118 Data Analysis with Mplus
Trait
Depression
γ1 γ3
γ2
ζ2
ζ1 ζ3
State 1: State 2: State 3:
Depression T1 Depression T2 Depression T3
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
λIS22 λIS23
λIS21
IS2
FIGURE 4.12. LST model for depression measured on three occasions of mea-
surement (T1–T3). Yik = observed variable i measured at time point k; eik = measure-
ment error variable; lik = first-order factor loading; gk = second-order factor load-
ing; zk = LS residual variable. The model additionally contains an indicator-specific
(method) factor (IS2) for the second indicator. The parameters lIS2k indicate factor
loadings on the indicator-specific factor. The variable names of the indicators used in
Mplus appear in parentheses.
Structural Equation Models for Measuring Variability and Change 119
FIGURE 4.13. Mplus model command for the specification of the LST model
with one indicator-specific factor (is2) for the second indicator shown in Figure
4.12.
The Mplus model statement for estimating the LST model in Figure
4.12 is shown in Figure 4.13. Note that the identification of the latent vari-
able mean structure is slightly more complicated in the LST model than
in the LS model. This is the case at least if one is interested in estimating
the mean of the latent trait factor. (A more detailed discussion of differ-
ent options for including a latent mean structure in LST analyses can be
found in Box 4.7.)
One way to identify the latent trait factor mean as well as the latent
intercept parameters g0k is to fix the intercept parameters pertaining to
the marker indicators to zero, that is, a11 = a12 = a13 = 0.
In the next step, the intercept term pertaining to an LS factor cho-
sen as reference indicator of the latent trait is set to zero as well—in our
case, the intercept pertaining to State 1 (i.e., g01 = 0). This state factor is
chosen for simplicity, because Mplus automatically sets the loading of this
state factor on the trait factor to 1 (i.e., g1 = 1). This is because Mplus, by
default, treats the first variable listed after the by statement as the refer-
ence variable and sets its loading to 1. As a consequence, the mean of
the latent trait factor is then directly identified through the mean of the
marker variable Y11. Finally, the intercepts of the remaining state factors
(in our case, g02 and g03) are freely estimated. The mean of the is2 factor
120 Data Analysis with Mplus
Value 3.541
Degrees of Freedom 3
P-Value 0.3155
Estimate 0.024
90 Percent C.I. 0.000 0.100
Probability RMSEA <= .05 0.615
122 Data Analysis with Mplus
CFI/TLI
CFI 1.000
TLI 0.998
Value 0.008
Information Criteria
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
DEPDAD11 1.000 0.000 999.000 999.000
DEPDAD21 0.700 0.049 14.248 0.000
Structural Equation Models for Measuring Variability and Change 123
STATE2 BY
DEPDAD12 1.000 0.000 999.000 999.000
DEPDAD22 0.751 0.043 17.260 0.000
STATE3 BY
DEPDAD13 1.000 0.000 999.000 999.000
DEPDAD23 0.755 0.047 16.087 0.000
IS2 BY
DEPDAD21 1.000 0.000 999.000 999.000
DEPDAD22 0.612 0.375 1.629 0.103
DEPDAD23 0.799 0.518 1.542 0.123
TRAIT BY
STATE1 1.000 0.000 999.000 999.000
STATE2 1.168 0.134 8.731 0.000
STATE3 1.067 0.121 8.805 0.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
TRAIT 0.000 0.000 999.000 999.000
Means
TRAIT 1.584 0.027 58.184 0.000
Intercepts
DEPDAD11 0.000 0.000 999.000 999.000
DEPDAD21 0.246 0.079 3.112 0.002
DEPDAD12 0.000 0.000 999.000 999.000
DEPDAD22 0.156 0.073 2.147 0.032
DEPDAD13 0.000 0.000 999.000 999.000
DEPDAD23 0.171 0.078 2.197 0.028
STATE1 0.000 0.000 999.000 999.000
STATE2 -0.209 0.214 -0.975 0.330
STATE3 -0.060 0.194 -0.311 0.756
Variances
TRAIT 0.103 0.019 5.553 0.000
IS2 0.011 0.008 1.390 0.164
Residual Variances
DEPDAD11 0.033 0.012 2.711 0.007
DEPDAD21 0.028 0.009 3.091 0.002
DEPDAD12 0.025 0.011 2.297 0.022
DEPDAD22 0.036 0.007 5.351 0.000
DEPDAD13 0.022 0.011 1.986 0.047
DEPDAD23 0.035 0.008 4.592 0.000
STATE1 0.103 0.016 6.308 0.000
STATE2 0.089 0.017 5.146 0.000
STATE3 0.098 0.017 5.901 0.000
124 Data Analysis with Mplus
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
DEPDAD11 0.928 0.028 32.979 0.000
DEPDAD21 0.849 0.029 29.046 0.000
STATE2 BY
DEPDAD12 0.950 0.023 41.661 0.000
DEPDAD22 0.874 0.024 36.051 0.000
STATE3 BY
DEPDAD13 0.953 0.025 38.686 0.000
DEPDAD23 0.862 0.026 32.966 0.000
IS2 BY
DEPDAD21 0.278 0.100 2.785 0.005
DEPDAD22 0.155 0.062 2.502 0.012
DEPDAD23 0.205 0.077 2.659 0.008
TRAIT BY
STATE1 0.707 0.046 15.245 0.000
STATE2 0.782 0.045 17.409 0.000
STATE3 0.739 0.046 16.188 0.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
TRAIT 0.000 0.000 999.000 999.000
Means
TRAIT 4.937 0.453 10.909 0.000
Intercepts
DEPDAD11 0.000 0.000 999.000 999.000
DEPDAD21 0.656 0.220 2.980 0.003
DEPDAD12 0.000 0.000 999.000 999.000
DEPDAD22 0.378 0.182 2.085 0.037
DEPDAD13 0.000 0.000 999.000 999.000
DEPDAD23 0.420 0.197 2.131 0.033
STATE1 0.000 0.000 999.000 999.000
STATE2 -0.435 0.440 -0.989 0.323
STATE3 -0.130 0.416 -0.313 0.755
Variances
TRAIT 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Structural Equation Models for Measuring Variability and Change 125
Residual Variances
DEPDAD11 0.139 0.052 2.668 0.008
DEPDAD21 0.201 0.067 3.026 0.002
DEPDAD12 0.098 0.043 2.267 0.023
DEPDAD22 0.212 0.042 5.081 0.000
DEPDAD13 0.092 0.047 1.966 0.049
DEPDAD23 0.215 0.049 4.417 0.000
STATE1 0.500 0.066 7.634 0.000
STATE2 0.389 0.070 5.534 0.000
STATE3 0.454 0.067 6.741 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
The latent mean structure is also shown in the output. In this model,
the unstandardized latent trait mean is estimated to be 1.584. The inter-
cepts of State 2 and State 3 are estimated to be ĝ 02 = –0.209 and ĝ 03 =
–0.060. These values reflect slight mean differences relative to the first
state factor. However, these mean differences are not significant (p ≥ .33).
Further steps of an LST analysis may involve tests of measurement
invariance to obtain more parsimonious and substantively meaningful
models. Given that we have already discussed the basic principles of test-
ing measurement invariance for the LS model in Section 4.1.4, I will not
go into the details here. The procedure shown for the LS model can be
easily transferred and applied to the LST model as well.
LST models were primarily developed as variability models to model
systematic, occasion- specific fluctuations of measurements around a
constant trait value. As we saw, these models can be used to estimate
variance components to separate stable influences from occasion-specific
126 Data Analysis with Mplus
For a long time, autoregressive models (e.g., Jöreskog, 1979a, 1979b) were
seen as the “gold standard” of methodology for analyzing longitudinal
data. With the development of latent growth curve models (LGCMs; see
Section 4.5), autoregressive models have become somewhat less popu-
lar (for a critique of autoregressive models, see Hertzog & Nesselroade,
1987). Nonetheless, autoregressive models are useful to answer a variety
of important research questions in longitudinal studies.
Autoregressive models are based on the (fairly plausible) idea that
previous behavior is the best predictor of present behavior. For this
reason, repeatedly measured manifest or latent variables are connected
through so-called autoregressions in autoregressive models (as reflected in
the path coefficients b1–b4 in the manifest autoregressive model shown
in Figure 4.14). The strengths of autoregressive (or autopredictive) effects
as well as the stability of individual differences across time can be exam-
ined by looking at the size of the autoregressive coefficients as well as
the amount of residual variance (variance of the residual variables ejk in
Figure 4.14). When the ratio of residual variance to total variance is large,
this means that the autoprediction is rather weak—indicating individual
differences in change over time. (In manifest autoregressive models that
do not account for measurement error, such error can be an additional or
even alternative explanation for a large residual variance.)
However, the goal of an analysis of autoregressive models is usually
not only to determine the degree of stability of (or changes in) a single
construct over time. Often, an additional goal is to predict individual dif-
ferences not explained by the autoregressive effect(s) through external
variables that are also included in the model. The impact of such addi-
tional variables can be examined by means of so-called cross-lagged effects
Structural Equation Models for Measuring Variability and Change 127
ε 12 ε 13
β5 β7
β6 β8
Y21 β3 Y22 β4 Y23
(c11) (c12) (c13)
ε 22 ε 23
FIGURE 4.14. Manifest autoregressive model for two manifest variables Yjk (j =
construct, k = time point) measuring two constructs (j = 1, 2; depression and com-
petence) on three measurement occasions (k = 1, 2, 3): b1–b4 = path coefficients rep-
resenting autoregressive effects; b5 –b8 = path coefficients representing cross-lagged
effects; ejk = residual variables. The variable names used in Mplus appear in paren-
theses.
(see path coefficients b5 –b8 in Figure 4.14). Cross-lagged effects are the
effects of additional, temporally preceding variables that are included in
the model in addition to the autoregressive effects. The goal is to explain
at least part of the “leftover” variability that is not explained by the autore-
gressive effect(s). In other words, one attempts to explain the “unstable”
part of individual differences by external variables that may be possible
causes of change.
*I would like to thank David A. Cole for providing the data used for this example.
Figure 4.15 shows the Mplus input file for the manifest autoregres-
sive model with cross-lagged effects from Figure 4.14. The autoregressive
and cross-lagged effects in the model are represented by partial regression
coefficients (also referred to as path coefficients; cf. Section 3.5). Therefore,
on statements are used in the Mplus model command to specify these
effects. As noted above, the model represents a special case of a manifest
path analysis, here specified for longitudinal data.
An important aspect in the model specification is that the residual
variables shown in Figure 4.14 are assumed to be uncorrelated. This has
FIGURE 4.15. Mplus input file for the specification of the simple manifest autore-
gressive model for three measurement occasions shown in Figure 4.14.
Structural Equation Models for Measuring Variability and Change 129
to be explicitly specified for the residuals on the third time point, because
Mplus otherwise estimates a residual correlation between these residuals
by default. For this reason, we add the following command in the model
specification:
This command overrides the Mplus default and sets the residual correla-
tion for the final outcome variables to zero. All other residual correlations
are automatically set to zero by Mplus. (As we will see shortly, the original
model with uncorrelated residuals does not fit the data in this example, so
that this restriction will be relaxed again in later analysis steps.)
The model fit indices reported by Mplus indicate a very bad fit of
the model to the data, c2 = 201.90, df = 6, p < .001, CFI = 0.86, RMSEA
= 0.24, SRMR = 0.09 (the full output can be found on the companion
website). The lack of fit can in this case be explained by the fact that the
model does not allow the residuals at the same time point to be corre-
lated. The estimation of these additional parameters is necessary in this
example, because there are shared occasion-specific effects between the
depression and competencies constructs at the same time point, in line
with LST theory. In other words, children who rated themselves as more
depressed also tended to report a lower level of competencies at the same
time point. These situation-specific effects are not sufficiently accounted
for in the standard autoregressive model with uncorrelated residual vari-
ables, shown in Figure 4.14. This can be changed by allowing the residual
variables at the same time point to correlate (see Figure 4.16).
In addition to situation-specific effects, the present example also
requires the inclusion of second-order autoregressive effects to achieve
an acceptable fit. Second-order autoregressive effects are autoregressive
effects between nonadjacent time points (represented by the path coef-
ficients b9 and b10 in Figure 4.16). The Mplus model statement for the
extended model is shown in Figure 4.17. This model fits the data well,
c2 = 2.74, df = 2, p = .25, CFI = 0.999, RMSEA = 0.03, SRMR = 0.01. In
addition, it fits significantly better than the original model with uncor-
related residual variables and no second-order autoregressive effects, Dc2
= 199.16, Ddf = 4, p < .001. The AIC value also indicates that the less
restrictive model in Figure 4.16 (AIC = 2493.08) should be preferred (the
AIC value for the model in Figure 4.14 was 2690.01). In the following, the
parameter estimates for the extended model are presented.
130 Data Analysis with Mplus
β9
ε 12 ε 13
β5 β7
β6 β8
Y21 β3 Y22 β4 Y23
(c11) (c12) (c13)
ε 22 ε 23
β10
FIGURE 4.16. Manifest autoregressive model for two manifest variables Yjk (j =
construct, k = time point) measuring two constructs (j = 1, 2; depression and com-
petence) on three measurement occasions (k = 1, 2, 3). b1–b4 = path coefficients
representing first-order autoregressive effects; b9 and b10 = latent path coefficients
representing second-order autoregressive effects; b5 –b8 = path coefficients repre-
senting cross-lagged effects; ejk = residual variables. In contrast to the model shown
in Figure 4.14, this model allows for second-order autoregressive effects as well as
correlated residual variables at the same measurement occasion. The variable names
used in Mplus appear in parentheses.
FIGURE 4.17. Mplus model command for the specification of the extended
autoregressive model with second-order autoregressive effects and correlated resid-
ual variables shown in Figure 4.16.
Structural Equation Models for Measuring Variability and Change 131
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
D12 ON
D11 0.526 0.037 14.225 0.000
C11 -0.085 0.020 -4.184 0.000
D13 ON
D12 0.510 0.040 12.782 0.000
D11 0.156 0.035 4.411 0.000
C12 -0.056 0.020 -2.749 0.006
C12 ON
C11 0.422 0.037 11.352 0.000
D11 -0.392 0.068 -5.786 0.000
C13 ON
C12 0.386 0.044 8.797 0.000
C11 0.192 0.037 5.181 0.000
D12 -0.436 0.072 -6.066 0.000
D12 WITH
C12 -0.040 0.005 -8.067 0.000
D13 WITH
C13 -0.035 0.005 -7.868 0.000
Intercepts
D12 0.288 0.050 5.789 0.000
D13 0.210 0.050 4.198 0.000
C12 1.335 0.091 14.640 0.000
C13 0.995 0.107 9.322 0.000
Residual Variances
D12 0.061 0.004 17.117 0.000
D13 0.051 0.003 17.117 0.000
C12 0.206 0.012 17.117 0.000
C13 0.207 0.012 17.117 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
D12 ON
D11 0.531 0.033 15.947 0.000
C11 -0.156 0.037 -4.209 0.000
132 Data Analysis with Mplus
D13 ON
D12 0.517 0.037 13.820 0.000
D11 0.160 0.036 4.426 0.000
C12 -0.100 0.036 -2.754 0.006
C12 ON
C11 0.440 0.036 12.214 0.000
D11 -0.224 0.038 -5.872 0.000
C13 ON
C12 0.361 0.040 9.128 0.000
C11 0.187 0.036 5.214 0.000
D12 -0.231 0.038 -6.150 0.000
D12 WITH
C12 -0.353 0.036 -9.778 0.000
D13 WITH
C13 -0.344 0.036 -9.434 0.000
Intercepts
D12 0.909 0.159 5.718 0.000
D13 0.674 0.163 4.146 0.000
C12 2.386 0.197 12.142 0.000
C13 1.663 0.200 8.315 0.000
Residual Variances
D12 0.610 0.031 19.374 0.000
D13 0.526 0.030 17.590 0.000
C12 0.657 0.032 20.657 0.000
C13 0.579 0.031 18.679 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
It turns out that all autoregressive effects (including both first- and
second-order effects) are statistically significant (see estimated path coef-
ficients and p-values under D12 ON D11, D13 ON D12 D11, C12 ON
C11, and C13 ON C12 C11). This indicates that a significant portion
of individual differences remained stable over time for both depression
and competencies. Nonetheless, individual differences were not perfectly
stable over time. Part of the “instability” in depression and competencies
Structural Equation Models for Measuring Variability and Change 133
3.5.2), namely, that measurement error is not explicitly accounted for and
that the path coefficients and their SEs may therefore be biased. In the
context of autoregressive models this can lead to bias in the estimation of
stability, change, and/or cross-lagged effects and corresponding tests of
statistical significance.
Autoregressive models with latent variables (latent autoregressive
models, LAMs) explicitly account for random measurement error by
using multiple indicators at each time point. LAMs have the additional
advantages that they allow testing of measurement models, modeling of
indicator-specific effects, and examining of the degree of measurement
invariance over time (cf. Section 4.1.4). These models were discussed in
detail by Jöreskog (1979a, 1979b) as well as by Hertzog and Nesselroade
(1987).
LAMs are based on the LS model described in Section 4.1.2. LAMs
restrict the covariances of the LS variables by imposing an autoregressive
structure. Cross-lagged models are based on multiconstruct versions of
the basic LS model and additionally include cross-lagged effects between
LS variables. Figure 4.18 shows a model for two constructs, each of which
is measured by two manifest variables Yijk at each of three time points (i
= indicator, j = construct, k = time point). Note that the structural model
(containing the paths between the LS variables) is equivalent to the struc-
ture of the manifest first-order autoregressive model with uncorrelated
residual variables in Figure 4.14.
A good starting point for a latent autoregressive analysis is the basic
LS model discussed in Section 4.1. This model can be used to clarify tech-
nical issues such as (1) the question of whether the basic measurement
β6 β8
State 21 β3 State 22 β4 State 23
(Comp1) (Comp2) (Comp3)
ζ22 ζ23
λ121 λ221 λ122 λ222 λ123 λ223
FIGURE 4.18. Latent autoregressive model for 12 manifest indicators Yijk (i = indi-
cator, j = construct, k = time point) measuring two constructs (j = 1, 2; depression
and competence) on three measurement occasions (k = 1, 2, 3). lijk = state factor
loadings; b1–b4 = latent path coefficients representing autoregressive effects; b5 –b8
= latent path coefficients representing cross-lagged effects; eijk = measurement error
variables; zjk = LS residual variable. The variable names used in Mplus appear in
parentheses.
model fits for each construct, (2) whether indicator-specific factors need
to be included due to indicator heterogeneity, and (3) which level of mea-
surement invariance can be assumed across time for each construct. The
final best-fitting LS model can then serve as the basis for specifying LAMs
with or without cross-lagged effects.
Figure 4.19 shows the Mplus input file for the estimation of the LAM
from Figure 4.18. For simplicity, we first show a simple model version
without indicator-specific factors. Later on, we discuss a more complex
model that accounts for indicator-specific effects.
136 Data Analysis with Mplus
FIGURE 4.19. Mplus input file for the specification of the simple latent autore-
gressive model for three measurement occasions shown in Figure 4.18.
model here (the full output can be found on the companion website). The
bad model fit can be explained by several factors. First, our analysis of the
manifest autoregressive model already indicated that the inclusion of cor-
related residuals and second-order autoregressive effects was necessary. It
is likely that these effects are also required in the LAM. Furthermore, in
our multiple-indicator LAM there may be indicator-specific effects due to
the use of the same indicators over time (cf. Section 4.1.3).
The LAM can be extended in a similar way as the manifest model.
In addition, we can include indicator-specific factors as discussed for the
LS model in Section 4.1.3. The extended model is shown in Figure 4.20
(p. 143). Figure 4.21 (p. 144) shows the corresponding Mplus model
command.
Even though the extended model still shows a significant chi-square
value, the fit of the model is much improved compared to the model in
Figure 4.18, c2 = 46.72, df = 28, p = .02; CFI = 0.996; RMSEA = 0.03; SRMR
= 0.02. A comparison of the AIC values also reveals that the more complex
model (AIC = 2601.53) should be preferred over the initial model (AIC =
2929.24). In the following, we show the parameter estimates obtained for
the more complex model.
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
DEPR1 BY
D11 1.000 0.000 999.000 999.000
D21 0.957 0.032 29.972 0.000
DEPR2 BY
D12 1.000 0.000 999.000 999.000
D22 0.909 0.030 30.500 0.000
DEPR3 BY
D13 1.000 0.000 999.000 999.000
D23 0.924 0.028 32.900 0.000
ISD BY
D21 1.000 0.000 999.000 999.000
D22 1.093 0.277 3.941 0.000
D23 1.385 0.425 3.256 0.001
COMP1 BY
C11 1.000 0.000 999.000 999.000
C21 1.031 0.059 17.501 0.000
138 Data Analysis with Mplus
COMP2 BY
C12 1.000 0.000 999.000 999.000
C22 0.913 0.054 16.840 0.000
COMP3 BY
C13 1.000 0.000 999.000 999.000
C23 0.857 0.049 17.346 0.000
ISC BY
C21 1.000 0.000 999.000 999.000
C22 1.167 0.165 7.084 0.000
C23 1.151 0.172 6.685 0.000
DEPR2 ON
DEPR1 0.720 0.055 13.141 0.000
COMP1 -0.030 0.034 -0.878 0.380
DEPR3 ON
DEPR2 0.590 0.063 9.406 0.000
DEPR1 0.153 0.051 3.033 0.002
COMP2 -0.049 0.033 -1.471 0.141
COMP2 ON
COMP1 0.693 0.071 9.715 0.000
DEPR1 -0.208 0.100 -2.080 0.037
COMP3 ON
COMP2 0.597 0.101 5.930 0.000
COMP1 0.195 0.073 2.679 0.007
DEPR2 -0.243 0.104 -2.338 0.019
ISD WITH
DEPR1 0.000 0.000 999.000 999.000
DEPR2 0.000 0.000 999.000 999.000
DEPR3 0.000 0.000 999.000 999.000
COMP1 0.000 0.002 0.151 0.880
COMP2 -0.002 0.002 -1.089 0.276
COMP3 0.000 0.002 -0.242 0.809
ISC WITH
COMP1 0.000 0.000 999.000 999.000
COMP2 0.000 0.000 999.000 999.000
COMP3 0.000 0.000 999.000 999.000
DEPR1 -0.011 0.004 -2.823 0.005
DEPR2 -0.007 0.003 -2.601 0.009
DEPR3 -0.008 0.003 -3.122 0.002
ISD -0.002 0.001 -1.622 0.105
DEPR2 WITH
COMP2 -0.035 0.004 -8.310 0.000
DEPR3 WITH
COMP3 -0.036 0.004 -8.584 0.000
Structural Equation Models for Measuring Variability and Change 139
COMP1 WITH
DEPR1 -0.093 0.008 -11.127 0.000
Intercepts
D11 0.306 0.013 22.757 0.000
D21 0.274 0.013 21.672 0.000
D12 0.274 0.013 20.497 0.000
D22 0.253 0.013 20.079 0.000
D13 0.283 0.013 21.352 0.000
D23 0.256 0.013 20.152 0.000
C11 2.074 0.024 85.370 0.000
C21 2.121 0.026 82.201 0.000
C12 2.091 0.024 88.301 0.000
C22 2.115 0.025 83.180 0.000
C13 2.082 0.025 82.382 0.000
C23 2.184 0.025 85.819 0.000
Variances
DEPR1 0.085 0.006 13.491 0.000
ISD 0.003 0.001 2.508 0.012
COMP1 0.225 0.021 10.628 0.000
ISC 0.054 0.012 4.578 0.000
Residual Variances
D11 0.018 0.002 8.220 0.000
D21 0.010 0.002 5.108 0.000
D12 0.014 0.002 6.928 0.000
D22 0.014 0.002 7.801 0.000
D13 0.010 0.002 5.251 0.000
D23 0.008 0.002 3.802 0.000
C11 0.111 0.012 9.473 0.000
C21 0.086 0.011 7.648 0.000
C12 0.084 0.011 7.546 0.000
C22 0.094 0.012 8.178 0.000
C13 0.087 0.013 6.982 0.000
C23 0.085 0.011 7.693 0.000
DEPR2 0.039 0.003 11.783 0.000
DEPR3 0.038 0.003 12.569 0.000
COMP2 0.096 0.013 7.650 0.000
COMP3 0.102 0.013 7.901 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
DEPR1 BY
D11 0.909 0.013 72.457 0.000
D21 0.925 0.012 75.929 0.000
DEPR2 BY
D12 0.929 0.011 82.978 0.000
D22 0.896 0.013 71.363 0.000
140 Data Analysis with Mplus
DEPR3 BY
D13 0.948 0.011 90.229 0.000
D23 0.914 0.012 76.546 0.000
ISD BY
D21 0.193 0.038 5.013 0.000
D22 0.211 0.041 5.215 0.000
D23 0.266 0.046 5.721 0.000
COMP1 BY
C11 0.818 0.022 36.435 0.000
C21 0.794 0.023 34.255 0.000
COMP2 BY
C12 0.858 0.021 40.148 0.000
C22 0.729 0.026 28.125 0.000
COMP3 BY
C13 0.871 0.021 42.281 0.000
C23 0.741 0.026 28.907 0.000
ISC BY
C21 0.377 0.040 9.400 0.000
C22 0.446 0.043 10.403 0.000
C23 0.439 0.043 10.224 0.000
DEPR2 ON
DEPR1 0.708 0.046 15.362 0.000
COMP1 -0.048 0.055 -0.880 0.379
DEPR3 ON
DEPR2 0.583 0.057 10.175 0.000
DEPR1 0.149 0.049 3.048 0.002
COMP2 -0.078 0.053 -1.475 0.140
COMP2 ON
COMP1 0.678 0.057 11.966 0.000
DEPR1 -0.125 0.060 -2.080 0.038
COMP3 ON
COMP2 0.551 0.084 6.591 0.000
COMP1 0.176 0.065 2.699 0.007
DEPR2 -0.137 0.059 -2.330 0.020
ISD WITH
DEPR1 0.000 0.000 999.000 999.000
DEPR2 0.000 0.000 999.000 999.000
DEPR3 0.000 0.000 999.000 999.000
COMP1 0.010 0.064 0.150 0.880
COMP2 -0.091 0.082 -1.111 0.267
COMP3 -0.020 0.083 -0.241 0.809
Structural Equation Models for Measuring Variability and Change 141
ISC WITH
COMP1 0.000 0.000 999.000 999.000
COMP2 0.000 0.000 999.000 999.000
COMP3 0.000 0.000 999.000 999.000
DEPR1 -0.169 0.051 -3.295 0.001
DEPR2 -0.157 0.059 -2.691 0.007
DEPR3 -0.185 0.056 -3.307 0.001
ISD -0.152 0.087 -1.741 0.082
DEPR2 WITH
COMP2 -0.570 0.051 -11.108 0.000
DEPR3 WITH
COMP3 -0.570 0.049 -11.625 0.000
COMP1 WITH
DEPR1 -0.675 0.032 -21.093 0.000
Intercepts
D11 0.954 0.051 18.870 0.000
D21 0.909 0.050 18.254 0.000
D12 0.859 0.049 17.521 0.000
D22 0.842 0.049 17.264 0.000
D13 0.895 0.050 18.049 0.000
D23 0.845 0.049 17.323 0.000
C11 3.579 0.114 31.374 0.000
C21 3.446 0.110 31.414 0.000
C12 3.702 0.117 31.514 0.000
C22 3.487 0.111 31.443 0.000
C13 3.454 0.111 31.226 0.000
C23 3.598 0.114 31.647 0.000
Variances
DEPR1 1.000 0.000 999.000 999.000
ISD 1.000 0.000 999.000 999.000
COMP1 1.000 0.000 999.000 999.000
ISC 1.000 0.000 999.000 999.000
Residual Variances
D11 0.174 0.023 7.652 0.000
D21 0.107 0.022 4.907 0.000
D12 0.136 0.021 6.538 0.000
D22 0.152 0.021 7.249 0.000
D13 0.101 0.020 5.060 0.000
D23 0.091 0.025 3.703 0.000
C11 0.331 0.037 9.004 0.000
C21 0.227 0.032 7.146 0.000
C12 0.265 0.037 7.219 0.000
C22 0.256 0.033 7.652 0.000
C13 0.241 0.036 6.702 0.000
C23 0.230 0.032 7.209 0.000
142 Data Analysis with Mplus
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
IS21
(isd)
λIS211 λIS213
λIS212
ε 111 ε 211 ε 112 ε 212 ε 113 ε 213
FIGURE 4.21. Mplus model command for the specification of the extended
latent autoregressive model with second- order autoregressive effects, correlated
latent residual variables, and indicator-specific factors shown in Figure 4.20.
Imperfect reliabilities are one reason for the differences seen in the results
between the manifest and the latent autoregressive models. An additional
reason may be the relatively large amounts of specific variance in the
competence test halves (the competence scale is composed of rather het-
erogeneous items), as indicated by the relatively large occasion-specific
effects (standardized loadings on the indicator- specific factor ISC:
.377 ≤ lˆ standardized
≤ .446).
As in an LS analysis, the degree of invariance of the parameters of
the measurement model over time should be examined for the LAM in
subsequent analysis steps. (The possibility to test for measurement invari-
ance is another advantage of the LAM over the manifest model.) I do not
discuss such an analysis here in detail given that the principle of invari-
ance testing was already illustrated in detail for the LS model in Section
4.1.4. The steps shown for the LS model can be transferred to the LAM.
The companion website contains a sample input file for a version of the
LAM assuming strong factorial invariance across time.
In latent change (LC) models (McArdle & Hamagami, 2001; Raykov, 1993;
Steyer et al., 1997, 2000), change is not measured “indirectly” through
the residual variables in autoregressions (as in autoregressive models), but
directly through so-called latent difference variables. Latent difference vari-
ables represent interindividual differences in true intraindividual change
over time—that is, change scores corrected for random measurement
error. For this reason, these types of models have also been referred to as
true change models (Steyer et al., 1997). Because latent difference variables
are of key importance in this approach, LC models are sometimes also
called latent difference models.
Latent difference variables can be included in SEMs in a relatively
simple way by reformulating LS models (see Section 4.1). Figure 4.22 illus-
trates the reparameterization of an LS model (A) to an LC model (B) based
on a single construct measured by two indicators on two measurement
occasions. The basic idea of the LC model is that the LS variable at time
2 (State 2) is a function of (1 times) the initial LS variable (State 1) plus (1
times) change (in terms of the latent difference variable State 2 – State 1):
A
State 1 State 2
(state1) (state2)
λ1 λ2 λ1 λ2
ε 11 ε 21 ε 12 ε 22
B State 2–State 1
(diff2_1)
1 0
State 1 1 State 2
(state1) (state2)
λ1 λ2 λ1 λ2
ε 11 ε 21 ε 12 ε 22
FIGURE 4.22. LS model and corresponding latent change (LC) model for depres-
sion measured by two indicators Yik (i = indicator, k = time point) on two measure-
ment occasions: li = time-invariant state factor loadings; eik = measurement error
variable. A: LS model. B: LC model. The variable names used in the Mplus appear
in parentheses.
Note that this decomposition is not at all restrictive: We add one variable
(the factor State 1) to another variable (here the factor State 2) and subtract
this variable again at the same time. This decomposition into initial state
plus change can be accomplished in an LS model through the inclusion
of an additional latent variable, the latent difference variable (State 2 –
State 1), as shown in Figure 4.22B. This means that the latent difference
becomes an explicit part of the model in the form of an additional latent
variable, which makes it possible to study interindividual differences in
intraindividual change at the latent (error-free) level.
A latent difference variable can be used like any other latent vari-
able in an SEM. For example, the latent difference variable can be cor-
related with other variables or it can serve as an exogenous (predictor)
Structural Equation Models for Measuring Variability and Change 147
We use the LS model for depression from Section 4.1.4 (data set depression.
dat) as the basis for our illustration of the analysis of LC models in Mplus.
As the starting point, we use the LS model version with an indicator-specific
factor for the second indicator and strong factorial invariance (in which the
factor loadings and intercepts are assumed to be equal across time). Given
that we have already tested the assumption of strong factorial invariance in
Section 4.1.4 (and found that we did not have to reject it), we can directly
start with the analysis of LC models. In general, the assumption of strong
factorial invariance has to be tested first before one can start an LC analysis.
The LS model is equivalent to both a baseline and a neighbor change model;
that is, all three models imply the same mean and covariance structure. The
reason is that the LC models are just reparameterizations of the LS model.
Therefore, the fit of all three models is also identical.
Structural Equation Models for Measuring Variability and Change 149
A
State 2–State 1 State 3–State 1
(diff2_1) (diff3_1)
1 1
State 1 State 2 0 State 3
0
1
(state1) (state2) 1 (state3)
λ1 λ2 λ1 λ2 λ1 λ2
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
B
State 2–State 1 State 3–State 2
(diff2_1) (diff3_2)
1 1 1
State 1 State 2 0 State 3
0
1
(state1) (state2) 1 (state3)
λ1 λ2 λ1 λ2 λ1 λ2
ε 11 ε 21 ε 12 ε 22 ε 13 ε 23
FIGURE 4.23. LC model for two indicators Yik (i = indicator, k = time point) and
three measurement occasions. li = time-invariant state factor loadings; eik = mea-
surement error variable. A: Baseline change model. B: Neighbor change model. The
models are equivalent and equivalent to an LS model with invariant parameters. The
(often required) indicator-specific factor for the second indicator is not shown here
to save space. The variable names used in the Mplus input appear in parentheses.
Figure 4.24 shows the Mplus model command for the specification
of the baseline change model. (The specification of the neighbor change
150 Data Analysis with Mplus
model is shown in Figure 4.25, page 158.) The remaining input does not
differ from the input for the LS model (cf. Figure 4.9).
The measurement models for the three LS factors are formulated in
analogy to the LS model with time-invariant loadings. According to the
parameterization in Figure 4.23, the latent difference variables are not
directly connected with observed indicators. Steyer et al. (1997, 2000)
discuss an alternative equivalent parameterization of LC models in which
the latent difference variables are directly connected to observed indica-
tors. For didactic reasons, I did not choose this parameterization here.
FIGURE 4.24. Mplus model command for the specification of the baseline
change model shown in Figure 4.23A. An additional indicator-specific factor for the
second indicator was included in the model. (This factor is not shown in Figure
4.23.)
Structural Equation Models for Measuring Variability and Change 151
diff2_1 by d11@0;
diff3_1 by d11@0;
These commands thus merely serve to introduce the names of the latent
variables diff2_1 and diff3_1—they have no deeper implication for
the model specification other than defining the names of the latent differ-
ence variables so that they can be used in the model. The relevant com-
mands that assign the meaning as difference scores to these variables are
shown next. We use on commands to specify that the LS variables at T2
and T3 are functions of the initial LS variable (state1) plus the latent
difference, respectively:
state2@0;
state3@0;
Given that the endogenous variables state2 and state3 are per-
fectly determined by initial state and change in this model, their residual
variables were fixed to zero. As a consequence, the residual terms can-
not be correlated with other variables. These correlations have to be sup-
pressed in Mplus, because otherwise the model would be misspecified
and would also, very likely, produce estimation problems:
Next, we request the estimation of the latent means for state1 as well
as for the two latent difference variables diff2_1 and diff3_1 (other-
wise, Mplus would keep these means fixed at 0 by default):
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 1.000 0.000 999.000 999.000
D21 0.738 0.030 24.997 0.000
STATE2 BY
D12 1.000 0.000 999.000 999.000
D22 0.738 0.030 24.997 0.000
STATE3 BY
D13 1.000 0.000 999.000 999.000
D23 0.738 0.030 24.997 0.000
DIFF2_1 BY
D11 0.000 0.000 999.000 999.000
DIFF3_1 BY
D11 0.000 0.000 999.000 999.000
IS2 BY
D21 1.000 0.000 999.000 999.000
D22 1.000 0.000 999.000 999.000
D23 1.000 0.000 999.000 999.000
STATE2 ON
STATE1 1.000 0.000 999.000 999.000
DIFF2_1 1.000 0.000 999.000 999.000
STATE3 ON
STATE1 1.000 0.000 999.000 999.000
DIFF3_1 1.000 0.000 999.000 999.000
STATE2 WITH
STATE3 0.000 0.000 999.000 999.000
DIFF3_1 0.000 0.000 999.000 999.000
STATE3 WITH
DIFF2_1 0.000 0.000 999.000 999.000
154 Data Analysis with Mplus
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
DIFF2_1 0.000 0.000 999.000 999.000
DIFF3_1 0.000 0.000 999.000 999.000
DIFF2_1 WITH
STATE1 -0.077 0.015 -5.324 0.000
DIFF3_1 WITH
STATE1 -0.088 0.015 -5.856 0.000
DIFF2_1 0.099 0.015 6.451 0.000
Means
STATE1 1.583 0.026 60.157 0.000
DIFF2_1 0.056 0.027 2.078 0.038
DIFF3_1 0.052 0.027 1.900 0.057
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.187 0.049 3.861 0.000
D12 0.000 0.000 999.000 999.000
D22 0.187 0.049 3.861 0.000
D13 0.000 0.000 999.000 999.000
D23 0.187 0.049 3.861 0.000
Variances
STATE1 0.195 0.019 10.510 0.000
DIFF2_1 0.193 0.021 9.385 0.000
DIFF3_1 0.198 0.021 9.453 0.000
IS2 0.007 0.002 3.433 0.001
Residual Variances
D11 0.040 0.009 4.477 0.000
D21 0.028 0.005 5.425 0.000
D12 0.022 0.009 2.453 0.014
D22 0.035 0.006 6.299 0.000
D13 0.019 0.009 2.192 0.028
D23 0.037 0.006 6.608 0.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 0.911 0.021 43.988 0.000
D21 0.867 0.022 39.895 0.000
Structural Equation Models for Measuring Variability and Change 155
STATE2 BY
D12 0.956 0.018 52.311 0.000
D22 0.867 0.020 43.500 0.000
STATE3 BY
D13 0.959 0.019 50.014 0.000
D23 0.854 0.021 41.331 0.000
DIFF2_1 BY
D11 0.000 0.000 999.000 999.000
DIFF3_1 BY
D11 0.000 0.000 999.000 999.000
IS2 BY
D21 0.220 0.033 6.732 0.000
D22 0.202 0.030 6.720 0.000
D23 0.205 0.031 6.733 0.000
STATE2 ON
STATE1 0.915 0.049 18.583 0.000
DIFF2_1 0.909 0.049 18.466 0.000
STATE3 ON
STATE1 0.946 0.052 18.195 0.000
DIFF3_1 0.953 0.052 18.430 0.000
STATE2 WITH
STATE3 999.000 999.000 999.000 999.000
DIFF3_1 999.000 999.000 999.000 999.000
STATE3 WITH
DIFF2_1 999.000 999.000 999.000 999.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 999.000 999.000 999.000 999.000
STATE3 999.000 999.000 999.000 999.000
DIFF2_1 0.000 0.000 999.000 999.000
DIFF3_1 0.000 0.000 999.000 999.000
DIFF2_1 WITH
STATE1 -0.399 0.054 -7.345 0.000
DIFF3_1 WITH
STATE1 -0.446 0.052 -8.646 0.000
DIFF2_1 0.508 0.050 10.167 0.000
Means
STATE1 3.583 0.181 19.792 0.000
DIFF2_1 0.127 0.061 2.069 0.039
DIFF3_1 0.116 0.061 1.892 0.058
156 Data Analysis with Mplus
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.498 0.134 3.706 0.000
D12 0.000 0.000 999.000 999.000
D22 0.456 0.124 3.677 0.000
D13 0.000 0.000 999.000 999.000
D23 0.465 0.126 3.686 0.000
Variances
STATE1 1.000 0.000 999.000 999.000
DIFF2_1 1.000 0.000 999.000 999.000
DIFF3_1 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Residual Variances
D11 0.171 0.038 4.532 0.000
D21 0.200 0.036 5.535 0.000
D12 0.086 0.035 2.448 0.014
D22 0.208 0.033 6.332 0.000
D13 0.081 0.037 2.191 0.028
D23 0.228 0.034 6.767 0.000
STATE2 0.000 999.000 999.000 999.000
STATE3 0.000 999.000 999.000 999.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
The positive signs of the latent difference factor means indicate that
the latent depression scores have (slightly) increased between time 1 and
time 2 as well as between time 1 and time 3.
The estimated p-values for the latent difference factor means indicate
that the mean difference between T1 and T2 is significantly different from
zero at the .05 level for a two-tailed test, z = 2.078, p = .038, indicating
that the increase in depression from T1 to T2 was statistically significant.
The mean difference between T1 and T3 failed to reach statistical signifi-
cance at the .05 level (at least for a two-tailed test), z = 1.900, p = 0.057.
The estimated variances of state1 (0.195), diff2_1 (0.193), and
diff3_1 (0.198) are significantly different from zero (all ps < .001), show-
ing that there are individual differences in the initial latent depression
state as well as in the LC scores. Notice that the R 2 values for state2 and
state3 equal 1, reflecting the fact that these state factors are, by defini-
tion, perfectly determined by initial state plus change in this model.
Figure 4.25 shows the model statement for the specification of the
neighbor change model for the same data example. The specification of
the latent difference variable diff2_1 is identical to the specification of
this variable in the baseline change model. This variable still represents
change from time 1 to time 2. The second latent difference variable in
the model is now labeled diff3_2 to make clear that in the neighbor
change version, we consider change between the adjacent time points 2
and 3 (rather than between time 1 and 3). According to the path diagram
in Figure 4.23B, the latent state variable state3 is now determined by
three components: (1) the initial state factor (state1), (2) LC between
time 1 and time 2 (diff2_1), and (3) LC between time 2 and time 3
(diff3_2). Therefore, the on statement for state3 now lists all three
components and sets the paths of all these components to 1 (using the
@1 command). The remaining commands are analogous to the baseline
change model.
The model fit of the neighbor change model is again identical to the
fit of the LS model with strong factorial invariance as well as to the fit of
the baseline change model (see Mplus outputs on the companion website
158 Data Analysis with Mplus
FIGURE 4.25. Mplus model command for the specification of the neighbor
change model shown in Figure 4.23B. An additional indicator-specific factor for the
second indicator was included in the model. (This factor is not shown in Figure
4.23.)
diff2_1 are identical to the means in the baseline change model (M1 =
1.583, M2–1 = 0.056). The mean of the latent difference variable diff3_2
is estimated to M3–2 = –0.004. Within rounding error, this is equal to the
difference in means between state3 and state2 in the LS model with
strong factorial invariance. The mean difference of –0.004 is not statisti-
cally significant, z = –0.163, p = .871. We can conclude that the average
level of depression has not significantly changed between T2 and T3.
In addition, we again obtain estimates for the variances of state1
(0.195), diff2_1 (0.193), and diff3_2 (0.192). These variances indi-
cate the degree to which individuals differ in their initial latent depression
scores as well as in their LC scores. In additional analysis steps, one could
add external variables (covariates) to the model that may be able to partly
or fully explain individual differences in initial status and change (for a
substantive example, see Steyer et al., 2000).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 1.000 0.000 999.000 999.000
D21 0.738 0.030 24.997 0.000
STATE2 BY
D12 1.000 0.000 999.000 999.000
D22 0.738 0.030 24.997 0.000
STATE3 BY
D13 1.000 0.000 999.000 999.000
D23 0.738 0.030 24.997 0.000
DIFF2_1 BY
D11 0.000 0.000 999.000 999.000
DIFF3_2 BY
D11 0.000 0.000 999.000 999.000
IS2 BY
D21 1.000 0.000 999.000 999.000
D22 1.000 0.000 999.000 999.000
D23 1.000 0.000 999.000 999.000
STATE2 ON
STATE1 1.000 0.000 999.000 999.000
DIFF2_1 1.000 0.000 999.000 999.000
160 Data Analysis with Mplus
STATE3 ON
STATE1 1.000 0.000 999.000 999.000
DIFF2_1 1.000 0.000 999.000 999.000
DIFF3_2 1.000 0.000 999.000 999.000
STATE2 WITH
STATE3 0.000 0.000 999.000 999.000
DIFF3_2 0.000 0.000 999.000 999.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
DIFF2_1 0.000 0.000 999.000 999.000
DIFF3_2 0.000 0.000 999.000 999.000
DIFF2_1 WITH
STATE1 -0.077 0.015 -5.324 0.000
DIFF3_2 WITH
STATE1 -0.010 0.012 -0.832 0.405
DIFF2_1 -0.093 0.015 -6.142 0.000
Means
STATE1 1.583 0.026 60.157 0.000
DIFF2_1 0.056 0.027 2.078 0.038
DIFF3_2 -0.004 0.026 -0.163 0.871
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.187 0.049 3.861 0.000
D12 0.000 0.000 999.000 999.000
D22 0.187 0.049 3.861 0.000
D13 0.000 0.000 999.000 999.000
D23 0.187 0.049 3.861 0.000
Variances
STATE1 0.195 0.019 10.510 0.000
DIFF2_1 0.193 0.021 9.385 0.000
DIFF3_2 0.192 0.021 9.341 0.000
IS2 0.007 0.002 3.433 0.001
Residual Variances
D11 0.040 0.009 4.477 0.000
D21 0.028 0.005 5.425 0.000
D12 0.022 0.009 2.453 0.014
D22 0.035 0.006 6.299 0.000
D13 0.019 0.009 2.192 0.028
D23 0.037 0.006 6.608 0.000
STATE2 0.000 0.000 999.000 999.000
STATE3 0.000 0.000 999.000 999.000
Structural Equation Models for Measuring Variability and Change 161
Two-Tailed
Estimate S.E. Est./S.E. P-Value
STATE1 BY
D11 0.911 0.021 43.988 0.000
D21 0.867 0.022 39.895 0.000
STATE2 BY
D12 0.956 0.018 52.311 0.000
D22 0.867 0.020 43.501 0.000
STATE3 BY
D13 0.959 0.019 50.014 0.000
D23 0.854 0.021 41.331 0.000
DIFF2_1 BY
D11 0.000 0.000 999.000 999.000
DIFF3_2 BY
D11 0.000 0.000 999.000 999.000
IS2 BY
D21 0.220 0.033 6.732 0.000
D22 0.202 0.030 6.720 0.000
D23 0.205 0.031 6.733 0.000
STATE2 ON
STATE1 0.915 0.049 18.583 0.000
DIFF2_1 0.909 0.049 18.466 0.000
STATE3 ON
STATE1 0.946 0.052 18.195 0.000
DIFF2_1 0.939 0.061 15.318 0.000
DIFF3_2 0.939 0.054 17.503 0.000
STATE2 WITH
STATE3 999.000 999.000 999.000 999.000
DIFF3_2 999.000 999.000 999.000 999.000
IS2 WITH
STATE1 0.000 0.000 999.000 999.000
STATE2 999.000 999.000 999.000 999.000
STATE3 999.000 999.000 999.000 999.000
DIFF2_1 0.000 0.000 999.000 999.000
DIFF3_2 0.000 0.000 999.000 999.000
DIFF2_1 WITH
STATE1 -0.399 0.054 -7.345 0.000
162 Data Analysis with Mplus
DIFF3_2 WITH
STATE1 -0.053 0.064 -0.835 0.403
DIFF2_1 -0.485 0.052 -9.383 0.000
Means
STATE1 3.583 0.181 19.792 0.000
DIFF2_1 0.127 0.061 2.069 0.039
DIFF3_2 -0.010 0.060 -0.163 0.871
Intercepts
D11 0.000 0.000 999.000 999.000
D21 0.498 0.134 3.706 0.000
D12 0.000 0.000 999.000 999.000
D22 0.456 0.124 3.677 0.000
D13 0.000 0.000 999.000 999.000
D23 0.465 0.126 3.686 0.000
Variances
STATE1 1.000 0.000 999.000 999.000
DIFF2_1 1.000 0.000 999.000 999.000
DIFF3_2 1.000 0.000 999.000 999.000
IS2 1.000 0.000 999.000 999.000
Residual Variances
D11 0.171 0.038 4.532 0.000
D21 0.200 0.036 5.535 0.000
D12 0.086 0.035 2.448 0.014
D22 0.208 0.033 6.332 0.000
D13 0.081 0.037 2.191 0.028
D23 0.228 0.034 6.767 0.000
STATE2 0.000 999.000 999.000 999.000
STATE3 0.000 999.000 999.000 999.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
The specific appeal of LC models lies in the fact that these models
allow us to analyze “true” change over time (i.e., change scores that are
corrected for measurement error) and to relate latent change variables to
other variables. For example, we may try to explain why depression scores
changed more in some individuals than in others. In addition, latent dif-
ference variables may serve as independent or mediator variables.
From a technical point of view, such extended analyses are easily car-
ried out in Mplus by adding additional variables to the model and includ-
ing additional on statements to either regress latent difference variables
on external variables or use the latent difference variables themselves as
predictors of other variables. For example, if age was another variable in
the data set (which is not the case here), we could examine whether indi-
vidual differences in latent change between time 1 and time 2 could be
explained by age differences. The additional specification would look as
follows, where age is an additional variable in the data set:
diff2_1 on age;
LGCMs (Bollen & Curran, 2006; Duncan et al., 2006; Meredith & Tisak,
1984, 1990) are increasingly used in the social sciences to analyze lon-
gitudinal data. In contrast to the LC models that were discussed in the
previous section, LGCMs explicitly focus on modeling the form of change
over time. One key question in growth curve modeling is which math-
ematical function adequately describes individual latent growth trajecto-
ries over time (e.g., linear or quadratic growth). Important research ques-
tions that can be answered by LGCMs are (among others):
• How large is the latent mean with regard to an attribute on the first
occasion of measurement (mean initial value)?
• Are there individual differences in the true initial value (“inter-
cept”) and, if yes, how large are they?
• Does the attribute change over time and, if yes, what is the form
and strength of this change? On average, is there an increase or
164 Data Analysis with Mplus
1 1 1 1 0 1 2 3
ε1 ε2 ε3 ε4
FIGURE 4.26. First-order LGCM for one observed variable Yk (k = time point)
measured on four occasions of measurement (k = 1, 2, 3, 4). ek = error variable. The
model assumes linear growth and equal spacing between time points. The variable
names used in the Mplus input appear in parentheses.
166 Data Analysis with Mplus
It can be seen in Figure 4.26 that all factor loadings in the model are
set to specific values (i.e., the loadings are not estimated as free param-
eters, but are fixed a priori). This is done so that the factors can be inter-
preted as intercept (or “initial status”) factor and linear slope factor.
Specifically, all Yk variables have unit loadings on the intercept fac-
tor and zero intercepts (all constant intercept parameters ak are set to 0;
this is not explicitly shown in the figure). Furthermore, Y1 loads only on
the intercept factor (but not on the slope factor). The intercept factor can
therefore be interpreted as the true score variable (t1) pertaining to Y1 in
the sense of classical test theory: Y1 = t1 + e1 = Intercept + e1. In other
words, the intercept factor represents individual differences in the true
(error-free) initial (T1) anxiety scores.
The linear slope factor can be interpreted as the true score difference
variable between the true score variable pertaining to Y2 and the true
score variable pertaining to Y1:
Linear Slope = t2 – t1
This is because Y2 has unit loadings on both the intercept and linear slope
factors (for more details on how to define factors in LGCMs based on true
score theory see Geiser, Keller, & Lockhart, in press): Y2 = t1 + (t2 – t1)
+ e2 = Intercept + Linear Slope + e2. This definition parallels the defini-
tion of latent difference factors in LC models (see Section 4.4). In contrast
to the LC factors considered in the previous section, however, the lin-
ear slope factor in LGCM represents true individual differences in linear
change over time, because the loadings of the (equally spaced) measure-
ments are set to 0, 1, 2, and 3, respectively.
With regard to our example, children with larger positive scores on
the slope factor show a stronger linear increase in their anxiety scores
than children with smaller slope factor scores. Children with negative
slope factor scores show decreasing anxiety values over time. The double-
headed arrow between the intercept and slope factors indicates that inter-
cept and slope can be correlated. (Their correlation can be estimated as a
separate model parameter in Mplus.) For example, individuals with larger
initial anxiety scores may show different trajectories than those with
smaller initial anxiety scores.
How can we answer the research questions discussed above with
the linear LGCMs? The estimated mean of the intercept factor provides
an answer to the question of what the mean true anxiety level was at T1.
A large intercept factor mean indicates a high level of anxiety at T1. The
Structural Equation Models for Measuring Variability and Change 167
FIGURE 4.27. Mplus input file for the specification of the linear first-order LGCM
for anxiety measured by one indicator across four measurement occasions shown in
Figure 4.26. In this input file, the simplified syntax for LGCMs is used.
slope of the growth curve is a random effect that can vary across individu-
als (i.e., the slope is a variable, not a constant; individuals can differ in
how much they change over time). The same is true for the intercept fac-
tor that also represents a random effect (i.e., individuals can differ in their
true initial anxiety scores). (The distinction of random vs. fixed effects is
revisited in Chapter 5, in which we discuss multilevel models.)
Behind the | sign, the repeatedly measured observed anxiety vari-
ables are listed that serve as indicators of the growth factors. The @ sign
is used to fix the loadings of the indicators on the slope factor so that the
slope factor properly represents the assumed form of the growth trajecto-
ries. In line with the assumption of linear growth over time, the loadings
are fixed to 0, 1, 2, and 3 on the slope factor (cf. Figure 4.26).
In the default specification of growth curve models with the spe-
cial syntax shown above, Mplus automatically fixes all loadings on the
intercept factor to 1 and sets all constant intercept terms (ak) to 0. In
addition, Mplus automatically estimates the means of the intercept and
slope factors. It is important to be aware of these default settings, because
growth models can also be specified using the conventional Mplus syntax
for standard CFA. If the conventional CFA syntax is used, the intercept
parameters ak have to be explicitly set to 0 by the user and the estimation
Structural Equation Models for Measuring Variability and Change 169
LGCMs can also be specified using the standard Mplus syntax for confirma-
tory factor models. This alternative specification is slightly more complicated
than using the special Mplus syntax for LGCMs, because the loading pat-
tern for both the intercept and slope factors has to be explicitly specified,
the intercepts ak have to be explicitly set to 0, and the estimation of the
latent growth factor means has to be explicitly requested by the user. On the
other hand, this method makes the underlying specification of LGCMs more
explicit and shows that these models are special cases of confirmatory factor
models. In the following, we show the alternative equivalent specification
of the same first-order LGCM for the anxiety data using the conventional
Mplus syntax:
The complete input files for both types of specifications are provided on the
companion website.
of the latent growth factor means has to be explicitly requested (see Box
4.13).
The Mplus plot option is very useful in the analysis of LGCMs,
because it can be used to analyze, for example, the individual observed
and model-implied growth trajectories. The plot option also allows the
plotting of the average observed trajectory along with the average model-
implied trajectories. These curves can be useful, for example, to exam-
ine model fit. If, for instance, many individuals appear to deviate from a
close-to-linear trajectory, this might be a sign that a linear growth model
may not be appropriate for the data. Note that the plot option can be used
only when individual data (and not summary data) are used as input for
the analyses. (The difference between individual and summary data is
explained in Chapter 2.)
In order to obtain plots of the growth curves, we request the Mplus
plot type 3 (type = plot3;). Using the series option, we specify
that the values of the variables a1–a4 are to be plotted as a function of
time (linear here refers to the name chosen for the linear slope factor in
the model statement):
170 Data Analysis with Mplus
One obtains a set of graphs that can be opened in the Mplus output
by selecting the menu option Graph → View Graphs, as described in
detail in Section 4.5.1.2. Before looking at the graphs, we first examine the
numerical output provided by Mplus. The model fit statistics indicate that
the linear LGCM shows a rather mediocre fit to the anxiety data. This is
shown especially by the large chi-square and RMSEA values: c2 = 27.288,
df = 5, p < .001, CFI = 0.98, RMSEA = 0.096, SRMR = 0.061.
For comparison, we can estimate a so-called intercept-only model
in which only an intercept, but no slope factor, is assumed. This model
implies that there is no true change in the latent anxiety scores over time.
(According to the intercept-only model, any change seen in the observed
scores over time is due only to measurement error.) By comparing the
linear LGCM to the intercept-only model, we can see whether the LGCM
at least fits better than a no-change model. Figure 4.28 shows the Mplus
input file for the specification of the intercept-only model.
FIGURE 4.28. Mplus input file for the specification of the first-order intercept-
only model for the anxiety data using the simplified syntax for LGCMs.
Structural Equation Models for Measuring Variability and Change 171
The basic model specification for LGCMs can be retained for the
specification of the intercept-only model. Three additional commands are
included to set the mean, variance, and covariance of the linear slope fac-
tor to zero, given that these parameters are not estimated in the intercept-
only model. According to this specification, only the intercept factor
remains in the model.
The plot option is unchanged, so that we can examine the model-
implied latent growth curves in the intercept-only model as well and com-
pare them to the model-implied growth curves in the linear LGCM. The
model fit statistics for the intercept-only model reveal that (as expected)
it shows a very poor fit to the data: c2 = 176.355, df = 8, p < .001, CFI =
0.857, RMSEA = 0.208, SRMR = 0.141. (The full output for the intercept-
only model can be found on the companion website.) A comparison of the
AIC values indicates that the intercept-only model (AIC = 1296.065) fits
worse than the linear LGCM (AIC = 1152.997). Hence, the assumption
of linear change is at least preferred over the assumption of no change.
(Nonetheless, the fit of the linear model is not optimal, indicating that the
assumption of linear change for every individual in the sample may be
too restrictive.)
In the following, we discuss the most important parameters for the
first-order linear LGCM, despite its modest fit. Under MODEL RESULTS
we see that the mean and variance of both the intercept and slope factor
are significantly different from zero (assuming an alpha level of .05). This
finding indicates that the mean initial latent level of anxiety is significantly
different from zero (MINTERCEPT = 0.698, z = 34.866, p < .001). There also
is significant variability in the initial latent anxiety scores (VarINTERCEPT
= 0.151, z = 11.996, p < .001).
With regard to the slope factor, we can see that, on average, there is
a significant linear decline in latent anxiety scores across time, because
the estimated slope factor mean has a negative sign (MLINEAR = –0.062,
z = –10.405, p < 0.001). This means that the mean model-implied lin-
ear growth curve has a negative slope. In addition, there is a signifi-
cant amount of individual differences in the slope values around the
mean growth curve, as indicated by the significant slope factor variance
(VarLINEAR = 0.007, z = 4.667, p < .001). This means that the children
differ not only in their initial latent anxiety scores, but also in their tra-
jectories over time: Some children show a stronger increase (or decrease)
in their anxiety scores than others. (This finding will be made clearer
shortly when we look at individual growth trajectories in the requested
plots.)
172 Data Analysis with Mplus
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
INTERC |
A1 1.000 0.000 999.000 999.000
A2 1.000 0.000 999.000 999.000
A3 1.000 0.000 999.000 999.000
A4 1.000 0.000 999.000 999.000
LINEAR |
A1 0.000 0.000 999.000 999.000
A2 1.000 0.000 999.000 999.000
A3 2.000 0.000 999.000 999.000
A4 3.000 0.000 999.000 999.000
L NEAR WITH
INTERC -0.011 0.003 -3.485 0.000
Means
INTERC 0.698 0.020 34.866 0.000
LINEAR -0.062 0.006 -10.405 0.000
Intercepts
A1 0.000 0.000 999.000 999.000
A2 0.000 0.000 999.000 999.000
A3 0.000 0.000 999.000 999.000
A4 0.000 0.000 999.000 999.000
Variances
INTERC 0.151 0.013 11.996 0.000
LINEAR 0.007 0.001 4.667 0.000
Residual Variances
A1 0.067 0.008 8.675 0.000
A2 0.048 0.004 10.965 0.000
A3 0.048 0.004 11.180 0.000
A4 0.040 0.006 6.545 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
INTERC |
A1 0.833 0.019 43.608 0.000
A2 0.906 0.023 40.204 0.000
A3 0.911 0.030 29.912 0.000
A4 0.903 0.040 22.597 0.000
LINEAR |
A1 0.000 0.000 999.000 999.000
A2 0.193 0.021 9.160 0.000
A3 0.389 0.042 9.259 0.000
A4 0.578 0.065 8.861 0.000
LINEAR WITH
INTERC -0.349 0.070 -5.001 0.000
Means
INTERC 1.796 0.091 19.789 0.000
LINEAR -0.753 0.108 -6.965 0.000
Intercepts
A1 0.000 0.000 999.000 999.000
A2 0.000 0.000 999.000 999.000
A3 0.000 0.000 999.000 999.000
A4 0.000 0.000 999.000 999.000
Variances
INTERC 1.000 0.000 999.000 999.000
LINEAR 1.000 0.000 999.000 999.000
Residual Variances
A1 0.306 0.032 9.622 0.000
A2 0.264 0.024 11.037 0.000
A3 0.265 0.023 11.747 0.000
A4 0.215 0.031 6.812 0.000
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
interc linear on x y z;
Using this command, the intercept and slope factors are regressed on
three independent variables named x, y, and z. The growth factors could
also serve as independent variables themselves. In this case, they would
appear on the right-hand side of an on statement.
symbol in the Mplus menu. Figure 4.30 shows different options avail-
able for viewing graphs. The option Sample means provides a plot of
the observed anxiety means as a function of time. The option Estimated
means results in a plot of the mean model-implied growth curve. When
choosing Sample and estimated means, both the average sample and
the average model-implied means are plotted in a single graph. Observed
individual values yields the observed individual growth curves for all
individuals in the sample. This option was chosen to produce the graph
shown in Figure 4.32.
After clicking on Observed individual values, the window Proper-
ties for viewing individual curves appears (see Figure 4.31). Here the
user can choose between different options for viewing individual growth
Structural Equation Models for Measuring Variability and Change 175
FIGURE 4.30. Choosing a specific graph in Mplus. In this case, we want to look
at the observed individual growth curves.
176 Data Analysis with Mplus
FIGURE 4.31. Choosing the type, number, and order of individual growth curves
to be displayed. Here: curves based on observed data (Individual data), 10 curves per
graph (Number of curves: 10), and consecutive order as in the data file, beginning
with the first individual in the data set (Consecutive order/Starting at index: 1).
FIGURE 4.32. Individual growth curves of the first 10 individuals in the data
set anxiety.dat. By moving the cursor over a specific data point, the number of the
individual as well as his or her numeric scores for all time points are shown.
FIGURE 4.33. Choosing a specific graph in Mplus. In this case, we want to look
at the observed and estimated mean growth curves.
(see Figure 4.33). Figure 4.34 shows the mean observed and the mean
model-implied growth curves for the linear LGCM. Figure 4.35 shows
the same curves for the intercept-only model, which assumes no change
over time. Note that the model-implied curve in the intercept-only model
is exactly parallel to the x-axis, a pattern that does not correspond well
to the observed means (which clearly show a decline over time). Mplus
graphs can be saved as a picture file in either *.DIB, *.EMF, or *.JPEG
format. This can be done by choosing the menu option Graph → Export
plot to, as shown in Figure 4.36.
Linear Quadratic
Intercept Factor
Slope Factor Slope Factor
(interc)
(linear) (quadr)
1 0 3 0 1 4 9
1 1 1 1 2
ε1 ε2 ε3 ε4
FIGURE 4.37. First-order LGCM for one observed variable Yk (k = time point)
measured on four occasions of measurement (k = 1, 2, 3, 4). ek = error variable.
The model assumes curvilinear (quadratic) growth and equal spacing between time
points. The variable names used in the Mplus input appear in parentheses.
Structural Equation Models for Measuring Variability and Change 181
FIGURE 4.38. Mplus input file for the specification of the quadratic first-order
LGCM for anxiety measured by one indicator across four measurement occasions
shown in Figure 4.37. In this input file, the simplified syntax for LGCMs is used.
182 Data Analysis with Mplus
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
INTERC |
A1 1.000 0.000 999.000 999.000
A2 1.000 0.000 999.000 999.000
A3 1.000 0.000 999.000 999.000
A4 1.000 0.000 999.000 999.000
LINEAR |
A1 0.000 0.000 999.000 999.000
A2 1.000 0.000 999.000 999.000
A3 2.000 0.000 999.000 999.000
A4 3.000 0.000 999.000 999.000
QUADR |
A1 0.000 0.000 999.000 999.000
A2 1.000 0.000 999.000 999.000
A3 4.000 0.000 999.000 999.000
A4 9.000 0.000 999.000 999.000
Structural Equation Models for Measuring Variability and Change 183
LINEAR WITH
INTERC -0.026 0.021 -1.201 0.230
QUADR WITH
INTERC 0.002 0.005 0.374 0.709
LINEAR -0.008 0.005 -1.565 0.117
Means
INTERC 0.720 0.020 35.617 0.000
LINEAR -0.114 0.018 -6.452 0.000
QUADR 0.016 0.005 3.113 0.002
Intercepts
A1 0.000 0.000 999.000 999.000
A2 0.000 0.000 999.000 999.000
A3 0.000 0.000 999.000 999.000
A4 0.000 0.000 999.000 999.000
Variances
INTERC 0.163 0.022 7.441 0.000
LINEAR 0.048 0.023 2.110 0.035
QUADR 0.001 0.001 0.673 0.501
Residual Variances
A1 0.036 0.019 1.870 0.061
A2 0.053 0.007 7.720 0.000
A3 0.035 0.006 5.498 0.000
A4 0.072 0.019 3.890 0.000
Two-Tailed
Estimate S.E. Est./S.E. P-Value
...
LINEAR WITH
INTERC -0.290 0.166 -1.746 0.081
QUADR WITH
INTERC 0.154 0.370 0.415 0.678
LINEAR -1.205 0.471 -2.558 0.011
Means
INTERC 1.784 0.128 13.928 0.000
LINEAR -0.518 0.146 -3.552 0.000
QUADR 0.514 0.415 1.237 0.216
(Meredith, 1993; Meredith & Horn, 2001; Millsap & Meredith, 2007).
As in LC models, at least strong measurement invariance is required for a
meaningful interpretation of growth parameters. Otherwise, changes over
time would be confounded with potential changes in the properties of the
measurement instrument.
In first-order LGCMs discussed in the previous section, the assump-
tion of strong measurement invariance is implicitly made, but is not test-
able in these models, because there is only a single indicator at each time
point. One of the strengths of second-order LGCMs (curve-of-factors models;
Geiser et al., in press; Hancock, Kuo, & Lawrence, 2001; McArdle, 1988;
Sayer & Cumsille, 2001) is that these models use multiple indicators at
each time point, so that the assumption of measurement invariance can
be tested in the same way as, for example, in LS and LC models.
Figure 4.39 shows a second-order LGCM for anxiety measured at
four time points using two indicators (e.g., two different scales or parallel
test halves) at each time point. Note that the second-order model is based
on more than just one observed variable per time point. Furthermore,
in contrast to first-order LGCMs, the indicators in second-order LGCMs
do not directly measure growth factors, but are indirectly related to the
intercept and slope factors through LS factors. The LS factors represent the
first-order factors in this model, whereas the growth factors are specified
as second-order latent variables similar to the latent trait factor in LST
models. In fact, Geiser et al. (in press) showed that second-order LGCMs
can be formulated based on the fundamental concepts of LST theory.
The use of multiple indicators per time points makes it possible to
empirically test the assumption of measurement invariance by constrain-
ing factor loadings, intercepts, and residual variables to be equal across
time, as demonstrated for the LS model in Section 4.1.4. At least strong
factorial invariance (equal loadings and intercepts across time) should
hold for the growth parameters to be interpretable.
Another advantage of second-order LGCMs is that the use of multiple
indicators makes it possible to separate reliable occasion-specific vari-
ance from variance due to the growth factors and measurement error. In
second-order LGCMs, occasion-specific effects are explicitly represented
by the LS residual variables (indicated as z1–z4 in Figure 4.39), whereas
they are confounded with the error variables ek in first-order models. The
reliabilities of the indicators therefore tend to be underestimated in first-
order models (Geiser et al., in press). In the model shown in Figure 4.39,
an indicator-specific residual factor (IS2) was included to reflect stable
specific variance in the second indicator not shared with the first indica-
tor (cf. Section 4.1.3).
Structural Equation Models for Measuring Variability and Change 185
1 1 1 1 0 1 2 3
ζ1 ζ2 ζ3 ζ4
State 1 State 2 State 3 State 4
(anx1) (anx2) (anx3) (anx4)
λ1 λ2 λ1 λ2 λ1 λ2 λ1 λ2
IS2
FIGURE 4.39. Second-order LGCM for two observed variables Yik (i = indicator, k
= time point) measured on four occasions of measurement. eik = measurement error
variable; li = time-invariant first-order state factor loading; zk = LS residual variable.
The model contains an indicator-specific (method) factor (IS2) for the second indica-
tor. The parameters lIS2k indicate factor loadings on the indicator-specific factor. The
model assumes linear growth and equal spacing between time points. The change
process is modeled at the level of the LS variables via second-order intercept and
slope factors. The variable names used in Mplus appear in parentheses.
Figure 4.40 shows the Mplus input for the specification of a second-
order linear LGCM for the anxiety test halves. The variable names corre-
spond to the names shown in parentheses in Figure 4.39. The first num-
ber indicates the test half (i = 1, 2), whereas the second index indicates the
time point (k = 1, 2, 3, 4).
In practice, it is useful to first study the degree of measurement invari-
ance of the indicators using a simple LS model (cf. discussion in Section
4.1) before imposing a restrictive growth structure on the LS variables.
Given that the steps in testing measurement invariance were discussed
in detail for the LS model, we proceed directly to the analysis of LGCMs,
assuming strong measurement invariance for the indicators.
The first part of the model specification is identical to an LS model
with equal loadings:
anx1 by a11
a21 (1);
anx2 by a12
a22 (1);
anx3 by a13
a23 (1);
anx4 by a14
a24 (1);
In the next part, the intercepts pertaining to the first indicator are fixed to
zero to identify the means of the LS factors:
The intercepts pertaining to the second indicator are set equal across
time so that—in conjunction with the constraints placed on the LS factor
loadings—strong factorial invariance is maintained:
The second- order factor structure that involves the growth factors
interc and linear can be specified using the same syntax as for first-
order LGCMs. The only difference is that the indicators of the growth fac-
tors now are the LS factors rather than the observed variables:
FIGURE 4.40. Mplus input file for the specification of linear second-order LGCM
for anxiety measured by two indicators across four measurement occasions shown in
Figure 4.39. In this input file, the simplified syntax for LGCMs is used.
188 Data Analysis with Mplus
In the next step, we request the estimation of the intercept and slope fac-
tor means:
[interc linear];
Defined as a residual factor, this factor is not allowed to correlate with any
of the remaining latent variables in the model. Therefore, these correla-
tions have to be set to zero:
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
INTERC |
ANX1 1.000 0.000 999.000 999.000
ANX2 1.000 0.000 999.000 999.000
ANX3 1.000 0.000 999.000 999.000
ANX4 1.000 0.000 999.000 999.000
Structural Equation Models for Measuring Variability and Change 189
LINEAR |
ANX1 0.000 0.000 999.000 999.000
ANX2 1.000 0.000 999.000 999.000
ANX3 2.000 0.000 999.000 999.000
ANX4 3.000 0.000 999.000 999.000
ANX1 BY
A11 1.000 0.000 999.000 999.000
A21 0.987 0.018 55.720 0.000
ANX2 BY
A12 1.000 0.000 999.000 999.000
A22 0.987 0.018 55.720 0.000
ANX3 BY
A13 1.000 0.000 999.000 999.000
A23 0.987 0.018 55.720 0.000
ANX4 BY
A14 1.000 0.000 999.000 999.000
A24 0.987 0.018 55.720 0.000
IS2 BY
A21 1.000 0.000 999.000 999.000
A22 1.000 0.000 999.000 999.000
A23 1.000 0.000 999.000 999.000
A24 1.000 0.000 999.000 999.000
IS2 WITH
ANX1 0.000 0.000 999.000 999.000
ANX2 0.000 0.000 999.000 999.000
ANX3 0.000 0.000 999.000 999.000
ANX4 0.000 0.000 999.000 999.000
INTERC 0.000 0.000 999.000 999.000
LINEAR 0.000 0.000 999.000 999.000
LINEAR WITH
INTERC -0.008 0.003 -2.875 0.004
Means
INTERC 0.693 0.019 35.827 0.000
LINEAR -0.059 0.006 -10.593 0.000
Intercepts
A11 0.000 0.000 999.000 999.000
A21 0.028 0.012 2.263 0.024
A12 0.000 0.000 999.000 999.000
A22 0.028 0.012 2.263 0.024
A13 0.000 0.000 999.000 999.000
A23 0.028 0.012 2.263 0.024
A14 0.000 0.000 999.000 999.000
A24 0.028 0.012 2.263 0.024
190 Data Analysis with Mplus
Variances
IS2 0.009 0.001 7.336 0.000
INTERC 0.141 0.012 11.993 0.000
LINEAR 0.007 0.001 5.408 0.000
Residual Variances
A11 0.021 0.004 5.561 0.000
A21 0.025 0.004 6.720 0.000
A12 0.019 0.003 6.491 0.000
A22 0.023 0.003 7.844 0.000
A13 0.019 0.003 7.209 0.000
A23 0.017 0.003 6.718 0.000
A14 0.020 0.003 7.352 0.000
A24 0.016 0.003 6.003 0.000
ANX1 0.048 0.007 7.063 0.000
ANX2 0.031 0.004 7.922 0.000
ANX3 0.030 0.004 8.199 0.000
ANX4 0.017 0.005 3.339 0.001
Two-Tailed
Estimate S.E. Est./S.E. P-Value
INTERC |
ANX1 0.864 0.019 46.518 0.000
ANX2 0.932 0.021 43.447 0.000
ANX3 0.922 0.030 31.217 0.000
ANX4 0.909 0.039 23.222 0.000
LINEAR |
ANX1 0.000 0.000 999.000 999.000
ANX2 0.205 0.020 10.387 0.000
ANX3 0.405 0.038 10.669 0.000
ANX4 0.600 0.059 10.193 0.000
ANX1 BY
A11 0.949 0.010 99.540 0.000
A21 0.918 0.010 90.264 0.000
ANX2 BY
A12 0.947 0.009 110.162 0.000
A22 0.911 0.009 97.051 0.000
ANX3 BY
A13 0.947 0.008 125.147 0.000
A23 0.927 0.009 108.099 0.000
ANX4 BY
A14 0.946 0.008 123.294 0.000
A24 0.932 0.009 106.026 0.000
192 Data Analysis with Mplus
IS2 BY
A21 0.208 0.015 13.758 0.000
A22 0.223 0.016 13.928 0.000
A23 0.224 0.016 13.857 0.000
A24 0.223 0.016 13.532 0.000
IS2 WITH
ANX1 0.000 0.000 999.000 999.000
ANX2 0.000 0.000 999.000 999.000
ANX3 0.000 0.000 999.000 999.000
ANX4 0.000 0.000 999.000 999.000
INTERC 0.000 0.000 999.000 999.000
LINEAR 0.000 0.000 999.000 999.000
LINEAR WITH
INTERC -0.259 0.071 -3.659 0.000
Means
INTERC 1.844 0.092 20.014 0.000
LINEAR -0.720 0.095 -7.582 0.000
Intercepts
A11 0.000 0.000 999.000 999.000
A21 0.059 0.027 2.230 0.026
A12 0.000 0.000 999.000 999.000
A22 0.064 0.029 2.226 0.026
A13 0.000 0.000 999.000 999.000
A23 0.064 0.029 2.221 0.026
A14 0.000 0.000 999.000 999.000
A24 0.064 0.029 2.219 0.027
ANX1 0.000 0.000 999.000 999.000
ANX2 0.000 0.000 999.000 999.000
ANX3 0.000 0.000 999.000 999.000
ANX4 0.000 0.000 999.000 999.000
Variances
IS2 1.000 0.000 999.000 999.000
INTERC 1.000 0.000 999.000 999.000
LINEAR 1.000 0.000 999.000 999.000
Residual Variances
A11 0.099 0.018 5.460 0.000
A21 0.115 0.017 6.746 0.000
A12 0.104 0.016 6.363 0.000
A22 0.120 0.015 7.793 0.000
A13 0.103 0.014 7.176 0.000
A23 0.090 0.014 6.465 0.000
A14 0.106 0.015 7.310 0.000
A24 0.082 0.014 5.696 0.000
ANX1 0.254 0.032 7.910 0.000
ANX2 0.188 0.023 8.188 0.000
Structural Equation Models for Measuring Variability and Change 193
R-SQUARE
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
A11 0.901 0.018 49.770 0.000
A21 0.885 0.017 52.066 0.000
A12 0.896 0.016 55.081 0.000
A22 0.880 0.015 57.118 0.000
A13 0.897 0.014 62.574 0.000
A23 0.910 0.014 65.336 0.000
A14 0.894 0.015 61.647 0.000
A24 0.918 0.014 63.977 0.000
Latent Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
FIGURE 4.41. Mean model-implied growth curve for the linear second-order
LGCM for the anxiety data generated via the Mplus plot option.
5
more outcome variables at each time point (see Section 4.5). A third level
(e.g., individuals nested within organizations or school classes) can then
again be accounted for using the type = twolevel; option. There-
fore, a total of three levels can be modeled in Mplus if one of the levels is
defined by repeated observations.
The following common notation is used in this chapter to indicate the param-
eters estimated in multilevel regression models. Note that actual estimates
based on sample data are indicated with a hat (^) to make clear that they are
sample estimates of population parameters.
σ 2B
rIC =
( σ 2W + σ 2B )
Even very small ICC values of, say, .05 or .10 can lead to considerable
bias in the results of tests of statistical significance in conventional OLS
regression analyses and ANOVA, which do not take the dependence of
observations into account (e.g., Cohen et al., 2003, p. 538). In a two-
level basic analysis, Mplus outputs the ICC as well as the estimated
level-1 (within clusters) and level-2 (between clusters) variances for all
variables included in the analysis (i.e., for all variables listed under use-
variables).
The individual data shown in Figure 5.1 were exported to an ASCII
file named math_kft.dat for use in Mplus following the procedure
described in Section 2.1. The Mplus input file for the specification of the
twolevel basic analysis for the variables math and kft is shown in
Figure 5.2. The cluster variable class has to be specifically defined using
the subcommand cluster = class; to inform Mplus that the level-2
units (here the school classes) are defined based on this variable. The
cluster variable does not have to appear under the usevariables list,
because this variable is a variable with a special function and is neither
used as an independent nor a dependent variable in the model.
Below, the most important part of the Mplus output for the two-
level basic analysis is shown. Under SUMMARY OF ANALYSIS,
we can see that N = 503 students (level-1 units) and two variables (math
and kft) were used in the analysis. Mplus also informs us that the vari-
able class served as the cluster variable in this analysis, as intended. This
allows us to once again check that our level-2 units are correctly defined
(the level-2 variable stype is included in the analysis in Sections 5.4.3
and 5.5.2).
FIGURE 5.1. Screen shot of the sample data set math_kft.sav. The variable class
indicates the different school classes and is used as cluster variable in Mplus. The
variable math is the continuous dependent variable, kft is a continuous level-1 pre-
dictor variable, and stype is a dummy-coded dichotomous level-2 predictor variable.
FIGURE 5.2. Mplus input file for a twolevel basic analysis for the vari-
ables math and kft.
202 Data Analysis with Mplus
SUMMARY OF ANALYSIS
Number of groups 1
Number of observations 503
Continuous
MATH KFT
SUMMARY OF DATA
Number of clusters 34
5 22
8 18
9 19
10 16 21 12 25
11 20 3
12 8
13 14 2 5 23 13
14 4
15 17 24 1
16 6 28 30
17 11 27
18 29 26 35 36
19 33
Multilevel Regression Analysis 203
20 31
21 34
22 10
23 32
24 7
Mplus next outputs the estimated ICCs for both the kft and math
variables. The values of rIC = .44 for kft and rIC = .378 for math indicate
that class membership accounts for a substantial portion of individual
differences in both the math and KFT test scores. In other words, a non-
trivial amount of the variance in math and KFT scores is due to average
performance differences between the school classes. This also shows us
that in this example it makes sense to consider predictor variables at the
class level, because there is a considerable amount of variability at this
level. The inclusion of level-2 predictors would not make sense if there
were no differences between the classes.
Intraclass Intraclass
Variable Correlation Variable Correlation
Means
MATH KFT
________ ________
1 0.000 0.000
Covariances
MATH KFT
________ ________
MATH 26.484
KFT 52.327 235.735
Correlations
MATH KFT
________ ________
MATH 1.000
KFT 0.662 1.000
Means
MATH KFT
________ ________
1 11.930 66.716
Covariances
MATH KFT
________ ________
MATH 20.806
KFT 49.983 142.985
Correlations
MATH KFT
________ ________
MATH 1.000
KFT 0.916 1.000
The variance estimates can be used to calculate the estimated ICCs (this
is not really necessary, given that Mplus automatically outputs the rIC
values):
σˆ 2B 20.806
For math: rˆ IC = = = 0.44
σˆ W + σˆ B 26.484 + 20.806
2 2
σˆ 2B 142.985
For kft: rˆ IC = = = 0.378
σˆ W + σˆ B 235.735 + 142.985
2 2
Multilevel Regression Analysis 205
In addition, we see that the correlation between math and kft at the indi-
vidual level is r = .662, whereas the relationship at the level of the school
classes is much higher (r = .916). In summary, we find that the use of
a multilevel analysis to analyze the present data must be highly recom-
mended, given the large amount of variability between school classes. If
we used a conventional OLS regression analysis without taking the cluster
structure into account, we would have to expect a significant underesti-
mation of the standard errors of the regression coefficients and therefore
a substantial increase of our type-I error rate as well as biased confidence
intervals. We now consider different types of multilevel regression models
following the taxonomy of Luke (2004; see his Table 2.1). We begin with
so-called random coefficient regression models, the most basic of which is the
null or intercept-only model.
σ 2rij = σ 2W
Large values of σ 2rij indicate that there are substantial individual dif-
ferences in the math scores among students who belong to the same class.
From the twolevel basic analysis (see Section 5.3) we already know
that σˆ 2rij = σˆ 2W = 26.484 for the math test score. The same estimate will
also be obtained based on the intercept-only model. At level 2, variability
in the cluster means b0j is modeled. The cluster means are decomposed
into the grand mean across all school classes (g00) plus the deviations of
the cluster means from the grand mean (u0j):
The level-2 equation shows that in the intercept-only model, the cluster
means are variables that have a mean g00 and a variance σ 2u0 j, both of
which are parameters at level 2. This also explains the name random inter-
cept model: Whereas in conventional OLS regression analysis the intercept
is a constant (fixed effect), the intercept b0j in the random intercept model
can vary across clusters and therefore represents a random effect. In our
example, the math test means are not constant, but can vary across school
classes.
The variance σ 2u0 j characterizes differences in the cluster means
across level-2 units (between clusters):
σ 2u0 j = σ 2B
that is, the degree of differences between the school classes. Large val-
ues of σ 2u0 j indicate that there are large differences in the mean math
scores between school classes. From the twolevel basic analysis, we
know that σˆ 2u0 j = σˆ 2B = 20.806 for the math test score. A similar estimate
is also obtained in the intercept-only model. Because in this model the
variability in the math test means across classes is not explained by exter-
nal variables (the level-2 equation does not contain any predictor vari-
ables), the intercept-only model is sometimes referred to as unconditional
cell means model, unconstrained model, empty model, or null model.
In summary, we obtain Mplus estimates for the following three types
of parameters in the intercept-only model:
Multilevel Regression Analysis 207
• At level 1 (WITHIN):
—The
— variance of the individual scores around the cluster means,
σ 2rij = σ 2W
• At level 2 (BETWEEN):
—The
— grand mean g00
—The
— variance of the cluster means around the grand mean,
σ u0 j = σ 2B
2
Figure 5.3 shows the Mplus input file for the estimation of the intercept-
only model using the outcome variable math. It is only necessary to list
the variable math under usevariables, specify the relevant cluster
variable (here, class), and request analysis: type = twolevel.
A specific model statement is not required for the “null model.” After run-
ning the input file, the following Mplus warning message occurs in the
output:
FIGURE 5.3. Mplus input file for the specification of the intercept-only model for
the variable math.
208 Data Analysis with Mplus
This message occurs due to the fact that we did not specify any relations
between variables. This is in line with our null model, however, which
does not relate either the individual scores or the cluster means to any
external variables. Therefore, the warning message can be ignored in this
particular case. (Note that in other models, the same or similar messages
may indicate a potentially serious model misspecification and should
therefore not be generally ignored.)
The Mplus output provides us with the estimate of the level-1 vari-
ance under MODEL RESULTS—Within Level (σˆ 2rij = 26.480). Under
Between Level we obtain estimates of the grand mean (gˆ 00 = 11.933)
and the level-2 variance σˆ 2u0 j = 20.853. The estimated values are only
slightly different from the values that were estimated in the twolevel
basic analysis (see Section 5.3). In addition, the intercept-only model
also provides us with standard errors of the model parameter estimates as
well as test statistics (z-scores) and corresponding p-values.
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
Variances
MATH 26.480 2.450 10.809 0.000
Between Level
Means
MATH 11.933 0.808 14.776 0.000
Variances
MATH 20.853 6.017 3.466 0.001
The intercept-only model provides only the level-1 and level-2 variances
that can be used for the calculation of the ICC. The model does not con-
tain any additional independent variables (“empty model”). Therefore,
variability in the scores at the individual level or the group means is not
accounted for by other variables in this model. In the following, we dis-
cuss more complex multilevel models in which additional independent
variables are present (1) only at level 1 (so-called random effects analysis
of covariance [ANCOVA] model), (2) only at level 2 (so-called means-as-
outcomes models), or (3) at both level 1 and level 2 (so-called intercepts-and-
slopes-as-outcomes models).
The level-2 equations show that in this model, the intercept b0j from the
level-1 regression of math score on KFT score can vary across school
classes (level-2 units). In contrast, the level-1 slope coefficient b1j is
assumed to be constant across school classes. This can be seen from the
fact that there is no residual score in the level-2 equation for b1j, whereas
for the intercept b0j, a residual term u0j is included. Note that this model
is too restrictive if the slopes of the level-1 regressions actually do vary
across school classes, because no such variation is allowed in the model.
In Section 5.5 we consider models that allow for variation in both the
intercept and slope coefficients across clusters.
210 Data Analysis with Mplus
• At level 1 (WITHIN):
—The
— constant slope (fixed effect) b1j = g10
—The
— residual variance σ 2rij
• At level 2 (BETWEEN):
—The
— mean intercept g00 across clusters
—The
— variance σ 2u0 j of the intercepts across clusters
Before we consider the results for this model in detail, I introduce a con-
cept that is of great importance in multilevel analysis: that of centering
predictor variables. Centering was briefly discussed in the Section 3.2 on
OLS regression analysis (cf. Box 3.4). For simplicity, we return to simple
OLS regression analysis to revisit the issue of centering:
Y = b0 + b1 · X + e
Note that the estimated OLS regression coefficients do not take cluster-
ing into account and therefore should not be used. The negative intercept
term in this equation (bˆ 0 = –6.486) means that for students who obtained
a KFT score of 0, we would expect a math test score of –6.486. There are
two problems associated with this estimate for our interpretation of the
results. First, negative scores lie outside of the possible range of values for
the outcome variable math (the smallest possible score on the math test is
0 points). Second, a KFT raw score of 0 is very unlikely to be observed for
any student (and, in fact, was not observed in the present study). There-
fore, a substantively meaningful interpretation of the intercept is not pos-
sible in this example—at least not with uncentered variables. By centering
the independent variable kft before the analysis, the zero point of this
variable becomes more meaningful. Centering means that a meaningful
constant (usually a type of mean) is subtracted from each raw score. In
the analysis, the centered (or deviation) scores are then used instead of
the raw scores.
Multilevel Regression Analysis 211
where XijC indicates the centered variable. If the student, for example, has
a score of 0 on the grand-mean-centered KFT variable, this means that the
KFT score of this student is exactly identical to the average in the entire
sample. Negative values on the centered variable indicate that a student
scored below average on the KFT, whereas positive values indicate scores
above average in the sample.
In the OLS regression model, centering leads to a different intercept
estimate that typically has a more meaningful interpretation when inde-
pendent variables have no meaningful zero point or when the value of 0 is
meaningful, but did not occur in the data (cf. Section 3.2). The intercept
in the regression equation with a centered predictor variable provides the
expected value of the dependent variable for an individual with an average
value on the independent variable. The slope is unaffected by centering in
the simple regression model. The OLS regression equation for the math–
KFT example changes as follows when centered, instead of uncentered,
KFT scores are used:
We see that only the intercept, but not the slope of the regression equa-
tion, has changed. The new intercept is 12.533, which means that for
students with an average KFT score, we expect a math test score of 12.533
based on the linear regression of math on kft. Substantively, this repre-
sents a more meaningful result than the negative intercept estimate in the
uncentered regression equation.
In Mplus, centering of variables is performed by adding the single
additional subcommand centering in the variable command. This
means that the independent variables do not have to be manually cen-
tered prior to the analysis in Mplus. Mplus offers two options for centering
variables: (1) centering at the grand mean (grand mean centering) and (2)
centering at the cluster mean (group mean centering). Grand mean center-
ing can be requested in Mplus with the following specification:
within = kft;
In addition, we use grand mean centering for the variable kft, as demon-
strated above. In the model command, we specify the regression equa-
tion at level 1. Since level 1 is referred to as the within level in Mplus,
we have to include a command that makes clear that this equation refers
to this level:
FIGURE 5.4. Mplus input file for the specification of the one-way random effects
ANCOVA model for the outcome variable math and the level-1 predictor variable kft.
Multilevel Regression Analysis 213
model: %within%
math on kft;
Because in our model, the level-2 equations only contain a random inter-
cept and a fixed slope with no additional structure, nothing has to be
specified for the between level. Mplus automatically estimates the
parameters g00 and σ 2u0 j at level 2. (The fixed slope b1j = g10 is treated
as a level-1 parameter in Mplus.) The subcommand stdyx in the out-
put command is used to request the fully standardized solution. Under
MODEL RESULTS—Within Level Mplus provides the estimated
slope coefficient bˆ 1 j = gˆ 10 = 0.234 as well as the estimated level-1 residual
variance σˆ 2r = 14.883. Under Between Level, the mean intercept
ij
(gˆ 00 = 12.310) and the level-2 residual variance for the random intercept
(σˆ 2u0 j = 5.403) are given.
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
MATH ON
KFT 0.234 0.012 20.042 0.000
Residual Variances
MATH 14.883 1.052 14.141 0.000
Between Level
Means
MATH 12.310 0.437 28.142 0.000
Variances
MATH 5.403 1.612 3.352 0.001
Enders & Tofighi, 2007, for details.) More detailed information on the
size of this effect is provided under the standardized solution (STDYX
Standardization). We can see that the standardized level-1 regres-
sion coefficient for the regression of math on KFT was estimated to 0.773.
The estimated R 2 value for the level-1 regression is .598, which shows that
almost 60% of the variability in individual math test scores is accounted
for by individual and group differences in the KFT scores.
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
MATH ON
KFT 0.773 0.020 38.617 0.000
Residual Variances
MATH 0.402 0.031 12.987 0.000
Between Level
Means
MATH 5.296 0.763 6.938 0.000
Variances
MATH 1.000 0.000 999.000 999.000
R-SQUARE
Within Level
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
means through macro variables such as, for example, characteristics of the
teacher or the type of school students attend. The equations for a model
with no level-1 and one level-2 predictor, Wj, can be written as follows:
• At level 1 (WITHIN):
—The
— residual variance σ 2rij
• At level 2 (BETWEEN):
—The
— fixed intercept g00 for the regression of b0j on Wj
—The
— fixed slope g01 for the regression of b0j on Wj
—The
— residual variance σ 2u0 j
In our example, the coefficients g00 and g01 have a special meaning,
because our level-2 predictor school type is a dichotomous variable with
the categories 0 = standard high school and 1 = gymnasium. For this rea-
son, g00 gives the expected mean math score for the standard high school
group. The slope coefficient g01 represents the expected mean difference
216 Data Analysis with Mplus
in math scores (gymnasium minus high school). Note that the interpreta-
tion of the coefficients would change if we used a different coding scheme
for the categories of the variable school type (e.g., effect or contrast cod-
ing; for details see Chapter 8 in Cohen et al., 2003).
Figure 5.5 shows the Mplus input for the means-as-outcomes model
for math test scores. It is important to note that the variable stype has to
be defined as a level-2 only variable. This is accomplished through the fol-
lowing specification in the variable command:
between = stype;
Note that between level variables can vary only between, but not
within clusters. If variables defined as between level variables show
variation within any of the clusters, Mplus issues a warning message in
the output. Further note that it is not necessary to center the variable
stype, because the value of 0 is generally meaningful for a dummy-coded
variable (in our case, it refers to the group conventional high school). In the
model statement, we specify that the math test score (variable math) is
predicted by the variable school type at level 2:
FIGURE 5.5. Mplus input file for the specification of the means-as-outcomes
model for the outcome variable math and the level-2 predictor variable stype.
Multilevel Regression Analysis 217
model: %between%
math on stype;
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
Variances
MATH 26.467 2.450 10.802 0.000
Between Level
MATH ON
STYPE 10.641 1.188 8.957 0.000
Intercepts
MATH 10.003 0.441 22.671 0.000
Residual Variances
MATH 3.980 0.887 4.488 0.000
For level 1, Mplus estimates only the level-1 residual variance (σˆ 2rij =
26.467). For level 2, the slope coefficient gˆ 10 is estimated at 10.641. This
means that the expected mean difference in math test scores between the
group gymnasium and the group standard high school is 10.641. This
value is statistically significant (z = 8.957, p < .001). We can conclude that
the mean math score in group gymnasium is significantly larger than
in group standard high school. Hence, the variable stype accounts for a
significant portion of the differences seen in the mean math scores across
school classes.
The intercept term of the level-2 regression is estimated to be gˆ 00 =
10.223. This means that the expected average math test score in the group
standard high school is 10.003 points. The estimated mean for group
gymnasium can be calculated as the sum of the level-2 intercept plus the
level-2 slope coefficient: gˆ 00 + gˆ 01 = 10.003 + 10.641 = 20.644.
218 Data Analysis with Mplus
32
24
22
γˆ 00 + γˆ 01 = 20.644
20
Math Score
18
16 γˆ 01 = 10.641
14
12
10 γˆ 00 = 10.003
8
0
1 3 5 7 10 12 14 17 19 21 23 25 27 29 31 33 35
2 4 6 8 11 13 16 18 20 22 24 26 28 30 32 34 36
FIGURE 5.6. Error bar plot for illustrating the relationship between cluster-level
math performance and school type. The figure also illustrates the meaning of the
level-2 regression coefficients estimated in the means-as-outcomes model.
Multilevel Regression Analysis 219
STDYX Standardization
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
Variances
MATH 1.000 0.000 999.000 999.000
Between Level
MATH ON
STYPE 0.897 0.033 27.184 0.000
Intercepts
MATH 2.213 0.246 9.013 0.000
Residual Variances
MATH 0.195 0.059 3.287 0.001
R-SQUARE
Within Level
Between Level
Observed Two-Tailed
Variable Estimate S.E. Est./S.E. P-Value
32
30
28
26
24
22
20
Math Score
18
16
14
12
10
8
6
4
2
0
0 20 40 60 80 100 120
KFT Score
FIGURE 5.7. OLS regression of math score on KFT score by school class. Each line
refers to the regression in one of the 34 school classes. The thick dotted line repre-
sents the (uncentered) average OLS regression line across all school classes.
Multilevel Regression Analysis 221
slopes across clusters are explained by level-2 predictor variables have been
referred to as intercepts-and-slopes-as-outcomes models by Luke (2004).
The level-2 equations show that variability is now allowed for both the
intercept and the slope, since residual terms are now included in both
level-2 equations. Such a random coefficient regression model, without
any level-2 predictors, is useful, for example, when a researcher is exclu-
sively interested in level-1 predictor variables, but in addition wants to
account for the effects of clustered data. If, in such a case, the researcher
used a conventional OLS regression analysis without taking the clustering
of observations into account, he or she would have to expect bias in the
estimated standard errors of the regression coefficients and, consequently,
inaccurate tests of significance and confidence intervals. In the random
coefficient regression model, in contrast, the dependencies among obser-
vations are taken into account and the standard errors, test statistics,
p-values, and confidence intervals are more accurate.
Note that the only difference between the one-way random effects
ANCOVA model and the random coefficient regression model is that now
there is also a residual term in the level-2 equation for the level-1 slope
coefficient. For our example, this means that the regression lines for the
regressions of math scores on KFT scores can now have a different inter-
cept and a different slope in each school class, as illustrated in Figure 5.7.
Mplus provides the following parameter estimates for the random coef-
ficient regression model:
• At Level 1 (WITHIN):
—The
— residual variance σ 2rij
• At level 2 (BETWEEN):
—The
— mean intercept g00 across clusters
—The
— variance σ 2u0 j in the intercepts across clusters
222 Data Analysis with Mplus
—The
— mean slope coefficient g10 across clusters
—The
— variance σ 2u1 j in the slopes across clusters
—The
— covariance t01 between intercept and slope (needs to be
specifically requested in Mplus)
within = kft;
centering = grandmean (kft);
FIGURE 5.8. Mplus input file for the specification of the random coefficient regres-
sion model for the outcome variable math and the level-1 predictor variable kft.
Multilevel Regression Analysis 223
model: %within%
beta1j | math on kft;
In this specification, beta1j denotes the (arbitrary) label chosen for the
random slope coefficient. This line of command means that the slope of
the level-1 regression of math test scores on KFT scores is treated as a
latent variable that can vary across level-2 units. At level 2, we include a
command for the estimation of the covariance t01 between the random
intercept and the random slope, because this parameter is not estimated
by default in Mplus:
%between%
math with beta1j;
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
Residual Variances
MATH 14.862 1.091 13.627 0.000
Between Level
MATH WITH
BETA1J 0.050 0.032 1.572 0.116
Means
MATH 12.195 0.456 26.771 0.000
BETA1J 0.233 0.012 19.401 0.000
Variances
MATH 4.801 1.356 3.542 0.000
BETA1J 0.001 0.001 0.672 0.502
224 Data Analysis with Mplus
We can see that the results are very similar. Nonetheless, the random
coefficient regression model is the preferred method in this case because
it takes the dependencies among observations into account that arise due
to clustering, such that the standard errors are correctly estimated and
potential differences in the intercepts and slopes across classes are per-
mitted.
Variability in the intercepts and slopes can be analyzed by the vari-
ance estimates for the between level. Once again, we can see that there
is variability in the intercepts (σˆ 2u0 j = 4.801). It can be seen that the vari-
ance in the slopes (σ 2u1 j = 0.001) is rather small. Nonetheless, one should
not rule out the possibility that differences in the slopes between school
classes may be explained through level-2 predictor variables if such effects
are theoretically predicted. Predicting differences in the slopes across clus-
ters can be done by specifying so-called intercepts-and-slopes-as-outcomes
models, which are described in the next section. Finally, the estimated
covariance between intercept and slope is tˆ 01 = 0.050, which is statisti-
cally nonsignificant (z = 1.572, p = .116).
line. The parameter g10 gives the estimated slope in the group that is
coded zero (high school). The parameter g11 indicates the difference in
the slopes (gymnasium minus high school). If g11 = 0, there is no differ-
ence in the slopes between the school types. If g11 > 0, the slope in group
gymnasium is larger than the slope in group high school. If g11 < 0, the
slope in group high school is larger than the slope in group gymnasium.
The parameter g11 is often of particular interest because this coef-
ficient reflects a potential interaction effect between the level-1 predictor
(here, KFT score) and the level-2 predictor (here, school type). This is
often referred to as a cross-level interaction, and it means that the rela-
tionship between the level-1 predictor and the outcome varies depending
upon the values of the level-2 predictor. In our example, this would mean
that if a cross-level interaction is present, the regression of math on kft
would be different for different school types. A significant g11 coefficient
would indicate a significant cross-level interaction.
The residual terms u0j and u1j reflect differences in the intercept and
slope coefficients between school classes that cannot be explained through
the level-2 predictor school type. In summary, we obtain the following
estimates for the intercepts-and-slopes-as-outcomes model in Mplus:
• At level 1 (WITHIN):
—The
— residual variance σ 2rij
• At level 2 (BETWEEN):
—The
— fixed intercept g00 for the regression of the random inter-
cept b0j on Wj
—The
— fixed slope g01 for the regression of the random intercept
b0j on Wj
—The
— fixed intercept g10 for the regression of the random slope
b1j on Wj
—The
— fixed slope g11 for the regression of the random slope b1j on
Wj
—The
— residual variance σˆ 2u0 j for the random intercepts
—The
— residual variance σ 2u1 j for the random slopes
—The
— residual covariance t01 between the random intercept and
random slope (provided only upon request in Mplus)
The Mplus input file for the specification of the intercepts- and-
slopes-as-outcomes model is shown in Figure 5.9. Note that the variable
kft was again defined as a level-1 variable. Furthermore, the variable kft
Multilevel Regression Analysis 227
FIGURE 5.9. Mplus input file for the specification of the intercepts-and-slopes-
as-outcomes model for the outcome variable math, the level-1 predictor variable kft,
and the level-2 predictor variable stype.
model: %within%
beta1j | math on kft;
%between%
math on stype;
beta1j on stype;
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Within Level
Residual Variances
MATH 14.608 1.070 13.658 0.000
Between Level
BETA1J ON
STYPE -0.058 0.018 -3.139 0.002
MATH ON
STYPE 5.394 0.969 5.568 0.000
MATH WITH
BETA1J 0.035 0.022 1.567 0.117
Intercepts
MATH 11.536 0.396 29.149 0.000
BETA1J 0.226 0.013 17.258 0.000
Residual Variances
MATH 3.022 0.724 4.171 0.000
BETA1J 0.001 0.001 0.509 0.611
32
24
22
20
Math Score
γˆ 00 + γˆ 01 = 16.93
18
16
14 γˆ 01 = 5.394
12 γˆ 00 = 11.536
10
0
–50.00 –25.00 .00 25.00 50.00
minus high school). The value of gˆ 01 is 5.394, which is also statistically sig-
nificant (z = 5.568, p < .001). In summary, the level-2 predictor school type
explains differences in both the intercept and slope coefficients between
school classes. Given that the predictor variable kft has been centered at
the grand mean, the estimated value of 5.394 indicates that students with
an average KFT score (relative to the entire sample) are expected to have
5.394 points more on the math test if they belong to the group gymna-
sium than if they belong to the group high school. The difference in the
intercepts is also illustrated in Figure 5.10. The regression lines for the
two school types intersect the zero point of the grand-mean-centered kft
variable at different levels (the dotted line for the gymnasium cuts the zero
point above the solid line of the high school group).
Under Intercepts we first obtain the estimated intercept, gˆ 00 , for
the regression of the random intercept, b0j, on school type (gˆ 00 = 11.536).
This value tells us that the estimated intercept for the regression of math
performance on KFT score in group high school is 11.536 (cf. Figure
5.10). Therefore, for high school students with an average KFT value, a
math test score of 11.536 is expected. The sum of the coefficients gˆ 00 +
gˆ 01 equals the expected intercept for students in the group gymnasium:
11.536 + 5.394 = 16.93. Consequently, for students in the gymnasium
with an average KFT score, a math test score of 16.93 is expected.
In addition, the estimated intercept gˆ 10 for the regression of the
random slope, b1j, on school type is estimated (gˆ 10 = 0.226). This value
gives the estimated slope of the regression line for the regression of the
math score on the KFT score in group high school (gˆ 10; this coefficient is
approximately equal to the slope of the solid OLS regression line in Figure
5.10). Therefore, we can now also calculate the expected slope in group
gymnasium, which is equal to the sum of gˆ 10 + gˆ 11 = 0.226 + (–0.058)
= 0.168. This shows that the slope of the regression line is flatter for stu-
dents in the group gymnasium than in the group high school. At level
2, Mplus additionally estimated the residual variances for the intercept
(σˆ 2u0 j = 3.022) and the slope (σˆ 2u1 j = 0.001) as well as the residual covari-
ance between intercept and slope (tˆ 01 = 0.035).
Note that Mplus can also be used to analyze multilevel regression
models with more than one level-1 and/or more than one level-2 predic-
tor. In such extended models, some or all intercepts and slopes can be
specified as random coefficients. For predictors for which the slopes are
not supposed to be random coefficients, a separate on statement needs to
be included in the model command, in which the | sign as well as the
name of the random slope is omitted. Consider the following example:
Multilevel Regression Analysis 231
model: %within%
beta1J | y on x1;
y on x2;
nature. For a dichotomous variable, the LCA model for a single item can
be written as follows (e.g., see Collins & Lanza, 2010; Rost, 2004):
G
p( X vi = 1) = ∑ p g pig .
g =1
G
∑ pg = 1
g =1
The specification of LCA models in Mplus requires the module for the
analysis of mixture distribution models (option analysis: type =
mixture). Either the mixture or the combination add-on upgrades are
needed to analyze LCA models in Mplus in addition to the Mplus base
version. Box 6.3 introduces the data example used in this chapter.
Figure 6.1 shows the Mplus input file for an LCA model with three
latent classes. (My rationale for preferring a three-class solution will
become clear in later sections when we discuss model fit.) The data set
used here has the name computergames.dat. It contains the subject ID
number (variable id) as well as the responses on the eight computer
game questionnaire items (variable names: c1–c8). Using the command
variable: auxiliary = id; we tell Mplus that the variable id
236 Data Analysis with Mplus
We illustrate the analysis of an LCA in Mplus using the data set computer-
games.dat. This data set contains the item responses of N = 861 German
children, who answered a questionnaire on their computer game preferences.
In the following, an English translation of this questionnaire is shown:
is used as an auxiliary variable that does not play a role in the statistical
analysis as such, but that should still be included in a newly generated
data set containing the individual class assignment probabilities pro-
duced by Mplus on request (see description of the savedata command
on p. 239). Without the auxiliary command, the subject ID would not be
saved to the new data set. This could be a disadvantage when, for exam-
ple, LCA is used for individual diagnostic purposes. Specific individuals
in the data set could no longer be linked to the estimated class member-
ship information if there is no ID variable. The command categorical
Latent Class Analysis 237
FIGURE 6.1. Mplus input file for the specification of an LCA with three classes for
the eight computer game items.
the likelihood function to find the solution with the largest possible log
likelihood value. The log likelihood value is a measure of the probability of
the observed data given the model and is used as the basis for calculating
various fit statistics (see Section 6.3).
One complication for iterative estimation procedures that is particu-
larly relevant to LCA (and mixture distribution analyses in general) is that,
often, a single set of parameter starting values will not allow us to find the
solution with the best possible log likelihood value. The consequence may
be that the model estimation terminates at a so-called local likelihood maxi-
mum, which is often associated with inaccurate parameter estimates. Such
a local solution can, for example, be associated with incorrect fit statistics
and biased parameter estimates (e.g., see Uebersax, 2000). Local solutions
should therefore be avoided (cf. Box 6.4).
If only a few sets of starting values are tried in an LCA, there is a
relatively high chance of encountering a local likelihood maximum.
Moreover, studies indicate that local maxima occur frequently in empiri-
cal applications, even if only four or five classes are extracted (Uebersax,
2000). The use of a sufficient number of random sets of starting values
increases the chances that an LCA program will find the optimal solution
with the highest log likelihood value. In fact, for relatively simple, clearly
identified LCA models with few classes, the use of a large number of start
values usually guarantees that the true maximum will be found.
The Mplus default setting is that 10 sets of random starting values
are generated by the program and used in the model estimation. This
relatively small number of start value sets can be insufficient in practice,
especially for more complex models when more than two or three latent
classes are extracted. To be on the safe side, we encourage users to specify
a larger number of starting value sets. A relatively safe setting is to request
at least 500 sets of start values in the first step of the optimization (for
complex models, an even larger number may be necessary). The com-
mand analysis: starts = can be used to increase the number of
starting values in Mplus. The specification starts = 500 50; causes
Mplus to use 500 sets of start values in the first step of the optimization
(instead of only 10) and to pick the 50 starting value sets that show the
largest log likelihood values in the first step for the second step of the opti-
mization process, until a convergence criterion (in Mplus the default is a
parameter change of less than 0.000001) is reached. The solution with the
highest log likelihood value is then output by Mplus as the final solution.
In addition to increasing the number of start value sets, it is also
recommended to choose a sufficient number of iterations for the initial
stage of the optimization (Uebersax, 2000). In Mplus, the command
Latent Class Analysis 239
The resulting plot is shown below, where we discuss the Mplus output for
an LCA.
Based on the estimated LCA model parameters, the probability with
which each individual belongs to each of the extracted classes can be
estimated. On the basis of these individual class assignment probabilities,
one can estimate the most likely class membership for each individual for
diagnostic purposes. An individual is assigned to the class for which his
or her assignment probability is highest. In Mplus, the savedata com-
mand can be used to save individual class assignment probabilities and
the most likely latent class membership information to an external file. In
order to save these values to an external file, we specify:
Through this command, Mplus saves a new file with the name
computergames_3classes.dat to the same directory that also contains
the input file. The new file contains the eight computer game variables,
the ID variable id (because we defined this variable as an auxiliary vari-
able earlier under the variable command), the individual class assignment
probabilities (new variables named CPROB1, CPROB2, and CPROB3 to be
generated by Mplus in the analysis), and the estimated class membership
(new variable C to be generated by Mplus).
For an LCA, Mplus by default estimates the parameters using ML esti-
mation with robust standard errors. Conventional ML standard errors can
be requested using the command analysis: estimator = ml;.
In the following, the most important part of the Mplus output for the
three-class solution of the computer game example is shown. First of all,
we obtain the unconditional item response probabilities for each category
in the total sample. It can be seen that overall, the category never/rarely
was chosen more frequently in all items. Of interest to us is whether the
This message indicates that in the estimation process, one or more condi-
tional response probabilities were estimated and set to boundary values of
0 or 1 (which corresponds to threshold values of +15 and –15 in Mplus).
242 Data Analysis with Mplus
C1
Category 1 0.822
Category 2 0.178
C2
Category 1 0.727
Category 2 0.273
C3
Category 1 0.746
Category 2 0.254
C4
Category 1 0.816
Category 2 0.184
C5
Category 1 0.739
Category 2 0.261
C6
Category 1 0.746
Category 2 0.254
C7
Category 1 0.649
Category 2 0.351
C8
Category 1 0.674
Category 2 0.326
Next is a summary of the log likelihood values for the 50 sets of start
values that showed the largest log likelihood values after the initial step
of the optimization (ordered from the best to the worst log likelihood
values), together with the associated seeds (seed = random start number
for generating start values) and the number of the corresponding starting
value set. It can be seen that the same log likelihood value of –3463.814
was found for each of the 50 best sets of start values from the first step.
This is a good sign because it shows that the solution is replicable across
a variety of different start values. This indicates that local maxima are
apparently not an issue for this solution. Such a result is typical for well-
defined models with a relatively small number of classes.
By default Mplus informs the user of the issue of local maxima with the
following message, which is always printed under the list of log likelihood
values:
If the model converged properly, the best log likelihood value for the final
solution is output once again under MODEL FIT INFORMATION:
Loglikelihood
H0 Value -3463.814
This log likelihood value is identical to the best value that appears on the
very top of the previous listing of log likelihood values for different sets
Latent Class Analysis 245
The listing of the log likelihood values in Mplus can also be useful for repli-
cating the solution with the best log likelihood value in subsequent analyses
steps. This allows the user to reproduce the best solution without having
to generate a large number of start value sets again, which can be time-
consuming for complex models. This approach can be useful, for example,
when one wants to request additional output for a solution that has not been
requested initially. Using the so-called optseed option in Mplus, one can
replicate a specific solution on the basis of the specific seed number associ-
ated with this solution. In the current example, we could specify the follow-
ing in the input:
where the number 605358 refers to the seed number that pertains to the
starting value set with the number 321, which is the starting value set that is
ranked first in the list of log likelihood values:
This starting value set is among those sets that produced the best log likeli-
hood value of –3463.814. Note that in this example one could have chosen
any one of the 50 seed numbers, because all 50 start value sets resulted in
the same best log likelihood value. This, however, is not always the case,
especially for models with more classes. Using the optseed command as
specified above, Mplus will replicate the exact same solution that has been
found based on the seed 605358, although this should be verified in the
subsequent output file.
of start values. Following are different model fit statistics, which are dis-
cussed in detail in Section 6.3:
Information Criteria
Pearson Chi-Square
Value 386.883
Degrees of Freedom 229
P-Value 0.0000
Value 340.605
Degrees of Freedom 229
P-Value 0.0000
Next, Mplus provides information about the estimated sizes of the latent
classes in the model. The following table contains the class sizes in terms
of the estimated p̂ g parameters:
Latent
Classes
1 149.48776 0.17362
2 327.71918 0.38063
3 383.79305 0.44575
Latent
Classes
1 149.48771 0.17362
2 327.71923 0.38063
3 383.79307 0.44575
The entropy statistic is a summary measure for the quality of the clas-
sification. Values close to 1 indicate high classification accuracy, whereas
values close to 0 indicate low classification certainty.
CLASSIFICATION QUALITY
Entropy 0.731
Latent
Classes
1 152 0.17654
2 317 0.36818
3 392 0.45528
The class sizes that are estimated based on the manifest classification
deviate slightly from the estimated class portions p̂ g in our example. A
certain amount of discrepancy between these two kinds of class propor-
tion estimates is usually observed, because the manifest classification of
persons using the maximum class assignment probability is almost always
248 Data Analysis with Mplus
associated with some amount of estimation error. The reason is that each
person has a certain (nonzero) membership probability for each class,
which is not taken into account when people are assigned to the one latent
class for which the probability is highest. For this reason, the p̂ g param-
eters should generally be preferred as estimates of the latent class sizes.
Following are the average latent class assignment probabilities for
individuals assigned to each class. These are important indicators of the
quality of an LCA solution. Values close to 1 on the main diagonal of the
classification matrix indicate a high precision or reliability of the clas-
sification. High values on the main diagonal of this matrix mean that
individuals, on average, are classified with high certainty into their most
likely latent class. In contrast, moderate to low values indicate that there
is a high uncertainty of the class assignment, on average. According to
Rost (2006), the values on the main diagonal should be .8 or larger for a
good class solution. In our example, we obtained high class assignment
probabilities for Classes 2 and 3 (.913 and .922, respectively), whereas the
mean assignment probability for members of Class 1 is only .76, which
is relatively low. Apparently, there is somewhat more uncertainty with
regard to the assignment to Class 1 compared to Classes 2 and 3. Mem-
bers of Class 1 also have a relatively high average probability of belonging
to Class 2 (.146). Therefore, there is a certain degree of overlap between
these two classes. Nonetheless, the classification quality for Class 1 can
still be seen as satisfactory.
1 2 3
The more latent classes are extracted, the lower in general are the
mean class assignment probabilities. The reason is that with many
extracted classes, there are more possibilities for an incorrect classifica-
tion. The classes often are not as clearly separated any more.
Under MODEL RESULTS, Mplus first provides the parameters of the
LCA model in the logistic threshold parameterization. First are the esti-
mated threshold parameters that correspond to the conditional response
probabilities p̂ig . Under Categorical Latent Variables—Means,
Latent Class Analysis 249
Mplus outputs the threshold parameters that correspond to the class sizes
p̂ g . The conversion of these parameters into probabilities is explained in
Chapter 13 of the Mplus User’s Guide (Muthén & Muthén, 1998–2012).
MODEL RESULTS
Two-Tailed
Estimate S.E. Est./S.E. P-Value
Latent Class 1
Thresholds
C1$1 1.398 0.506 2.761 0.006
C2$1 1.920 0.706 2.717 0.007
C3$1 1.675 0.423 3.961 0.000
C4$1 1.489 0.523 2.850 0.004
C5$1 -0.981 0.406 -2.413 0.016
C6$1 -1.480 0.654 -2.263 0.024
C7$1 0.187 0.357 0.524 0.600
C8$1 0.685 0.334 2.052 0.040
Latent Class 2
Thresholds
C1$1 0.555 0.144 3.852 0.000
C2$1 -0.591 0.191 -3.101 0.002
C3$1 -0.097 0.163 -0.593 0.553
C4$1 0.463 0.143 3.244 0.001
C5$1 1.154 0.247 4.681 0.000
C6$1 1.542 0.291 5.290 0.000
C7$1 -0.628 0.152 -4.134 0.000
C8$1 -0.576 0.159 -3.633 0.000
Latent Class 3
Thresholds
C1$1 4.582 0.780 5.872 0.000
C2$1 4.334 0.804 5.388 0.000
C3$1 2.723 0.322 8.455 0.000
C4$1 4.583 0.927 4.947 0.000
C5$1 2.216 0.300 7.391 0.000
C6$1 2.166 0.366 5.924 0.000
C7$1 2.873 0.326 8.817 0.000
C8$1 2.845 0.410 6.937 0.000
Means
C#1 -0.943 0.247 -3.817 0.000
C#2 -0.158 0.151 -1.048 0.294
250 Data Analysis with Mplus
Latent Class 1
C1
Category 1 0.802 0.080 9.968 0.000
Category 2 0.198 0.080 2.463 0.014
C2
Category 1 0.872 0.079 11.068 0.000
Category 2 0.128 0.079 1.623 0.105
C3
Category 1 0.842 0.056 14.991 0.000
Category 2 0.158 0.056 2.808 0.005
C4
Category 1 0.816 0.078 10.399 0.000
Category 2 0.184 0.078 2.345 0.019
C5
Category 1 0.273 0.081 3.383 0.001
Category 2 0.727 0.081 9.020 0.000
C6
Category 1 0.185 0.099 1.878 0.060
Category 2 0.815 0.099 8.247 0.000
C7
Category 1 0.547 0.088 6.179 0.000
Category 2 0.453 0.088 5.125 0.000
C8
Category 1 0.665 0.074 8.941 0.000
Category 2 0.335 0.074 4.508 0.000
Latent Class 2
C1
Category 1 0.635 0.033 19.021 0.000
Category 2 0.365 0.033 10.915 0.000
Latent Class Analysis 251
C2
Category 1 0.356 0.044 8.152 0.000
Category 2 0.644 0.044 14.721 0.000
C3
Category 1 0.476 0.041 11.727 0.000
Category 2 0.524 0.041 12.916 0.000
C4
Category 1 0.614 0.034 18.144 0.000
Category 2 0.386 0.034 11.423 0.000
C5
Category 1 0.760 0.045 16.916 0.000
Category 2 0.240 0.045 5.332 0.000
C6
Category 1 0.824 0.042 19.466 0.000
Category 2 0.176 0.042 4.165 0.000
C7
Category 1 0.348 0.034 10.094 0.000
Category 2 0.652 0.034 18.917 0.000
C8
Category 1 0.360 0.037 9.847 0.000
Category 2 0.640 0.037 17.521 0.000
Latent Class 3
C1
Category 1 0.990 0.008 126.531 0.000
Category 2 0.010 0.008 1.295 0.195
C2
Category 1 0.987 0.010 96.062 0.000
Category 2 0.013 0.010 1.259 0.208
C3
Category 1 0.938 0.019 50.383 0.000
Category 2 0.062 0.019 3.309 0.001
C4
Category 1 0.990 0.009 106.680 0.000
Category 2 0.010 0.009 1.090 0.276
C5
Category 1 0.902 0.027 33.919 0.000
Category 2 0.098 0.027 3.699 0.000
C6
Category 1 0.897 0.034 26.596 0.000
Category 2 0.103 0.034 3.049 0.002
C7
Category 1 0.947 0.017 57.362 0.000
Category 2 0.053 0.017 3.242 0.001
C8
Category 1 0.945 0.021 44.386 0.000
Category 2 0.055 0.021 2.580 0.010
252 Data Analysis with Mplus
FIGURE 6.2. First step in viewing the item profile plot in Mplus.
Given that there are limited options to edit the Mplus line graph (e.g.,
it does not contain labels for the y- and x-axes), it is useful for a first
inspection of the class solution, but not suitable for inclusion in publica-
tions, presentations, etc. It is therefore usually a good idea to plot the
relevant conditional response probabilities again in an external graph-
ics program that allows for more nicely designed graphs. As an example,
Figure 6.5 shows the solution in a Microsoft Excel® line graph. In this
FIGURE 6.4. Mplus profile plot showing the conditional response probabilities for
the second category of the computer game items in each of the three latent classes.
The items are on the x-axis, whereas the y-axis shows the conditional probability of
endorsing the category “often/very often.”
0.9
Conditional response probability
0.8
for category "often/very often"
0.7
0.6
Class 1 (44.6%)
0.5 Class 2 (38.1%)
Class 3 (17.4%)
0.4
0.3
0.2
0.1
FIGURE 6.5. Microsoft Excel line graph for the same data shown in Figure 6.4.
Latent Class Analysis 255
example, the order of classes was changed from “largest first” to “smallest
last,” so that the smallest class in the Mplus solution is now Class 3 and
the largest class is now Class 1. In addition, the x-axis was labeled accord-
ing to the item content to facilitate the interpretation of the solution.
Using the class profile plot, we can once again see how the classes
differ from one another: The largest class (now Class 1) is the class of
nonplayers, in which members have very low probabilities for the cat-
egory “often/very often” across all items. The second largest class (Class
2) represents the group of children who play more often in general and
who have a specific preference for action, sport, and simulation games.
The smallest class (now Class 3) consists of logic and skill-training game
players.
In addition to the class profile plot, we have also requested that the
individual class assignment probabilities and estimated class member-
ship based on the maximum class assignment probability be saved to an
external file. At the end of the Mplus output file, we obtain a list of the
old and new variables that were saved to the new file by Mplus as well as
information on the format of the variables in the new data file. This infor-
mation is important, because it tells us the order in which the variables
(columns) appear in the new data file, so that the data can be processed
correctly in further analyses in external statistic programs such as, for
example, SPSS. As an example, note that the variable ID (subject ID) now
is no longer the first variable in the data set, but is now the ninth vari-
able:
SAVEDATA INFORMATION
C1 F10.3
C2 F10.3
C3 F10.3
C4 F10.3
C5 F10.3
C6 F10.3
C7 F10.3
C8 F10.3
ID F10.3
CPROB1 F10.3
CPROB2 F10.3
CPROB3 F10.3
C F10.3
256 Data Analysis with Mplus
Save file
computergames_3classes.dat
The four new variables that Mplus added to the data file are
CPROB1, CPROB2, CPROB3, and C. The newly created data file is named
computergames_3classes.dat (the name was specified in the input file)
and can be used for further analyses. For instance, one could use the class
membership information saved by Mplus to find out about the prefer-
ences of specific children in the data set. To which computer game class
is a specific child most likely to belong? This and similar questions can
be answered based on the individual class assignment probabilities that
Mplus saved for each child. Figure 6.6 shows an excerpt from that part of
the newly created data file (after it was read into SPSS) that contains the
four newly generated variables CPROB1, CPROB2, CPROB3, and C. An
SPSS syntax file for reading the data set generated by Mplus can be found
on the companion website.
The variables CPROB1, CPROB2, and CPROB3 indicate the esti-
mated probability of class membership for each child for Classes 1, 2, and
3, respectively. For example, based on his or her response pattern, the
first child in the data set (first row) has a probability of .218 of belong-
ing to Class 1, .028 of belonging to Class 2, and .754 of belonging to
Class 3. Note that the sum of the probabilities is 1. This is because LCA
assumes that each child belongs to one (and only one) of the extracted
latent classes. Given that this child’s probability is largest for Class 3, the
child is assigned to this class and therefore obtains the value of 3 on the
variable C, which indicates the most likely latent class membership for
each child.
The second child (second row) has the highest assignment probabil-
ity for Class 1 (.988) and is therefore assigned to Class 1, etc. In this
way, one can (1) examine how reliably each individual can be classified
and (2) obtain an estimate of the most likely latent class membership for
each individual based on the estimated class assignment probabilities.
The average latent class assignment probabilities for the most likely latent
class membership that Mplus reports in the standard output are calcu-
lated based on these individual class assignment probabilities.
Latent Class Analysis 257
Often, researchers do not have a strong theory that allows them to make
a firm assumption about the number of classes to extract. Given that the
number of classes is not a model parameter to be estimated in LCA, the
number of classes has to be determined indirectly through model tests
and/or model comparisons. In practice, this means that researchers usu-
ally estimate a series of LCA models with different numbers of classes
(e.g., the two- to five-class solutions) and then compare the absolute and
relative fit of these models. The model with the best model fit is usually
selected.
For the selection of the best fitting model, a number of different cri-
teria are available. Here I discuss only those criteria that are available
in Mplus. One can distinguish between indices for evaluating absolute
fit (how well does a model fit the observed data in absolute terms?) and
258 Data Analysis with Mplus
indices for evaluating relative fit (how well does a model fit compared to
other models?). More detailed information on indices for evaluating model
fit in LCA can be found in Collins, Fidler, Wugalter, and Long (1993),
Formann (2003), and Rost (2004).
Pearson Chi-Square
Value 386.883
Degrees of Freedom 229
P-Value 0.0000
Value 340.605
Degrees of Freedom 229
P-Value 0.0000
The values of the test statistics are relatively similar (LR = 386.883;
Pearson X2 = 340.605), so that we can suspect that the asymptotic con-
ditions for the two tests are approximately met. As a consequence, the
p-values can be interpreted, albeit with caution.
Both statistics show significant p-values in our example. Therefore,
the three-class solution would have to be rejected based on these sta-
tistics. However, neither the LR nor the Pearson X2 value is extremely
large relative to their degrees of freedom (df = 229). This indicates that the
degree of misfit of the model is not extremely large. In addition, we are
using a relatively large sample size (N = 861), so that there is considerable
statistical power to detect even relatively minor misspecifications. Addi-
tional model fit indices as well as model comparisons should be used to
obtain a more precise picture of the fit of the model.
Pearson X2 statistic and log likelihood value. Residual statistics also allow
us to detect “outliers,” that is, cases with unique or aberrant response
patterns. Residual statistics indicate the contribution of each observed
response pattern to the global Pearson X2 value and the log likelihood.
In addition, Mplus outputs standardized residuals (z-scores) for each pat-
tern. Large absolute z-values indicate response pattern frequencies that
are particularly under- or overestimated by the model. Residual statistics
are not part of the Mplus standard output, but can be requested using the
following line of code in the Mplus input file:
output: tech10;
To save time, we use the optseed function (see Box 6.6) to reestimate
the model with the additional request for the tech10 output. For this
purpose, we specify the seed that was associated with the best log likeli-
hood value in our previous analysis, so that Mplus automatically repro-
duces the best solution (without having to iterate through all 500 sets of
start values again):
The complete input file can be found on the companion website. We now
obtain additional output (TECHNICAL 10 OUTPUT) under the esti-
mated model parameters. Mplus first provides us with a detailed overview
of the observed response patterns. Each observed response pattern across
the eight computer game items is listed and enumerated by Mplus. In our
example, a total of 163 (of the total of 28 = 256 possible) response pat-
terns were actually observed. For example, the pattern with the number 1
(00100001) refers to the pattern in which the children selected the second
category (“often/very often”) only for the third (C3) and last item (C8). In
pattern number 2 (00000000), the category “often/very often” was not
selected for any of the eight items, and so on.
TECHNICAL 10 OUTPUT
MODEL FIT INFORMATION FOR THE LATENT CLASS INDICATOR MODEL PART
RESPONSE PATTERNS
each pattern to the overall Pearson X2 statistic and log likelihood value.
For our pattern number 1, we find relatively large contributions of 16.68
Pearson X2 statistic and 28.07 for the log likelihood. These values have
a similar interpretation as do modification indices in SEMs (cf. Box 3.7).
If, for example, pattern number 1 had not been observed, the overall
Pearson X2 value would decrease by 16.68. Residual statistics can provide
valuable information about outliers (e.g., “unscalable” cases) and other
peculiarities in the data. However, these statistics should never be used
to exclude cases with “inconvenient” response patterns in a purely data-
driven way.
At the end of the residual statistics table, Mplus also provides the over-
all contribution to the global Pearson X2 value that stems from model-
implied response patterns that were not observed in the actual data (so-
called “empty cells”):
p-value indicates that the model with G classes fits the data better than
the more parsimonious model with one class less. For a nonsignificant
p-value, the more parsimonious model with one class less is preferred.
In our example, we first test the three-class model against a two-class
solution. The bootstrap LR difference test procedure is requested in Mplus
by adding the following command to the input file for the three-class
model in Figure 6.1:
output: tech14;
TECHNICAL 14 OUTPUT
Random Starts Specification for the k-1 Class Model for Generated Data
Number of initial stage random starts 50
Number of final stage optimizations 20
Random Starts Specification for the k Class Model for Generated Data
Number of initial stage random starts 50
Number of final stage optimizations 20
Number of bootstrap draws requested 500
Latent Class Analysis 265
The log likelihood value for the G – 1 model (here, the two-class model
that is tested against the three-class model) is –3522.517 (it can be found
under H0 Loglikelihood Value). It is important to compare the log
likelihood value for the G – 1 model to the log likelihood value found in a
previous run of the G – 1 model that used a large number of start values
to make sure that the model estimated in the bootstrap procedure does
not represent a local maximum. The output for the separate analysis of the
two-class model is not shown here, but can be found on the companion
website. It turns out that the separate estimation of the two-class solution
with 500 sets of start values resulted in the same overall log likelihood
value, which indicates that we can trust the present bootstrap results.
Mplus also reports the LR difference value, which is equal to the dif-
ference in LR values between the two models. For the two-class model, the
LR value is 458.012 (see output on the companion website). For the three-
class model, we obtained an LR value of 340.605 (see Section 6.3.1.1).
Hence, the LR difference is 117.407. In the three-class model, nine addi-
tional parameters are estimated, compared to the two-class model (one
additional class size parameter and eight additional conditional response
probabilities). The LR difference of 117.407 is significant according to the
bootstrap p-value (Approximate P-Value), which is smaller than
.0001. Therefore, according to this criterion, the three-class model fits
significantly better than the two-class model.
For the comparison of three versus four classes, we additionally have
to estimate the four-class solution, including tech14 output (see com-
panion website for the detailed input specification). We obtain the follow-
ing result (only the most relevant output is shown):
We can see that the loglikelihood value for the three-class model
(–3463.814) has again been replicated exactly, so that we can trust the
266 Data Analysis with Mplus
bootstrap results. It turns out that the difference in the LR values between
the three- and the four-class solution is also significant, LR D = 54.239, df
= 9, p < .0001. According to this criterion, we would prefer the four-class
model over the three-class solution. However, the decision as to the num-
ber of classes to retain should not be based only on statistical, but also on
substantive criteria and interpretability of a solution. In our example, it
turns out that the three-class solution is easier to interpret than the four-
class solution (see Section 6.3.3).
TECHNICAL 11 OUTPUT
Value 115.508
P-Value 0.0038
versus four classes (using an extended input file for the four-class model
with tech11; see companion website) yields a result different from the
bootstrap LR difference test:
TECHNICAL 11 OUTPUT
Value 53.362
P-Value 0.0925
The p-values associated with both the VLMR test and its adjusted version
are nonsignificant (p = .0896 and p = .0925, respectively). This means
that according to these tests, the four-class model does not fit significantly
better than the more parsimonious three-class solution. (In contrast, the
bootstrap LR test found the four-class model to fit significantly better than
the three-class model.)
The results of a simulation study by Nylund, Asparouhov, and Muthén
(2007) indicate that, in general, the bootstrap LR test is a more accurate
indicator of the number of classes than the VLMR test. Therefore, the
bootstrap LR test should be preferred in practical applications.
Information Criteria
These values are not meaningful in an absolute sense. However, they can
be compared to the corresponding IC values calculated for other mod-
els. For example, we can compare them to the IC values obtained for the
two- and four-class solutions. The IC values for all three solutions are
shown in Table 6.1. It can be seen that the BIC index is smallest for three
classes, whereas AIC and aBIC favor four (or even more) classes. Based on
their simulation study, Nylund et al. (2007) recommend the BIC index for
determining the number of classes in LCA (in addition to the bootstrap
LR test; see Section 6.3.2.1).
6.3.3 Interpretability
For model fit evaluation and model comparisons in LCA, not only statis-
tical criteria play an important role. Equally important is the interpret-
ability of a solution (Collins & Lanza, 2010). This point can be illustrated
using the data sample discussed in this chapter. Even though the boot-
strap LR test indicated that the four-class solution fit the data better than
the three-class solution, close inspection of the latent class profiles in the
four-class solution reveals that one of the classes in the four-class solution
Meaning/ Chapters/
Command explanation Comments Section
271
272 Appendix A
Meaning/ Chapters/
Command explanation Comments Section
Example 2
(covariance
matrix): data:
type = cova;
variable: Specification of 6
categorical = variables in the
<list of ordered analysis that are
categorical variables ordered categorical
to be used in the (ordinal) or
analysis >; dichotomous
variable: Specification 6
nominal = <list of variables
of nominal variables as unordered
to be used in the categorical
analysis >; (nominal)
Meaning/ Chapters/
Command explanation Comments Section
variable: Definition of a 5
within = <name of variable that is
the level-1 variable(s) used only at level
in the analysis >; 1 of a multilevel
analysis
variable: Definition of a 5
between = variable that is
< name of the level-2 used only at level
variable(s) in the 2 of a multilevel
analysis>; analysis
variable: Centering of a 3; 5
centering = variable at the
grandmean × grand mean
(<variable name >);
variable: Centering of a 5
centering = variable at the
groupmean × cluster mean
(<variable name >); (group mean
centering)
274 Appendix A
Meaning/ Chapters/
Command explanation Comments Section
Meaning/ Chapters/
Command explanation Comments Section
Meaning/ Chapters/
Command explanation Comments Section
Meaning/ Chapters/
Command explanation Comments Section
Meaning/ Chapters/
Command explanation Comments Section
output: Model-implied
residual; covariance and
mean structure
and residual
statistics
output: Model
modindices; modification
indices
279
280 Appendix B
Further Readings
A .pdf version of the Mplus User’s Guide can be downloaded for free from the
Mplus homepage (http://www.statmodel.com).
Muthén, L. K., & Muthén, B. O. (1998–2007). Mplus user’s guide, fifth edition. Los
Angeles: Muthén & Muthén. http://www.statmodel.com/ugexcerpts.shtml.
The following book by Barbara Byrne introduces basic and advanced structural
equation modeling with Mplus:
Byrne, B. M. (2011). Structural equation modeling with Mplus: Basic concepts, appli-
cations, and programming. New York: Routledge.
Specific readings that deal with the specific statistical details of methods dis-
cussed in the book can be found at the beginning of each chapter.
283
References
Aiken, L. S., & West, S. G. (1991). Multiple regression: Testing and interpreting inter-
actions. Newbury Park, CA: Sage.
Baron, R. M., & Kenny, D. A. (1986). The moderator–mediator variable distinc-
tion in social psychological research: Conceptual, strategic, and statistical
considerations. Journal of Personality and Social Psychology, 51, 1173–1182.
Bollen, K. A. (1989). Structural equations with latent variables. New York: Wiley.
Bollen, K. A., & Curran, P. J. (2006). Latent curve models: A structural equation
perspective. New York: Wiley.
Bollen, K. A., & Long, S. (Eds.). (1993). Testing structural equation models. New-
bury Park, CA: Sage.
Bryk, A. S., & Raudenbush, S. W. (1987). Application of hierarchical linear mod-
els to assessing change. Psychological Bulletin, 101, 147–158.
Byrne, B. M., Shavelson, R. J., & Muthén, B. O. (1989). Testing for the equiva-
lence of factor covariance and mean structures: The issue of partial mea-
surement invariance. Psychological Bulletin, 105, 456–466.
Chen, F., Bollen, K. A., Paxton, P., Curran, P., & Kirby, J. (2001). Improper solu-
tions in structural equation models: Causes, consequences, and strategies.
Sociological Methods and Research, 29, 468–508.
Clogg, C. C. (1995). Latent class models: Recent developments and prospects
for the future. In G. Arminger, C. C. Clogg, & M. E. Sobel (Eds.), Handbook
of statistical modeling in the social sciences (pp. 311–359). New York: Plenum
Press.
Cohen, J., Cohen, P., West, S. G., & Aiken, L. S. (2003). Applied multiple regres-
sion/correlation analysis for the behavioral sciences. Mahwah, NJ: Erlbaum.
Cole, D. A., Martin, J. M., & Powers, B. (1997). A competency-based model of
child depression: A longitudinal study of peer, parent, teacher, and self-
evaluations. Journal of Child Psychology and Psychiatry and Allied Disciplines,
38, 505–514.
285
286 References
Cole, D. A., Martin, J. M., Powers, B., & Truglio, R. (1996). Modeling causal
relations between academic and social competence and depression: A
multitrait–multimethod longitudinal study of children. Journal of Abnormal
Psychology, 105, 258–270.
Cole, D. A., & Maxwell, S. E. (2003). Testing mediational models with longitu-
dinal data: Questions and tips in the use of structural equation modeling.
Journal of Abnormal Psychology, 112, 558–577.
Collins, L. M., Fidler, P. L., Wugalter, S. E., & Long, J. D. (1993). Goodness-of-fit
testing for latent class models. Multivariate Behavioral Research, 28, 375–389.
Collins, L. M., & Lanza, S. T. (2010). Latent class and latent transition analysis with
applications in the social, behavioral, and health sciences. New York: Wiley.
Crayen, C. (2010). Chi-square distributions calculator version 3 [Computer soft-
ware]. Berlin: Freie Universitat.
Duncan, T. E., Duncan, S. C., & Strycker, L. A. (2006). An introduction to latent
variable growth curve modeling: Concepts, issues, and applications (2nd ed.).
Mahwah, NJ: Erlbaum.
Eid, M. (2000). A multitrait–multimethod model with minimal assumptions.
Psychometrika, 65, 241–261.
Eid, M., Courvoisier, D. S., & Lischetzke, T. (2011). Structural equation model-
ing of ambulatory assessment data. In M. R. Mehl & T. S. Connor (Eds.),
Handbook of research methods for studying daily life (pp. 384–406). New York:
Guilford Press.
Eid, M., Gollwitzer, M., & Schmitt, M. (2010). Forschungsmethoden und Statistik
[Research methods and statistics]. Weinheim, Germany: Beltz.
Eid, M., Langeheine, R., & Diener, E. (2003). Comparing typological structures
across cultures by multigroup latent class analysis. Journal of Cross-Cultural
Psychology, 34, 195–210.
Eid, M., Lischetzke, T., & Nussbeck, F. W. (2006). Structural equation models
for multitrait–multimethod data. In M. Eid & E. Diener (Eds.), Handbook
of multimethod measurement in psychology (pp. 283–299). Washington, DC:
American Psychological Association.
Eid, M., Nussbeck, F. W., Geiser, C., Cole, D. A., Gollwitzer, M., & Lischetzke,
T. (2008). Structural equation modeling of multitrait–multimethod data:
Different models for different types of methods. Psychological Methods, 13,
230–253.
Eid, M., Schneider, C., & Schwenkmezger, P. (1999). Do you feel better or worse?:
The validity of perceived deviations of mood states from mood traits. Euro-
pean Journal of Personality, 13, 283–306.
Enders, C. K. (2010). Applied missing data analysis. New York: Guilford Press.
Enders, C. K., & Tofighi, D. (2007). Centering predictor variables in cross-
sectional multilevel models: A new look at an old issue. Psychological Meth-
ods, 12, 121–138.
Finney, S. J., & DiStefano, C. (2006). Non-normal and categorical data in struc-
tural equation modeling. In G. R. Hancock & R. O. Mueller (Eds.), Structural
References 287
Raudenbush, S. W., & Bryk, A. S. (2002). Hierarchical linear models (2nd ed.).
Thousand Oaks, CA: Sage.
Raykov, T. (1993). On estimating true change interrelationships with other vari-
ables. Quality and Quantity, 27, 353–370.
Raykov, T., & Marcoulides, G. A. (2006). A first course in structural equation model-
ing (2nd ed.). Mahwah, NJ: Erlbaum.
Reuter, T., Ziegelmann, J. P., Wiedemann, A. U., Geiser, C., Lippke, S., Schüz,
B., et al. (2010). Changes in intentions, planning, and self-efficacy predict
changes in behaviors: An application of latent true change modeling. Jour-
nal of Health Psychology, 15, 935–947.
Roosa, M. W., Liu, F., Torres, M., Gonzales, N., Knight, G., & Saenz, D. (2008).
Sampling and recruitment in studies of cultural influences on adjustment:
A case study with Mexican Americans. Journal of Family Psychology, 22,
293–302.
Rost, J. (2004). Lehrbuch Testtheorie–Teskonstruktion (2nd ed.) [Textbook test the-
ory and test construction]. Bern, Germany: Huber.
Rost, J. (2006). Latent-Class-Analyse [Latent class analysis]. In F. Petermann &
M. Eid (Eds.), Handbuch de Psycologischen Dianostik [Handbook of Psycho-
logical Assessment] (pp. 275–287). Göttingen, Germany: Hogrefe.
Rost, J., & Langeheine, R. (Eds.). (1997). Applications of latent trait and latent class
models in the social sciences. Munster, Germany: Waxmann.
Sayer, A. G., & Cumsille, P. E. (2001). Second-order latent growth models. In
L. M. Collins & A. G. Sayer (Eds.), New methods for the analysis of change
(pp. 177–200). Washington, DC: American Psychological Association.
Schafer, J. L., & Graham, J. W. (2002). Missing data: Our view of the state of the
art. Psychological Methods, 7, 147–177.
Schermelleh-Engel, K., Moosbrugger, H., & Müller, H. (2003). Evaluating the fit
of structural equation models: Test of significance and descriptive goodness-
of-fit measures. Methods of Psychological Research—Online, 8, 23–74. http://
www.dgps.de/fachgruppen/methoden/mpr-online.
Schumacker, R. E., & Lomax, R. G. (1996). A beginner’s guide to structural equation
modeling. Mahwah, NJ: Erlbaum.
Singer, J. D., & Willett, J. B. (2003). Applied longitudinal data analysis: Modeling
change and event occurrence. New York: Oxford University Press.
Snijders, T. A. B., & Bosker, R. J. (1999). Multilevel analysis: An introduction to basic
and advanced multilevel modeling. London: Sage.
Sobel, M. E. (1982). Asymptotic confidence intervals for indirect effects in struc-
tural equation models. In S. Leinhardt (Ed.), Sociological methodology 1982
(pp. 290–312). Washington, DC: American Sociological Association.
Sörbom, D. (1975). Detection of correlated errors in longitudinal data. British
Journal of Mathematical and Statistical Psychology, 28, 138–151.
Steyer, R. (1988). Experiment, Regression und Kausalitat: Die logische Struktur kau-
saler Regressionsmodelle [Experiment, regression, and causality: On the logi-
cal structure of causal regression models]. Unpublished habitation thesis,
University of Trier, Trier, Germany.
References 291
Steyer, R., Eid, M., & Schwenkmezger, P. (1997). Modeling true intraindividual
change: True change as a latent variable. Methods of Psychological Research—
Online, 2, 21–33. http://www.dgps.de/fachgruppen/methoden/mpr-online.
Steyer, R., Ferring, D., & Schmitt, M. J. (1992). States and traits in psychological
assessment. European Journal of Psychological Assessment, 8, 79–98.
Steyer, R., Partchev, I., & Shanahan, M. (2000). Modeling true intra-individual
change in structural equation models: The case of poverty and children’s
psychosocial adjustment. In T. D. Little, K. U. Schnabel, & J. Baumert
(Eds.), Modeling longitudinal and multiple-group data: Practical issues, applied
approaches, and specific examples (pp. 109–126). Hillsdale, NJ: Erlbaum.
Steyer, R., Schmitt, M. J., & Eid, M. (1999). Latent state–trait theory and research
in personality and individual differences. European Journal of Personality, 13,
389–408.
Uebersax, J. (2000). A brief study of local maximum solutions in latent class analysis.
Available at http://www.johnuebersax.com/stat/local.htm.
von Davier, M. (1997). Bootstrapping goodness-of-fit statistics for sparse categor-
ical data: Results of a Monte Carlo study. Methods of Psychological Research—
Online, 2, 29–48.
Ware, J. E., & Sherbourne, C. D. (1992). The MOS 36-item short-form health
survey (SF-36): I. Conceptual framework and item selection. Medical Care,
30, 473–483.
Whitelaw, N. A., & Liang, J. (1991). The structure of the OARS physical health
measures. Medical Care, 29, 332–347.
Widaman, K. F., & Reise, S. P. (1997). Exploring the measurement invariance
of psychological instruments: Applications in the substance use domain. In
K. J. Bryant, M. Windle, & S. G. West (Eds.), The science of prevention: Meth-
odological advances from alcohol and substance abuse research (pp. 281–324).
Washington, DC: American Psychological Association.
Author Index
G
C
Gaedicke, A.-K., 2
Chen, F., 59, 61 Gavin, M. B., 212
Clogg, C. C., 234 Geiser, C., 2, 26, 28, 32, 89, 91, 95, 166, 184,
Cohen, J., 28, 38, 64, 195, 196, 197, 216, 220 191, 231, 235, 236, 259
Cohen, P., 28 Goodman, L. A., 232
Cole, D. A., 88, 128 Graham, J. W., 19, 28
Collins, L. M., 234, 258, 268
Courvoisier, D. S., 81
Crayen, C., 45, 103 H
Cumsille, P. E., 184
Curran, P. J, 26, 59, 82, 163, 181 Hagenaars, J. A., 234–235
Hamagami, F., 145
Hancock, G. R., 184
D Hau, K. T., 64
Heller, K., 2
de Leeuw, J., 197, 212 Henry, N. W., 232
Diener, E., 235 Hertzog, C., 82, 126, 134
293
294 Author Index
K
N
Kaplan, D., 27
Keller, B.T., 166 Nesselroade, J. R., 82, 126, 134
Kenny, D. A., 63, 64 Noble, C. L., 88
Kirby, J., 59 Nussbeck, F. W., 26, 95
Kline, R. B., 27 Nylund, K., 267, 268
Kovacs, M., 128
Kreft, I. G. G., 197, 212
Kuo, W., 184 P
Schneider, C., 90 U
Schumacker, R. E., 27
Schwenkmezger, P., 26, 90 Uebersax, J., 238, 240
Scullen, S. E., 88
Shanahan, M., 82
Shavelson, R. J., 100 V
Sheets, V., 65
Sherbourne, C. D., 63 van de Pol, F., 258
Singer, J. D., 195, 197 von Davier, M., 258, 263
Snijders, T. A. B., 196, 197
Sobel, M. E., 70
Sörbom, D., 88 W
Steyer, R., 26, 81, 82, 85, 87, 116, 145, 148,
150, 159, 181 Ware, J. E., 63
Strycker, L. A., 26 Weinlader, H., 2
Wen, Z., 64
West, S. G., 28, 38, 64, 65
T Whitelaw, N. A., 62, 63, 73, 74
Widaman, K. F., 100
Tisak, J., 163 Willett, J. B, 195, 197
Tofighi, D., 212, 214 Williams, J, 65
Truglio, R., 128 Wugalter, S. E., 258
Subject Index
Note. Page numbers followed by b, t, and f indicate boxes, tables, and figures.
Commands and subcommands appear in Courier font.
296
Subject Index 297
CDC (Chi-Square Difference Calculator), 103, first-order, 51–52, 52f–54f, 54–58, 55t
104b–105b for longitudinal data, 82 (See also Latent
Center for Epidemiological Studies Depression state model)
Scale (CES-D), 83b second-order, 58–60, 59f–60f, 61b
Centering Convergence criterion, 240b
definition of, 210 Correlated uniqueness model, 88
grand mean, 211 Correlation matrix, 21–23, 22f, 23f
group mean, 211–212 Correlations
of independent variables, 38b between residual variables, 133
of predictor variables, 210 between state factors, 95
variable: centering =, 38b, 211 between state variables, 84
CES-D (Center for Epidemiological Studies between two factors, setting to zero, 52,
Depression Scale), 83b 54
CFA. See Confirmatory factor analysis Covariance matrix, 9, 21–23, 22f, 23f
CFI. See Comparative Fit Index Covariance residuals, 46b
Change models Covariances
functional form over time (See Latent coverage of data, 17–18
growth curve models) standardized, 55
goal of, 81 between two factors, setting to zero, 52,
LAM (See Latent autoregressive model) 54
latent (See Latent change model (latent Covariance structure models, 24
difference model)) Cross-lagged effects, 126–127, 133, 134,
Chi-Square Difference Calculator (CDC), 103, 144
104b–105b Cross-level interactions, 197
Chi-square difference test
automated using CDC software, 103,
104b–105b D
description of, 45b
Chi-square test data: file =, 271
description of, 45b data: listwise = on, 271
for independence model, 35 data: nobservations =, 272
latent path analysis, 76 data: type =, 271–272
for latent regression model, 44 data command, 12
latent state model, 87 Data set
manifest path model, 67 sample, 2
classes =, 237 unreadable in Mplus, 281–282
Class profile plot, 252–253, 253f .dat file format, 1, 7f
cluster = class, 200 Deletion, 2
Clustered data, 195–196 Dependent (endogenous) variables, 24, 26
Coding missing values, 2, 3f–6f, 4 Descriptive statistics
Coefficient of determination, 37 implausible or incorrect, 280
Command lines, 14b SPSS, 19, 21f
Comments, 12 Directional relationships (regressions), in path
Comparative Fit Index (CFI) diagrams, 26
description of, 45b–46b
latent state model, 88
one-factor and three-factor models, 55, E
55t
Conditional response or conditional solution EFA (exploratory factor analysis), vs. latent
probability, 234 class analysis, 233
Configural invariance, 100, 100b Empty cells, 263
Confirmatory factor analysis (CFA models) Empty model, 209. See also Null model
definition of, 25–26 (intercept-only model)
example of, 51b Endogenous (dependent) variables, 24, 26
298 Subject Index
Entropy, 269b G
Equality constraint, 99–100
Equality of latent means across time, testing General factor (g-factor model), 51, 52f–53f
for, 114b General factor model (g-factor model), 51,
Error messages, 279–280 52f–53f
missing semicolons, 12 G–1 model, 265
related to actual estimation of model, 34 Goodness of fit. See also model fit information,
WARNING in Model command, under specific models
30–31 in structural equation modeling, 45b–47b
warnings (See Warning messages) Grand mean centering, 38, 211
Error variables (residual variables), 27 Green font, for comments, 12
Exclamation mark, 14b Group mean centering, 211–212
Exogenous variables, 24, 26
Exploratory factor analysis (EFA), vs. latent
class analysis, 233 H
Exporting
coding missing values, 2, 3f–6f, 4 Heywood case, 59–61, 61b
Mplus graph, 179, 180f Hierarchical linear models. See Mplus,
SPSS data to Mplus, 1–2, 7, 7f–8f multilevel models
Hierarchical structure, of clustered data,
195–196
F
Factorial invariance I
strict, 100b, 114–115, 115f
strong, 100, 100b, 109–114 ICC (interclass correlation coefficient), 198,
weak, 100b, 102 200, 201f, 202–205
Factor loadings, 25 Implementation of constraints, checking, 105b
File formats, 1–2, 7f Importing, into Mplus, 7, 8f, 10
FIML estimation, 15–16, 19, 20f Incremental fit indices, 45b
First-order latent growth curve models, Independence model (null model), 23, 35, 45b
164–183 Independent variables, 38b, 196
example of, 165 Indicator-specific effects, 88–89
external variables, influence of, 167 Indicator-specific factor or method, 89–92,
intercept factor mean, 166–167 89f–93f, 102
linear Indicator specificity, 85–86
graphical analysis of growth curves, 174, Indirect effects, 63, 65b, 70–71
175f–179f, 176–177, 179 Individual data format, 9–10
intercept-only model, 170, 170f Information Criteria (IC), 46b, 267–268. See
model fit information, 170–171 also AIC; BIC
MODEL RESULTS, 171–172 *.inp, 14
path diagram, 164–166, 165f INPUT READING TERMINATED
plot option, 169–170, 171, 174 NORMALLY, 32–33
slope factor, 166 Intercept-only model (null model), 23,
specification, 167–168, 168f, 169b, 170f 205–209, 208b
STANDARDIZED MODEL RESULTS, Intercepts, 230
172–174 Intercepts-and-slopes-as-outcomes model,
quadratic, 179, 180f–181f, 181–183 221, 224–231
exporting Mplus graph, 179, 180f cross-level interaction effect, 228–229, 229f
MODEL RESULTS, 182–183 output, 228–231, 229f
path diagram, 180f parameters, 225–226
specification, 181–182, 181f specification of, 226–228, 227f
vs. second-order LGCMs, 164, 184 Interclass correlation coefficient (ICC), 198,
Formative measurement models, 25 200, 201f, 202–205
Subject Index 299
305