Handout StructuralEquationModelling
Handout StructuralEquationModelling
Page 1
24/05/00
Structural Equation Modelling (4) When cause is absent effect is absent: when I dont talk about causality no boredom is observed. (5) The manipulation of cause leads to an associated change in effect. So, if we manipulated whether someone is listening to me talking about causality or to my cat is mewing, the effect elicited should change according to the manipulation. This final manipulation serves to rule out external variables that might affect the cause-effect relationship. Historically, Psychology has two traditions: (a) experimental (variables are manipulated systematically to infer causality), and (b) situations are observed and variables correlated (see Field, 2000, Chapter 7; Cronbach, 1957). The first approach allows inferences about causality because possible causes are being compared systematically (i.e. a situation in which cause is observed is compared against a situation in which cause is absent). In the later approach no such manipulation is made, therefore cause cannot be inferred. For example, we can observe that night always follows day, but this doesnt necessarily mean that day causes night (Hume). So, in situations in which cause cannot be manipulated we cannot make causal attributions about our variables. Statistically speaking, this means that when we analyze data from non-experimental situations we cannot conclude anything about cause an effect. Structural Equation Modeling (SEM) is an attempt to provide a flexible framework within which causal models can be built.
A Simple SEM
SEM is an attempt to model causal relations between variables by including all variables that are known to have some involvement in the process of interest. As a simple example, we could test the effect of a drug on some psychological disorder (e.g. obsessive compulsive disorder, OCD). Obsessive checking behaviour sometimes characterizes this disorder. So, we could take a baseline of checking behaviour, administer the drug and then re-measure checking to ascertain the reduction in this behaviour. We might find a pretty strong relation (correlation) between drug dose and reduction in checking across patients. However, there are several problems. First, this is a non-directional relationship because although drug dose could cause the reduction in checking behaviour, the drug dose could be caused by the behaviour (because higher doses are likely to be given to the more severe cases). Second, drug dose might not have a direct effect on behaviour; instead, the drug might inhibit some cognitive process (such as worrying that a task hasnt been done) that in turn prevents behaviour. So, although the effect might be causal it is part of a chain of events. Finally, the drug might have no effect and both drug dose and behaviour are moderated by a third variable. For example, the trait of restraint might (a) reduce the number of times that a obsessive thoughts lead to an actually behaviour, and (b) affect the amount of symptoms reported to the psychiatrist (and hence affects the amount of drug prescribed. This example should hopefully illustrate several points about building a causal pathway model about the effect of a drug on behaviour. 1. Third Variables: In non-experimental research, all variables that may cause a spurious relationship between the hypothesized cause and effect must be collected and included in the data. In this example, restraint might cause a spurious relationship between the drug and checking behaviour. Therefore, we need to collect data about patients levels of cognitive restraint and include this variable in the model. This is similar to what we do with partial correlations (see Field, 2000, Chapter3). We collect information about possible third variables and include them in the model. 2. Disturbance Terms: There may be variables that have an influence over either the cause of effect variables, but not both. For example, worry (as a trait) may affect checking behaviour
Page 2
24/05/00
Structural Equation Modelling but will not affect the dose of drug. Similarly, the psychiatrist prescribing the dose is likely to directly affect the dose (because doctors will vary in their preferred dose for a given level of symptoms) but not the outcome (the checking behaviour). These variables need not be included in the model because they affect only the cause or effect, but not both. These are known as disturbance terms and are sometimes notated by the letter i for the ith variable.
Model Specification
The first step in SEM is to specify a model. A model is simply a statement (or set of statements) about the relations between variables. Although you may not realise it you will have already specified many models during your statistics course. For example, a correlation implies a non-directional single relation between two variables, whereas ANOVA and Regression imply directional relations between variables (however, neither technique can statistically test the directionality of these relations). More complex models (in ANOVA for example) involve the use of planned comparisons to specify particular relations between levels of a variable. In all of these techniques, an implicit model underlies the analysis. In SEM, the model is much more central to the process, so rather than relying on a fixed framework based on an implicit model (such as Regression), in SEM we specify a model first, and then use SEM to test the model specified. Specification involves several processes. First parameters need to be specified. Parameters are a set of values (constants) that indicate the nature of the relation between variables. So, these are similar to correlation coefficients, or factor loadings in which the magnitude and direction of the parameter indicate something about the strength and direction of the relationship. There are two types of parameter: Fixed Parameters: These are parameters not estimated from the data collected. Typically, their value is fixed at zero. These parameters indicate no relation between variables. Free Parameters: These are parameters that are estimated from the data collected and are believed, by the investigator, to be different from zero. The parameters indicate the presence of a relationship between variables.
Variables: There are two types of variables that are used in SEM. The first type is measured variables, which are variables that can be directly measured in some way. So, these variables consist of data that have actually been collected that map directly to the construct of interest (we could measure heart rate directly though a pulse monitor and the data collected maps directly onto the variable of interest). A second type of variable is a latent variable, which is a variable that cannot be measured directly but is implied by the covariances between two or more indicator variables. A latent variable is, therefore, just another name for a factor (in factor analysis) or component (in principal components). So, it is just a construct that cannot be measured directly (e.g. depression) but can be accessed through measuring other related variables (such as questions on the BDI). Latent variables are free of the uniqueness and measurement error associated with the indicators. Within a model, measured variables can either be included as indicators of a latent variable, or as stand-alone variables. Model Components: Within a model there are two components: the measurement model and the structural model. The measurement model is basically the part of the general model in which latent variables are prescribed. Therefore, we determine which measured variables are indicators of a latent variable (or factor). The structural model is the part of the model in which we define the relationship between latent variables and other measured variables that are not indicators of some other latent variable. These two parts of the model are combined to make a whole model that comprehensively describes the relationships between variables that are free of measurement error (i.e. latent variables).
Page 3
24/05/00
Structural Equation Modelling Types of Relations between Variables: So far we have learnt about the different types of variables we can use in SEM, and the basic types of models that we can use. The next important feature is how we specify relations between types of variables. There are basically three types of relations that can be specified, two of which you should have come across during basic statistics courses. 1. Association: This is a nondirectional relation between variables. This is what is measured by the correlation coefficient with which you should not be familiar. The parameters representing these effects are simply the covariances between variables (or the correlation coefficient if variables are standardized). 2. Direct Effect: This is a directional relation between two variables. You have come across this sort of relation in ANOVA and regression in which one variable predicts another. Just like multiple regression and factorial ANOVA it is possible to have a dependent variable related to several independent variables, and as in MANOVA and discriminant analysis it is possible to have one independent variable related to several dependent variables. The parameters representing these effects are values of the regression coefficients (i.e. weights in the regression equation (see Field, 2000, Chapter4). 3. Indirect Effect: This is an indirect relation between variables and stems from the possibility that a variable can be both an independent and a dependent variable within SEM. In simple terms, a given variable can be the cause of some change in another variable whilst itself is influenced by some different variable. As a simple example, if we know that practice makes perfect (i.e. practise leads to a perfect performance) and we know that the trait dedication (a latent variable) leads to more practise, then we know that dedication must have an indirect effect of perfection. In this example, practise is both an independent variable (because variation in practise leads to variation in perfect performance) and a dependent variable (variation in practise is affected by variation in dedication). If the direct and indirect effects of an independent variable are summed, we gain the total effect of that variable. Variables that receive a directional influence from some other variable in the system are known as endogenous variables. These variables may also exert their own directional influence on other variables in the system. Variables that do not receive any directional influence from another variable are known as exogenous variables. Exogenous variables are usually associated with each other by nondirectional relations (although it is possible that they are not associated with any other variables in the model. The variance in endogenous variables is not completely explained by other variables in the system. Instead it is assumed that they are also influenced by an error term which consists of both random error and systematic variable that could, for example, represent the effect of some other variable that has not been measured or included in the present model.
Additional Considerations
In SEM we are hoping to model causal relationships between variables. To do this several conditions should be met. 1. The cause and effect variables of interest must be included. 2. All variables that affect both the cause and effect variables must be included. Therefore, any variable that could produce a spurious relation between the cause and effect variable must be considered within the model. 3. The direct effects between variables must be included to indicate the causal relations of interest.
Page 4
24/05/00
Structural Equation Modelling 4. Relations between exogenous variables (non-direct effects) must be included. 5. Disturbance terms should be included that represent any variables that need not be explicitly represented in the model.
X1
1
X2
2
X3
3
Restraint (1)
2 4 1 6
Drug Dose
3
1.0
OCD (1)
5
Actions (Y2) Thoughts (Y1)
1.0
1.0
Figure 1.
Page 5
24/05/00
Structural Equation Modelling This diagram looks very complex even though it describes a relatively simply set of relations between variables. Primarily we are interested in the directional relation between drug dose and OCD. Drug dose can be directly measured and so is a measured variable (hence represented by a rectangle). We are interested in the effect that drug dose has on OCD. Now, in this model OCD is a latent variable in that it cannot be measured directly. To keep things simple we assume that we can measure OCD symptoms in terms of the frequency of obsessive thoughts and the frequency of obsession-related actions. As such, OCD is represented by a circle (because it is a latent variable) that is measured by two measured variables (thoughts and actions) both represented by rectangles. The latent variable representing OCD has directional arrows coming into it and so is said to be endogenous (i.e. dependent on other variables within the model). The final variable in the model is the trait of restraint, which cannot be measured directly. This latent variable is represented by an ellipse and is measured through three measured variables (X1, X2, and X3). Restraint has a direct effect on both drug dose and OCD symptoms and so is connected to them both with a single arrow. Restraint does not have any directional arrows coming into it and so is said to be exogenous (i.e. doesnt depend on other variables within the model). All of the measured variables (the squares) have associated error terms, but the latent variables do not (because they are assumed to be free from measurement error). However, both OCD and drug dose have disturbance terms attached to them. In the case of OCD this might be some kind of comorbidity factor such as worry or anxiety (that affect only OCD symptoms but not drug dose). For Drug dose this disturbance term might be the psychiatrist (because different psychiatrists will have different dose preferences for a given level of reported symptoms). The letters in the path diagram follow usual conventions, which are listed as follows (from Pedhazer & Schmelkin, 1991): X measured variable that indicates an exogenous latent variable.
error term associated with Y. Exogenous latent variable. Endogenous latent variable. Parameter of a directional relation between a latent variable and its indicators. Parameter of a directional relation between an endogenous and exogenous latent variable. Parameter of a directional relation between two endogenous latent variables. Parameter of a non-directional relation between two latent variables (i.e. the covariance). disturbance terms (or error in an endogenous latent variable).
Page 6
24/05/00
Structural Equation Modelling 1. Fix the variance of each latent variable: Fixing the variance of each latent variable has the effect of standardizing these variables. Typically, the variance of each latent variable is set at 1 so that any resulting parameter estimates can be interpreted like standardized regression coefficients (directional relations) or correlations (non-directional relations). If measured variables are also standardized (converted to Z-scores) then relations between latent variables and measured variables can also be interpreted in this way. 2. Fix the value of one directional relation emitted by a given latent variable: In effect this means setting the value of the relation for each error term in the model at 1.0. So, we assign a value of 1 to the influence of each error term on its associated endogenous variable. Remember that error terms are typically exogenous and so their variances are estimated from the model. By fixing the value of their influence on an endogenous variable, it is possible to estimate this error variance, which in turn tells us how much variance in the endogenous variable cannot be explained other variables within the model. Youll notice in the path diagram that the influence of all error terms has been set at 1. Also, although not explicitly drawn on the diagram, the variance of restraint would also be set at 1. The question remains of how to establish a scale for endogenous latent variables because the variance of these variables is not estimated by the model but implied by other variables within the model. One way is to fix the value of one influence of an endogenous latent variable on another variable at 1. Typically, we fix the value of the influence of the latent variable on one of its indicators. An alternative is to set the implied variance of these variables at 1.0, which has the desirable effect of simplifying the interpretation of the resulting parameter estimates. However, current software typically uses the former approach.
Estimation
Once the model is constructed estimates of the free parameters must be obtained from a set of observed data. The simplest way to get these estimates is to use simple regression techniques to estimate the values of relations between variables in the model. However, computer software packages use iterative procedures in which a series of solutions are generated and then compared to the actual data. Typically, the process involves generating a series of estimates of the free parameters that imply a covariance matrix like the observed one. The implied covariance matrix is the covariance matrix that would result if the values of the fixed parameters and current estimates of the free parameters were substitutive into the structural equations that make up the defined model. So, the program begins with a set of start values that are in some cases estimated from the data (LISREL) or can be a set of default values (EQS), and the implied covariance matrix is estimated and compared to the actual covariance matrix. The differences between these matrices are stored in a residual matrix. This matrix represents the error between the model and the real data. Estimates are changed to try to minimize the values in the residual matrix (so that values in the implied covariance matrix and actual covariance matrix converge). When the residuals cannot be reduced any further the iterations finish. The residuals are expressed as a single value that represents the degree of correspondence between the implied covariance matrix and the actual covariance matrix. This value is sometimes called the value of the fitting function and in a perfect model would be zero (indicating no difference between the implied and actual covariance matrices).
Goodness-of-Fit
As with even the simplest models, it is essential to establish how well the model fits the observed data (see Field, 2000, Chapter 1). The simplest gauge of how well the model fits the data would be to
Page 7
24/05/00
Structural Equation Modelling inspect the residual matrix. A good model should have a high number of residuals close to zero (implying high correspondence between elements of the implied covariance matrix and the actual covariance matrix). The fitting function already provides a summary of the information in this matrix and so, in itself, is a measure of how well the model fits the data. Indeed, if the fitting function is multiplied by the sample size minus 1 we get statistic that is distributed with a 2 distribution if the data are multivariate normal and the specified model is the correct one (the 2 goodness-of-fit test). This later assumption is unlikely to be true which has led to considerable debate about the appropriateness of this test. Alternatives are the adjunct fit indexes pioneered by Bentler and Bonett (1980). These work on the principle of comparing the fitted model against a null model (and so are rather like hypothesis testing). The null model (sometimes called the independence model) used is one in which no relations between variable are specified and instead only variable variances are estimated. As such these statistics represent the improvement in fit of the actual model over a model in which all structural parameters are set at zero. Unlike the goodness-of-fit test, these indexes cannot be statistically tested and instead are treated as global indexes of model adequacy. Bentler and Bonett indexes vary between zero and 1 and an acceptable critical value is 0.9. Indexes less than this value reflect inadequate models. However, not all adjunct fit indexes follow these rules. The goodness-of-fit and adjunct fit indexes are in many ways opposite. The goodness-of-fit gives us an estimate of the error in the model and as such large values represent large amounts of error. Therefore, it is more accurate to call these measures badness-of-fit indexes because large values represent a bad fit. When we put these values to statistical tests we, are, therefore looking fir a null result (i.e. the 2 is not significantly different from zero). The adjunct fit index represents the goodness-of-fit more literally in that high values demonstrate a better fit, however, they cannot be put to a statistical test.
Page 8
24/05/00
Structural Equation Modelling true. The estimates obtained are no different from those obtained from regression, ANOVA or a simple correlation. They tell us nothing about causality. The difference between SEM and these other techniques is in the flexibility with which causal models can be built. In ANOVA models, causality is inferred because one variable is systematically manipulated to see the effect on another. Likewise, in SEM, causality can be inferred, but only from the model originally constructed (and not from the statistical test of that model). The benefit of SEM over other approaches such as ANOVA or regression is simply the flexibility with which models can be built.
Further Reading
Hoyle, R. H. (ed.)(1995). Structural Equation Modelling: concepts, issues and applications. Sage: Thousand Oaks, CA. Pedhazur, E. & Schmelkin, L. (1991). Measurement, design and analysis. Hillsdale, NJ: Erlbaum. Chapters 23 & 24. Saris, W. E. (1999). Structural Equation Modelling. In H. Adr & G. J. Mellenbergh, (eds.), Research methodology in the social, behavioural and life sciences (pp. 220-239). London: Sage.
References
Bentler, P. M., & Bonett, D. G. (1980). Significance tests and goodness-of-fit in the analysis of covariance structures. Psychological Bulletin, 88, 588-606. Cronbach, L. J. (1957). The two disciplines of scientific psychology. The American Psychologist, 12, 671 684. Field, A. P. (2000). Discovering statistics using SPSS for windows: advanced techniques for the beginner. London: Sage.
Page 9
24/05/00