Factor
Factor
reduction
William Revelle
Department of Psychology
Northwestern University
June 1, 2019
Contents
1 Overview of this and related documents 4
1.1 Jump starting the psych package–a guide for the impatient . . . . . . . . . 4
3 Getting started 7
1
5.2 Confidence intervals using bootstrapping techniques . . . . . . . . . . . . . 31
5.3 Comparing factor/component/cluster solutions . . . . . . . . . . . . . . . . 31
5.3.1 Factor correlations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
5.4 Determining the number of dimensions to extract. . . . . . . . . . . . . . . 37
5.4.1 Very Simple Structure . . . . . . . . . . . . . . . . . . . . . . . . . . 39
5.4.2 Parallel Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.5 Factor extension . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.6 Comparing multiple solutions . . . . . . . . . . . . . . . . . . . . . . . . . . 41
8 Multilevel modeling 74
8.1 Decomposing data into within and between level correlations using statsBy 74
8.2 Generating and displaying multilevel data . . . . . . . . . . . . . . . . . . . 76
10 Simulation functions 79
11 Graphical Displays 81
12 Miscellaneous functions 81
13 Data sets 83
15 Psychometric Theory 85
2
16 SessionInfo 85
3
1 Overview of this and related documents
To do basic and advanced personality and psychological research using R is not as compli-
cated as some think. This is one of a set of “How To” to do various things using R (R Core
Team, 2019), particularly using the psych (Revelle, 2019) package.
The current list of How To’s includes:
1. Installing R and some useful packages
2. Using R and the psych package to find omegah and ωt .
3. Using R and the psych for factor analysis and principal components analysis. (This
document).
4. Using the score.items function to find scale scores and scale statistics.
5. An overview (vignette) of the psych package
Several functions are meant to do multiple regressions, either from the raw data or
from a variance/covariance matrix, or a correlation matrix. This is discussed in more
detail in
6. How to do mediation and moderation analysis using mediate and setCor is dis-
cuseded in the mediation, moderation and regression analysis tutorial.
1.1 Jump starting the psych package–a guide for the impatient
You have installed psych (section 3) and you want to use it without reading much more.
What should you do?
1. Activate the psych package:
R code
library(psych)
library(psychTools)
2. Input your data (section 4.1). Go to your friendly text editor or data manipulation
program (e.g., Excel) and copy the data to the clipboard. Include a first line that has
the variable labels. Paste it into psych using the read.clipboard.tab command:
R code
myData <- read.clipboard.tab()
\end{Rnput}
\item Make sure that what you just read is right. Describe it (section~\ref{sect:describe}) and perh
\begin{Rinput}
describe(myData)
headTail(myData)
4
3. Look at the patterns in the data. If you have fewer than about 10 variables, look at
the SPLOM (Scatter Plot Matrix) of the data using pairs.panels (section 4.4.1).
R code
pairs.panels(myData)
5. Test for the number of factors in your data using parallel analysis (fa.parallel,
section 5.4.2) or Very Simple Structure (vss, 5.4.1) .
R code
fa.parallel(myData)
vss(myData)
6. Factor analyze (see section 5.1) the data with a specified number of factors (the
default is 1), the default method is minimum residual, the default rotation for more
than one factor is oblimin. There are many more possibilities (see sections 5.1.1-5.1.3).
Compare the solution to a hierarchical cluster analysis using the ICLUST algorithm
(Revelle, 1979) (see section 5.1.6). Also consider a hierarchical factor solution to find
coefficient ω (see 5.1.5). Yet another option is to do a series of factor analyses in
what is known as the “bass akward” procedure (Goldberg, 2006) which considers the
correlation between factors at multiple levels of analysis (see ??).
R code
fa(myData)
iclust(myData)
omega(myData)
bassAckward(myData)
7. Some people like to find coefficient α as an estimate of reliability. This may be done
for a single scale using the alpha function (see 6.1). Perhaps more useful is the
ability to create several scales as unweighted averages of specified items using the
scoreIems function (see 6.4) and to find various estimates of internal consistency for
these scales, find their intercorrelations, and find scores for all the subjects.
R code
alpha(myData) #score all of the items as part of one scale.
myKeys <- make.keys(nvar=20,list(first = c(1,-3,5,-7,8:10),second=c(2,4,-6,11:15,-16)))
my.scores <- scoreItems(myKeys,myData) #form several scales
my.scores #show the highlights of the results
5
At this point you have had a chance to see the highlights of the psych package and to
do some basic (and advanced) data analysis. You might find reading the entire overview
vignette helpful to get a broader understanding of what can be done in R using the psych.
Remember that the help command (?) is available for every function. Try running the
examples for each help page.
6
2 Overview of this and related documents
The psych package (Revelle, 2019) has been developed at Northwestern University since
2005 to include functions most useful for personality, psychometric, and psychological re-
search. The package is also meant to supplement a text on psychometric theory (Revelle,
prep), a draft of which is available at http://personality-project.org/r/book/.
Some of the functions (e.g., read.clipboard, describe, pairs.panels, scatter.hist,
error.bars, multi.hist, bi.bars) are useful for basic data entry and descriptive analy-
ses.
Psychometric applications emphasize techniques for dimension reduction including factor
analysis, cluster analysis, and principal components analysis. The fa function includes
five methods of factor analysis (minimum residual , principal axis, weighted least squares,
generalized least squares and maximum likelihood factor analysis). Determining the num-
ber of factors or components to extract may be done by using the Very Simple Structure
(Revelle and Rocklin, 1979) (vss), Minimum Average Partial correlation (Velicer, 1976)
(MAP) or parallel analysis (fa.parallel) criteria. Item Response Theory (IRT) models for
dichotomous or polytomous items may be found by factoring tetrachoric or polychoric
correlation matrices and expressing the resulting parameters in terms of location and dis-
crimination using irt.fa. Bifactor and hierarchical factor structures may be estimated by
using Schmid Leiman transformations (Schmid and Leiman, 1957) (schmid) to transform
a hierarchical factor structure into a bifactor solution (Holzinger and Swineford, 1937).
Scale construction can be done using the Item Cluster Analysis (Revelle, 1979) (iclust)
function to determine the structure and to calculate reliability coefficients α (Cronbach,
1951)(alpha, scoreItems, score.multiple.choice), β (Revelle, 1979; Revelle and Zin-
barg, 2009) (iclust) and McDonald’s ωh and ωt (McDonald, 1999) (omega). Guttman’s six
estimates of internal consistency reliability (Guttman (1945), as well as additional estimates
(Revelle and Zinbarg, 2009) are in the guttman function. The six measures of Intraclass
correlation coefficients (ICC) discussed by Shrout and Fleiss (1979) are also available.
Graphical displays include Scatter Plot Matrix (SPLOM) plots using pairs.panels, corre-
lation “heat maps” (cor.plot) factor, cluster, and structural diagrams using fa.diagram,
iclust.diagram, structure.diagram, as well as item response characteristics and item
and test information characteristic curves plot.irt and plot.poly.
3 Getting started
Some of the functions described in this overview require other packages. Particularly useful
for rotating the results of factor analyses (from e.g., fa or principal) or hierarchical
factor models using omega or schmid, is the GPArotation package. These and other useful
packages may be installed by first installing and then using the task views (ctv ) package
7
to install the “Psychometrics” task view, but doing it this way is not necessary.
Alternatively, if you have a file name for a remote file, you can read it using read.file as
well.
myData <- read.file(fn) # where file name is the the remote address of the file
8
read.clipboard.csv for reading text that is comma delimited.
read.clipboard.tab for reading text that is tab delimited (e.g., copied directly from an
Excel file).
read.clipboard.lower for reading input of a lower triangular matrix with or without a
diagonal. The resulting object is a square matrix.
read.clipboard.upper for reading input of an upper triangular matrix.
read.clipboard.fwf for reading in fixed width fields (some very old data sets)
For example, given a data set copied to the clipboard from a spreadsheet, just enter the
command
R code
my.data <- read.clipboard()
This will work if every data field has a value and even missing data are given some values
(e.g., NA or -999). If the data were entered in a spreadsheet and the missing values
were just empty cells, then the data should be read in as a tab delimited or by using the
read.clipboard.tab function.
R code
my.data <- read.clipboard(sep="\t") #define the tab option, or
my.tab.data <- read.clipboard.tab() #just use the alternative function
For the case of data in fixed width fields (some old data sets tend to have this format),
copy to the clipboard and then specify the width of each field (in the example below, the
first variable is 5 columns, the second is 2 columns, the next 5 are 1 column the last 4 are
3 columns).
R code
my.data <- read.clipboard.fwf(widths=c(5,2,rep(1,5),rep(3,4))
9
> library(psych)
> library(psychTools)
> data(sat.act)
> describe(sat.act) #basic descriptive statistics
10
> png( 'pairspanels.png' )
> pairs.panels(sat.act,pch='.')
> dev.off()
null device
1
Figure 1: Using the pairs.panels function to graphically show relationships. The x axis
in each scatter plot represents the column variable, the y axis the row variable. Note the
extreme outlier for the ACT. The plot character was set to a period (pch=’.’) in order to
make a cleaner graph.
11
calls cor with use=‘pairwise’, method=‘pearson’ as default values and returns (invisibly)
the full correlation matrix and displays the lower off diagonal matrix.
> lowerCor(sat.act)
When comparing results from two different groups, it is convenient to display them as one
matrix, with the results from one group below the diagonal, and the other group above the
diagonal. Use lowerUpper to do this:
> female <- subset(sat.act,sat.act$gender==2)
> male <- subset(sat.act,sat.act$gender==1)
> lower <- lowerCor(male[-1])
It is also possible to compare two matrices by taking their differences and displaying one (be-
low the diagonal) and the difference of the second from the first above the diagonal:
> diffs <- lowerUpper(lower,upper,diff=TRUE)
> round(diffs,2)
12
4.4.3 Heatmap displays of correlational structure
Perhaps a better way to see the structure in a correlation matrix is to display a heat map
of the correlations. This is just a matrix color coded to represent the magnitude of the
correlation. This is useful when considering the number of factors in a data set. Consider
the Thurstone data set which has a clear 3 factor solution (Figure 2) or a simulated data
set of 24 variables with a circumplex structure (Figure 3). The color coding represents a
“heat map” of the correlations, with darker shades of red representing stronger negative
and darker shades of blue stronger positive correlations. As an option, the value of the
correlation can be shown.
13
> png('corplot.png')
> cor.plot(Thurstone,numbers=TRUE,main="9 cognitive variables from Thurstone")
> dev.off()
null device
1
Figure 2: The structure of correlation matrix can be seen more clearly if the variables are
grouped by factor and then the correlations are shown by color. By using the ’numbers’
option, the values are displayed as well.
14
> png('circplot.png')
> circ <- sim.circ(24)
> r.circ <- cor(circ)
> cor.plot(r.circ,main='24 variables in a circumplex')
> dev.off()
null device
1
Figure 3: Using the cor.plot function to show the correlations in a circumplex. Correlations
are highest near the diagonal, diminish to zero further from the diagonal, and the increase
again towards the corners of the matrix. Circumplex structures are common in the study
of affect.
15
5.1 Dimension reduction through factor analysis and cluster analysis
Parsimony of description has been a goal of science since at least the famous dictum
commonly attributed to William of Ockham to not multiply entities beyond necessity1 . The
goal for parsimony is seen in psychometrics as an attempt either to describe (components)
or to explain (factors) the relationships between many observed variables in terms of a
more limited set of components or latent factors.
The typical data matrix represents multiple items or scales usually thought to reflect fewer
underlying constructs2 . At the most simple, a set of items can be be thought to represent
a random sample from one underlying domain or perhaps a small set of domains. The
question for the psychometrician is how many domains are represented and how well does
each item represent the domains. Solutions to this problem are examples of factor analysis
(FA), principal components analysis (PCA), and cluster analysis (CA). All of these pro-
cedures aim to reduce the complexity of the observed data. In the case of FA, the goal is
to identify fewer underlying constructs to explain the observed data. In the case of PCA,
the goal can be mere data reduction, but the interpretation of components is frequently
done in terms similar to those used when describing the latent variables estimated by FA.
Cluster analytic techniques, although usually used to partition the subject space rather
than the variable space, can also be used to group variables to reduce the complexity of
the data by forming fewer and more homogeneous sets of tests or items.
At the data level the data reduction problem may be solved as a Singular Value Decom-
position of the original matrix, although the more typical solution is to find either the
principal components or factors of the covariance or correlation matrices. Given the pat-
tern of regression weights from the variables to the components or from the factors to the
variables, it is then possible to find (for components) individual component or cluster scores
or estimate (for factors) factor scores.
Several of the functions in psych address the problem of data reduction.
fa incorporates five alternative algorithms: minres factor analysis, principal axis factor
analysis, weighted least squares factor analysis, generalized least squares factor anal-
ysis and maximum likelihood factor analysis. That is, it includes the functionality of
three other functions that will be eventually phased out.
(bassAckward) will do multiple factor and principal components analyses and then show
the relationship between factor levels by finding the interfactor correlations.
fa.extend will extend the factor solution for an X set of variables into a Y set (perhaps
1 Although probably neither original with Ockham nor directly stated by him (Thorburn, 1918), Ock-
factors than observed variables, but are willing to estimate the major underlying factors.
16
of criterion variables).
principal Principal Components Analysis reports the largest n eigen vectors rescaled by
the square root of their eigen values.
factor.congruence The congruence between two factors is the cosine of the angle between
them. This is just the cross products of the loadings divided by the sum of the squared
loadings. This differs from the correlation coefficient in that the mean loading is not
subtracted before taking the products. factor.congruence will find the cosines
between two (or more) sets of factor loadings.
vss Very Simple Structure Revelle and Rocklin (1979) applies a goodness of fit test to
determine the optimal number of factors to extract. It can be thought of as a quasi-
confirmatory model, in that it fits the very simple structure (all except the biggest c
loadings per item are set to zero where c is the level of complexity of the item) of a
factor pattern matrix to the original correlation matrix. For items where the model is
usually of complexity one, this is equivalent to making all except the largest loading
for each item 0. This is typically the solution that the user wants to interpret. The
analysis includes the MAP criterion of Velicer (1976) and a χ 2 estimate.
fa.parallel The parallel factors technique compares the observed eigen values of a cor-
relation matrix with those from random data.
fa.plot will plot the loadings from a factor, principal components, or cluster analysis
(just a call to plot will suffice). If there are more than two factors, then a SPLOM
of the loadings is generated.
nfactors A number of different tests for the number of factors problem are run.
fa.diagram replaces fa.graph and will draw a path diagram representing the factor struc-
ture. It does not require Rgraphviz and thus is probably preferred.
fa.graph requires Rgraphviz and will draw a graphic representation of the factor struc-
ture. If factors are correlated, this will be represented as well.
iclust is meant to do item cluster analysis using a hierarchical clustering algorithm
specifically asking questions about the reliability of the clusters (Revelle, 1979). Clus-
ters are formed until either coefficient α Cronbach (1951) or β Revelle (1979) fail to
increase.
17
n Fk and its transpose plus a diagonal matrix of uniqueness.
R = FF 0 +U 2 (1)
The maximum likelihood solution to this equation is found by factanal in the stats pack-
age. Five alternatives are provided in psych, all of them are included in the fa function
and are called by specifying the factor method (e.g., fm=“minres”, fm=“pa”, fm=“”wls”,
fm=”gls” and fm=”ml”). In the discussion of the other algorithms, the calls shown are to
the fa function specifying the appropriate method.
factor.minres attempts to minimize the off diagonal residual correlation matrix by ad-
justing the eigen values of the original correlation matrix. This is similar to what is done
in factanal, but uses an ordinary least squares instead of a maximum likelihood fit func-
tion. The solutions tend to be more similar to the MLE solutions than are the factor.pa
solutions. min.res is the default for the fa function.
A classic data set, collected by Thurstone and Thurstone (1941) and then reanalyzed by
Bechtoldt (1961) and discussed by McDonald (1999), is a set of 9 cognitive variables with
a clear bi-factor structure Holzinger and Swineford (1937). The minimum residual solution
was transformed into an oblique solution using the default option on rotate which uses
an oblimin transformation (Table 1). Alternative rotations and transformations include
“none”, “varimax”, “quartimax”, “bentlerT”, and “geominT” (all of which are orthogonal
rotations). as well as “promax”, “oblimin”, “simplimax”, “bentlerQ, and“geominQ” and
“cluster” which are possible oblique transformations of the solution. The default is to do
a oblimin transformation, although prior versions defaulted to varimax. The measures of
factor adequacy reflect the multiple correlations of the factors with the best fitting linear
regression estimates of the factor scores (Grice, 2001).
18
Table 1: Three correlated factors from the Thurstone 9 variable problem. By default, the
solution is transformed obliquely using oblimin. The extraction method is (by default)
minimum residual.
> f3t <- fa(Thurstone,3,n.obs=213)
> f3t
Factor Analysis using method = minres
Call: fa(r = Thurstone, nfactors = 3, n.obs = 213)
Standardized loadings (pattern matrix) based upon correlation matrix
MR1 MR2 MR3 h2 u2 com
Sentences 0.90 -0.03 0.04 0.82 0.18 1.0
Vocabulary 0.89 0.06 -0.03 0.84 0.16 1.0
Sent.Completion 0.84 0.03 0.00 0.74 0.26 1.0
First.Letters 0.00 0.85 0.00 0.73 0.27 1.0
Four.Letter.Words -0.02 0.75 0.10 0.63 0.37 1.0
Suffixes 0.18 0.63 -0.08 0.50 0.50 1.2
Letter.Series 0.03 -0.01 0.84 0.73 0.27 1.0
Pedigrees 0.38 -0.05 0.46 0.51 0.49 2.0
Letter.Group -0.06 0.21 0.63 0.52 0.48 1.2
The degrees of freedom for the null model are 36 and the objective function was 5.2 with Chi Square of 10
The degrees of freedom for the model are 12 and the objective function was 0.01
The harmonic number of observations is 213 with the empirical chi square 0.52 with prob < 1
The total number of observations was 213 with Likelihood Chi Square = 2.98 with prob < 1
19
an arbitrary target matrix. The default is to transform towards an independent cluster
solution.
Using the Thurstone data set, three factors were requested and then transformed into an
independent clusters solution using target.rot (Table 2).
Table 2: The 9 variable problem from Thurstone is a classic example of factoring where
there is a higher order factor, g, that accounts for the correlation between the factors. The
extraction method was principal axis. The transformation was a targeted transformation
to a simple cluster solution.
> f3 <- fa(Thurstone,3,n.obs = 213,fm="pa")
> f3o <- target.rot(f3)
> f3o
Call: NULL
Standardized loadings (pattern matrix) based upon correlation matrix
PA1 PA2 PA3 h2 u2
Sentences 0.89 -0.03 0.07 0.81 0.19
Vocabulary 0.89 0.07 0.00 0.80 0.20
Sent.Completion 0.83 0.04 0.03 0.70 0.30
First.Letters -0.02 0.85 -0.01 0.73 0.27
Four.Letter.Words -0.05 0.74 0.09 0.57 0.43
Suffixes 0.17 0.63 -0.09 0.43 0.57
Letter.Series -0.06 -0.08 0.84 0.69 0.31
Pedigrees 0.33 -0.09 0.48 0.37 0.63
Letter.Group -0.14 0.16 0.64 0.45 0.55
20
Table 3: The 9 variable problem from Thurstone is a classic example of factoring where
there is a higher order factor, g, that accounts for the correlation between the factors. The
factors were extracted using a weighted least squares algorithm. All loadings are shown by
using the cut=0 option in the print.psych function.
> f3w <- fa(Thurstone,3,n.obs = 213,fm="wls")
> print(f3w,cut=0,digits=3)
Factor Analysis using method = wls
Call: fa(r = Thurstone, nfactors = 3, n.obs = 213, fm = "wls")
Standardized loadings (pattern matrix) based upon correlation matrix
WLS1 WLS2 WLS3 h2 u2 com
Sentences 0.905 -0.034 0.040 0.822 0.178 1.01
Vocabulary 0.890 0.066 -0.031 0.835 0.165 1.01
Sent.Completion 0.833 0.034 0.007 0.735 0.265 1.00
First.Letters -0.002 0.855 0.003 0.731 0.269 1.00
Four.Letter.Words -0.016 0.743 0.106 0.629 0.371 1.04
Suffixes 0.180 0.626 -0.082 0.496 0.504 1.20
Letter.Series 0.033 -0.015 0.838 0.719 0.281 1.00
Pedigrees 0.381 -0.051 0.464 0.505 0.495 1.95
Letter.Group -0.062 0.209 0.632 0.527 0.473 1.24
The degrees of freedom for the null model are 36 and the objective function was 5.198 with Chi Square of
The degrees of freedom for the model are 12 and the objective function was 0.014
The harmonic number of observations is 213 with the empirical chi square 0.531 with prob < 1
The total number of observations was 213 with Likelihood Chi Square = 2.885 with prob < 0.996
21
> plot(f3t)
Factor Analysis
0.0 0.2 0.4 0.6 0.8
● ● ● ●
● ●
0.8
0.6
MR1
0.4
● ●
0.2
● ●
0.0
● ●
● ● ● ●
● ●
● ●
0.8
● ●
0.6
● ●
MR2
0.4
0.2
● ●
● ●
● ●
0.0
● ● ● ●
● ●
● ●
0.8
0.6
● ●
● ●
0.4
MR3 0.2
● ●
● ●
0.0
● ● ● ●
● ●
● ●
0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8
Figure 4: A graphic representation of the 3 oblique factors from the Thurstone data using
plot. Factors were transformed to an oblique solution using the oblimin function from the
GPArotation package.
22
> fa.diagram(f3t)
Factor Analysis
Sentences
0.9
Vocabulary 0.9
MR1
Sent.Completion 0.8
0.6
First.Letters
0.9
0.6
Suffixes
0.5
Letter.Series 0.8
MR3
Letter.Group 0.6
0.5
Pedigrees
Figure 5: A graphic representation of the 3 oblique factors from the Thurstone data using
fa.diagram. Factors were transformed to an oblique solution using oblimin.
23
to a maximum likelihood solution and fits slightly better than the pa or wls solutions.
Comparisons with SPSS suggest that the pa solution matches the SPSS OLS solution, but
that the minres solution is slightly better. At least in one test data set, the weighted least
squares solutions, although fitting equally well, had slightly different structure loadings.
Note that the rotations used by SPSS will sometimes use the “Kaiser Normalization”. By
default, the rotations used in psych do not normalize, but this can be specified as an option
in fa.
24
Table 4: The Thurstone problem can also be analyzed using Principal Components Anal-
ysis. Compare this to Table 2. The loadings are higher for the PCA because the model
accounts for the unique as well as the common variance.The fit of the off diagonal elements,
however, is much worse than the fa results.
> p3p <-principal(Thurstone,3,n.obs = 213,rotate="Promax")
> p3p
Principal Components Analysis
Call: principal(r = Thurstone, nfactors = 3, rotate = "Promax", n.obs = 213)
Standardized loadings (pattern matrix) based upon correlation matrix
RC1 RC2 RC3 h2 u2 com
Sentences 0.92 0.01 0.01 0.86 0.14 1.0
Vocabulary 0.90 0.10 -0.05 0.86 0.14 1.0
Sent.Completion 0.91 0.04 -0.04 0.83 0.17 1.0
First.Letters 0.01 0.84 0.07 0.78 0.22 1.0
Four.Letter.Words -0.05 0.81 0.17 0.75 0.25 1.1
Suffixes 0.18 0.79 -0.15 0.70 0.30 1.2
Letter.Series 0.03 -0.03 0.88 0.78 0.22 1.0
Pedigrees 0.45 -0.16 0.57 0.67 0.33 2.1
Letter.Group -0.19 0.19 0.86 0.75 0.25 1.2
25
> om.h <- omega(Thurstone,n.obs=213,sl=FALSE)
> op <- par(mfrow=c(1,1))
Omega
Sentences
0.9
Vocabulary
0.9
F1
0.8
Sent.Completion
0.8
First.Letters
0.9
0.6 0.4
Suffixes
0.7
Letter.Series
0.2 0.8
F3
0.6
Letter.Group
0.5
Pedigrees
26
> om <- omega(Thurstone,n.obs=213)
Omega
Sentences
0.6
0.7 Vocabulary
0.6
F1*
0.7 0.5
Sent.Completion
0.7
First.Letters
0.6 0.6
Pedigrees
27
Yet another approach to the bifactor structure is do use the bifactor rotation function in
either psych or in GPArotation. This does the rotation discussed in Jennrich and Bentler
(2011).
28
may be specified a priori, or found empirically. The resulting statistics include the average
split half reliability, α (Cronbach, 1951), as well as the worst split half reliability, β (Revelle,
1979), which is an estimate of the general factor saturation of the resulting scale (Figure 8).
Cluster loadings (corresponding to the structure matrix of factor analysis) are reported
when printing (Table 7). The pattern matrix is available as an object in the results.
> data(bfi)
> ic <- iclust(bfi[1:25])
ICLUST
O2 0.57
0.57 C7
O5 0.32 C21
−0.42 α = 0.41
O4 β = 0.27
C3 C14
0.7 α = 0.63
C2 0.65 C1 0.84 β = 0.58
0.65 0.77
C1 C15
α = 0.73
C5 0.69 −0.79 β = 0.67
0.69 C2
C4
N5 C13
0.75 α = 0.71
N4 0.72 0.86 β = 0.65
0.72 C8 0.86
N3 C16
α = 0.81
N2 0.84 0.76 β = 0.76
0.84 C3
N1
E4 0.72
−0.72 C6
E2 0.77 C11
−0.77 α = 0.72
C19
E1 β = 0.69 0.6
α = 0.76
O3 0.63 β = 0.64
0.63 C4 0.79
O1 0.73 C12
α = 0.68 0.63
E5 0.63 0.9 β = 0.64
0.63 C9
E3 C20
α = 0.81
C17
A4 0.7 β = 0.63
α = 0.72
C10
A2 0.77 0.78 β = 0.61
α = 0.72 0.69
A5 0.71 0.91 β = 0.68 0.54
0.71 C5
A3 C18
−0.61 α = 0.71
A1 β = 0.5
Figure 8: Using the iclust function to find the cluster structure of 25 personality items
(the three demographic variables were excluded from this analysis). When analyzing many
variables, the tree structure may be seen more clearly if the graphic output is saved as a
pdf and then enlarged using a pdf viewer.
The previous analysis (Figure 8) was done using the Pearson correlation. A somewhat
cleaner structure is obtained when using the polychoric function to find polychoric corre-
29
Table 5: The summary statistics from an iclust analysis shows three large clusters and
smaller cluster.
> summary(ic) #show the results
ICLUST (Item Cluster Analysis)Call: iclust(r.mat = bfi[1:25])
ICLUST
Purified Alpha:
C20 C16 C15 C21
0.80 0.81 0.73 0.61
Guttman Lambda6*
C20 C16 C15 C21
0.82 0.81 0.72 0.61
Original Beta:
C20 C16 C15 C21
0.63 0.76 0.67 0.27
Cluster size:
C20 C16 C15 C21
10 5 5 5
30
lations (Figure 9). Note that the first time finding the polychoric correlations some time,
but the next three analyses were done using that correlation matrix (r.poly$rho). When
using the console for input, polychoric will report on its progress while working using
progressBar.
Table 6: The polychoric and the tetrachoric functions can take a long time to finish
and report their progress by a series of dots as they work. The dots are suppressed when
creating a Sweave document.
> data(bfi)
> r.poly <- polychoric(bfi[1:25]) #the ... indicate the progress of the function
A comparison of these four cluster solutions suggests both a problem and an advantage of
clustering techniques. The problem is that the solutions differ. The advantage is that the
structure of the items may be seen more clearly when examining the clusters rather than
a simple factor solution.
31
> ic.poly <- iclust(r.poly$rho,title="ICLUST using polychoric correlations")
> iclust.diagram(ic.poly)
ICLUST using polychoric correlations
N5 C13
0.78 α = 0.74
N4 0.75 0.86 β = 0.69
0.75 C8 0.88
N3 C16
α = 0.84
N2 0.87 0.77 β = 0.79
0.87 C5
N1
O5 0.63
0.63 C7
O2 0.4 C21
−0.52 α = 0.48
C22
O4 β = 0.35
0.31
α = 0.81
O3 0.67 β = 0.29
0.67 C2
O1 0.75 C12
α = 0.72
E5 0.66 0.9 β = 0.68
0.66 C9 0.7
E3 C19 −0.31
α = 0.8
C11
E1 0.8 −0.68 β = 0.66
α = 0.77
E4 0.76 −0.91 β = 0.73 0.66
−0.76 C6 0.6
E2 C20
α = 0.83
C18−0.69
A1 0.65 β = 0.66
α = 0.76
C17
A4 0.73 β = 0.56
−0.68
α = 0.77
C10
A2 0.81 0.79 β = 0.65
α = 0.77
A5 0.76 0.92 β = 0.73
0.76 C3
A3 C23
α = 0.83
C2 0.7 β = 0.58
0.7 C4
C1 0.69 C14 −0.49
0.72 α = 0.67
C15
C3 β = 0.61
0.8
α = 0.77
C5 0.73 C1 −0.81 β = 0.71
C4 0.73
Figure 9: ICLUST of the BFI data set using polychoric correlations. Compare this solution
to the previous one (Figure 8) which was done using Pearson correlations.
32
> ic.poly <- iclust(r.poly$rho,5,title="ICLUST using polychoric correlations for nclusters=5")
> iclust.diagram(ic.poly)
ICLUST using polychoric correlations for nclusters=5
O5 0.63
0.63 C7
O2
O4
C2 0.7
0.7 C4
C1 0.69 C14
0.72 α = 0.67
C15
C3 β = 0.61
0.8
α = 0.77
C5 0.73 C1 −0.81 β = 0.71
C4 0.73
N5 C13
0.78 α = 0.74
N4 0.75 0.86 β = 0.69
0.75 C8 0.88
N3 C16
α = 0.84
N2 0.87 0.77 β = 0.79
0.87 C5
N1
O3 0.67
0.67 C2
O1 0.75 C12
α = 0.72
E5 0.66 0.9 β = 0.68
0.66 C9 0.7
E3 C19
α = 0.8
C11
E1 0.8 −0.68 β = 0.66
α = 0.77
E4 0.76 −0.91 β = 0.73 0.66
−0.76 C6
E2 C20
α = 0.83
C18−0.69
A1 0.65 β = 0.66
α = 0.76
C17
A4 0.73 β = 0.56
−0.68
α = 0.77
C10
A2 0.81 0.79 β = 0.65
α = 0.77
A5 0.76 0.92 β = 0.73
0.76 C3
A3
Figure 10: ICLUST of the BFI data set using polychoric correlations with the solution
set to 5 clusters. Compare this solution to the previous one (Figure 9) which was done
without specifying the number of clusters and to the next one (Figure 11) which was done
by changing the beta criterion.
33
> ic.poly <- iclust(r.poly$rho,beta.size=3,title="ICLUST beta.size=3")
ICLUST beta.size=3
N5 C13
0.78 α = 0.74
N4 0.75 0.86 β = 0.69
0.75 C8 0.88
N3 C16
α = 0.84
N2 0.87 0.77 β = 0.79
0.87 C5
N1
O5 0.63
0.63 C7
O2 0.4 C21
−0.52 α = 0.48
C22
O4 β = 0.35
0.31
α = 0.81
O3 0.67 β = 0.29
0.67 C2
O1 0.75 C12
α = 0.72
E5 0.66 0.9 β = 0.68
0.66 C9 0.7
E3 C19 −0.31
α = 0.8
C11
E1 0.8 −0.68 β = 0.66
α = 0.77
E4 0.76 −0.91 β = 0.73 0.66
−0.76 C6 0.6
E2 C20
α = 0.83
C18−0.69
A1 0.65 β = 0.66
α = 0.76
C17
A4 0.73 β = 0.56
−0.68
α = 0.77
C10
A2 0.81 0.79 β = 0.65
α = 0.77
A5 0.76 0.92 β = 0.73
0.76 C3
A3 C23
α = 0.83
C2 0.7 β = 0.58
0.7 C4
C1 0.69 C14 −0.49
0.72 α = 0.67
C15
C3 β = 0.61
0.8
α = 0.77
C5 0.73 C1 −0.81 β = 0.71
C4 0.73
Figure 11: ICLUST of the BFI data set using polychoric correlations with the beta criterion
set to 3. Compare this solution to the previous three (Figure 8, 9, 10).
34
Table 7: The output from iclustincludes the loadings of each item on each cluster. These
are equivalent to factor structure loadings. By specifying the value of cut, small loadings
are suppressed. The default is for cut=0.su
> print(ic,cut=.3)
ICLUST (Item Cluster Analysis)
Call: iclust(r.mat = bfi[1:25])
Purified Alpha:
C20 C16 C15 C21
0.80 0.81 0.73 0.61
G6* reliability:
C20 C16 C15 C21
0.83 1.00 0.67 0.38
Original Beta:
C20 C16 C15 C21
0.63 0.76 0.67 0.27
Cluster size:
C20 C16 C15 C21
10 5 5 5
MR1 MR2
SS loadings 1.80 1.78
Proportion Var 0.18 0.18
Cumulative Var 0.18 0.36
Proportion Explained 0.50 0.50
Cumulative Proportion 0.50 1.00
The degrees of freedom for the null model are 45 and the objective function was 2.03 with Chi Square of 5664.89
The degrees of freedom for the model are 26 and the objective function was 0.17
The harmonic number of observations is 2762 with the empirical chi square 397.07 with prob < 5e-68
The total number of observations was 2800 with Likelihood Chi Square = 468.37 with prob < 1.2e-82
36
PA3 0.35 -0.24 0.88 -0.37
PA4 0.29 -0.12 0.27 -0.90
A more complete comparison of oblique factor solutions (both minres and principal axis), bi-
factor and component solutions to the Thurstone data set is done using the factor.congruence
function. (See table 9).
Table 9: Congruence coefficients for oblique factor, bifactor and component solutions for
the Thurstone problem.
> factor.congruence(list(f3t,f3o,om,p3p))
MR1 MR2 MR3 PA1 PA2 PA3 g F1* F2* F3* h2 RC1 RC2 RC3
MR1 1.00 0.06 0.09 1.00 0.06 0.13 0.72 1.00 0.06 0.09 0.73 1.00 0.08 0.04
MR2 0.06 1.00 0.08 0.03 1.00 0.07 0.60 0.06 1.00 0.08 0.57 0.04 0.99 0.12
MR3 0.09 0.08 1.00 0.01 0.01 1.00 0.52 0.09 0.08 1.00 0.51 0.06 0.02 0.99
PA1 1.00 0.03 0.01 1.00 0.04 0.05 0.67 1.00 0.03 0.01 0.68 1.00 0.06 -0.04
PA2 0.06 1.00 0.01 0.04 1.00 0.00 0.57 0.06 1.00 0.01 0.54 0.04 0.99 0.05
PA3 0.13 0.07 1.00 0.05 0.00 1.00 0.54 0.13 0.07 1.00 0.54 0.10 0.01 0.99
g 0.72 0.60 0.52 0.67 0.57 0.54 1.00 0.72 0.60 0.52 0.99 0.69 0.58 0.50
F1* 1.00 0.06 0.09 1.00 0.06 0.13 0.72 1.00 0.06 0.09 0.73 1.00 0.08 0.04
F2* 0.06 1.00 0.08 0.03 1.00 0.07 0.60 0.06 1.00 0.08 0.57 0.04 0.99 0.12
F3* 0.09 0.08 1.00 0.01 0.01 1.00 0.52 0.09 0.08 1.00 0.51 0.06 0.02 0.99
h2 0.73 0.57 0.51 0.68 0.54 0.54 0.99 0.73 0.57 0.51 1.00 0.71 0.56 0.50
RC1 1.00 0.04 0.06 1.00 0.04 0.10 0.69 1.00 0.04 0.06 0.71 1.00 0.06 0.00
RC2 0.08 0.99 0.02 0.06 0.99 0.01 0.58 0.08 0.99 0.02 0.56 0.06 1.00 0.05
RC3 0.04 0.12 0.99 -0.04 0.05 0.99 0.50 0.04 0.12 0.99 0.50 0.00 0.05 1.00
37
Table 10: Factor correlations and factor congruences between “minres” factor analysis and
“pca” principal components using “oblimin” rotation for both.
> faCor(Thurstone,c(3,3),fm=c("minres","pca"), rotate=c("oblimin","oblimin"))
Call: faCor(r = Thurstone, nfactors = c(3, 3), fm = c("minres", "pca"),
rotate = c("oblimin", "oblimin"))
2) Extracting factors until the change in chi square from factor n to factor n+1 is not
significant.
3) Extracting factors until the eigen values of the real data are less than the corresponding
eigen values of a random data set of the same size (parallel analysis) fa.parallel (Horn,
1965).
4) Plotting the magnitude of the successive eigen values and applying the scree test (a
sudden drop in eigen values analogous to the change in slope seen when scrambling up the
talus slope of a mountain and approaching the rock face (Cattell, 1966).
5) Extracting factors as long as they are interpretable.
6) Using the Very Structure Criterion (vss) (Revelle and Rocklin, 1979).
7) Using Wayne Velicer’s Minimum Average Partial (MAP) criterion (Velicer, 1976).
8) Extracting principal components until the eigen value < 1.
Each of the procedures has its advantages and disadvantages. Using either the chi square
test or the change in square test is, of course, sensitive to the number of subjects and leads
to the nonsensical condition that if one wants to find many factors, one simply runs more
subjects. Parallel analysis is partially sensitive to sample size in that for large samples the
eigen values of random factors will be very small. The scree test is quite appealing but can
lead to differences of interpretation as to when the scree“breaks”. Extracting interpretable
factors means that the number of factors reflects the investigators creativity more than the
data. vss, while very simple to understand, will not work very well if the data are very
factorially complex. (Simulations suggests it will work fine if the complexities of some of
the items are no more than 2). The eigen value of 1 rule, although the default for many
38
programs, seems to be a rough way of dividing the number of variables by 3 and is probably
the worst of all criteria.
An additional problem in determining the number of factors is what is considered a factor.
Many treatments of factor analysis assume that the residual correlation matrix after the
factors of interest are extracted is composed of just random error. An alternative con-
cept is that the matrix is formed from major factors of interest but that there are also
numerous minor factors of no substantive interest but that account for some of the shared
covariance between variables. The presence of such minor factors can lead one to extract
too many factors and to reject solutions on statistical grounds of misfit that are actually
very good fits to the data. This problem is partially addressed later in the discussion of
simulating complex structures using sim.structure and of small extraneous factors using
the sim.minor function.
39
> vss <- vss(bfi[1:25],title="Very Simple Structure of a Big 5 inventory")
4 4
0.8
4
3 4
4
3 3 3 3
2 2
Very Simple Structure Fit
3
2 2 2
2
2
0.6
1
1 1
1 1
1 1
1
0.4
0.2
0.0
1 2 3 4 5 6 7 8
Number of Factors
Figure 12: The Very Simple Structure criterion for the number of factors compares solutions
for various levels of item complexity and various numbers of factors. For the Big 5 Inventory,
the complexity 1 and 2 solutions both achieve their maxima at four factors. This is in
contrast to parallel analysis which suggests 6 and the MAP criterion which suggests 5.
40
6 649 0.020 0.027 -661
7 435 0.016 0.023 -724
8 278 0.013 0.020 -738
41
> fa.parallel(bfi[1:25],main="Parallel Analysis of a Big 5 inventory")
Parallel analysis suggests that the number of factors = 6 and the number of components = 6
PC Actual Data
PC Simulated Data
PC Resampled Data
FA Actual Data
4
FA Simulated Data
FA Resampled Data
3
2
1
0
5 10 15 20 25
Factor/Component Number
Figure 13: Parallel analysis compares factor and principal components solutions to the real
data as well as resampled data. Although vss suggests 4 factors, MAP 5, parallel analysis
suggests 6. One more demonstration of Kaiser’s dictum.
42
> v16 <- sim.item(16)
> s <- c(1,3,5,7,9,11,13,15)
> f2 <- fa(v16[,s],2)
> fe <- fa.extension(cor(v16)[s,-s],f2)
> fa.diagram(f2,fe=fe)
Factor analysis and extension
V13 V16
0.7 0.6
V5 V14
−0.6 0.6
0.6 −0.5
V15 V8
V9 V2
0.6 −0.6
−0.6 −0.6
V1 V4
−0.5 0.6
V3 V10
Figure 14: Factor extension applies factors from one set (those on the left) to another set
of variables (those on the right). fa.extension is particularly useful when one wants to
define the factors with one set of variables and then apply those factors to another set.
fa.diagram is used to show the structure.
43
> fe <- fa.extend(bfi,5,ov=1:25,ev=26:28)
> diagram(fe)
Factor analysis and extension
N1
N2 0.8
0.8
N3
0.7
N5 0.5 MR2
N4 0.5
E2
−0.7 −0.1
E4 age
0.6
E1 −0.6
0.4 MR1
E5
0.4 0.1
E3
C2
0.7
C4 −0.6
C3 0.6 MR3 0.3
−0.6 education
C5
0.5
C1
A3
0.1
0.1
0.7
A2 0.6 MR5
A5 0.5 0.1
0.4
A4 0.3
−0.4
gender
A1 −0.2
O3 0.6
−0.5 MR4
O5
0.5
O1
−0.5
O2 0.4
O4
Figure 15: Factor extension applies factors from one set (those on the left) to another set
of variables (those on the right). fa.extend is particularly useful when one wants to define
the factors with one set of variables and then apply those factors to another set. diagram
is used to show the structure.
44
two solutions from the same correlation matrix, R , F1 and F2 are found by using the two
weights matrices, W1 and W2 (for finding factor scores for the first and second model) and
then finding the factor covariances, C = W10 RW2 which may then be converted to factor
correlations by dividing by the square root of the diagonal of C. By default bassAckward
uses the correlation preserving weights discussed by ten Berge et al. (1999), although other
options (e.g. regression weights) may also be used.
> ba5 <- bassAckward(bfi[1:25], nfactors =c(2,3,4,5),plot=FALSE)
> baf <- bassAckward.diagram(ba5)
BassAckward
E2
E4
E1 −0.68
−0.56
0.59
E5
0.42
E3 0.42 F1 0.82
A3 F1
A2 −0.34 0.99
0.640.66 0.81 F1
A5 0.95
A4 0.53
0.43 F2 F1
A1 −0.41
N1 F2
N2 0.81
−0.39 0.99 1
N3 0.78
0.71 F3 F2
N5 0.490.47 0.34
N4 −0.32
1
C2 F3
C4 0.67 0.99 0.59
−0.61 F4 F2
C3 0.57 0.94
−0.55
C5 0.55
C1 F3
0.49
O3 0.61 0.99 F4
O5 −0.54 F5
0.51
O1 −0.46
O2 0.37
O4
Figure 16: bassAckward compares solutions at multiple levels by successive factoring and
the finding the factor correlations across levels. Compare the three factor solution to the
five factor solution. The dimensions of social approach, withdrawal, and constraint seen at
the three factor level become the more traditional CANOE factors at the five factor level.
And we show the items associated with this solution by using fa.lookup (Table 11)
45
Table 11: bfi items sorted in the order of the five factors from bassAckward
> fa.lookup(baf$bass.ack[[5]],dictionary=bfi.dictionary[2])
F1 F2 F3 F4 F5 Item
E2 -0.68 -0.05 0.10 -0.02 -0.06 Find it difficult to approach others.
E4 0.59 0.29 0.01 0.02 -0.08 Make friends easily.
E1 -0.56 -0.08 -0.06 0.11 -0.10 Don't talk a lot.
E5 0.42 0.05 0.15 0.27 0.21 Take charge.
E3 0.42 0.25 0.08 0.00 0.28 Know how to captivate people.
A3 0.12 0.66 -0.03 0.02 0.03 Know how to comfort others.
A2 0.00 0.64 -0.02 0.08 0.03 Inquire about others' well-being.
A5 0.23 0.53 -0.11 0.01 0.04 Make people feel at ease.
A4 0.06 0.43 -0.06 0.19 -0.15 Love children.
A1 0.17 -0.41 0.21 0.07 -0.06 Am indifferent to the feelings of others.
N1 0.10 -0.11 0.81 0.00 -0.05 Get angry easily.
N2 0.04 -0.09 0.78 0.01 0.01 Get irritated easily.
N3 -0.10 0.08 0.71 -0.04 0.02 Have frequent mood swings.
N5 -0.20 0.21 0.49 0.00 -0.15 Panic easily.
N4 -0.39 0.09 0.47 -0.14 0.08 Often feel blue.
C2 -0.09 0.08 0.15 0.67 0.04 Continue until everything is perfect.
C4 0.00 0.04 0.17 -0.61 -0.05 Do things in a half-way manner.
C3 -0.06 0.09 0.03 0.57 -0.07 Do things according to a plan.
C5 -0.14 0.02 0.19 -0.55 0.09 Waste my time.
C1 -0.03 -0.02 0.07 0.55 0.15 Am exacting in my work.
O3 0.15 0.08 0.03 0.02 0.61 Carry the conversation to a higher level.
O5 0.10 0.04 0.13 -0.03 -0.54 Will not probe deeply into a subject.
O1 0.10 0.02 0.02 0.07 0.51 Am full of ideas.
O2 0.06 0.16 0.19 -0.08 -0.46 Avoid difficult reading material.
O4 -0.32 0.17 0.13 -0.02 0.37 Spend time reflecting on things.
>
46
6 Classical Test Theory and Reliability
Surprisingly, 107 years after Spearman (1904) introduced the concept of reliability to psy-
chologists, there are still multiple approaches for measuring it. Although very popular,
Cronbach’s α (Cronbach, 1951) underestimates the reliability of a test and over estimates
the first factor saturation (Revelle and Zinbarg, 2009).
α (Cronbach, 1951) is the same as Guttman’s λ 3 (Guttman, 1945) and may be found
by
n tr(V )x n Vx − tr(V x )
λ3 = 1− = =α
n−1 Vx n−1 Vx
∑ e2j 2 )
∑(1 − rsmc
λ6 = 1 − = 1− .
Vx Vx
The squared multiple correlation is a lower bound for the item communality and as the
number of items increases, becomes a better estimate.
G6 is also sensitive to lumpiness in the test and should not be taken as a measure of
unifactorial structure. For lumpy tests, it will be greater than alpha. For tests with equal
item loadings, alpha > G6, but if the loadings are unequal or if there is a general factor,
G6 > alpha. G6 estimates item reliability by the squared multiple correlation of the other
items in a scale. A modification of G6, G6*, takes as an estimate of an item reliability
the smc with all the items in an inventory, including those not keyed for a particular scale.
This will lead to a better estimate of the reliable variance of a particular item.
Alpha, G6 and G6* are positive functions of the number of items in a test as well as the av-
erage intercorrelation of the items in the test. When calculated from the item variances and
total test variance, as is done here, raw alpha is sensitive to differences in the item variances.
Standardized alpha is based upon the correlations rather than the covariances.
47
More complete reliability analyses of a single scale can be done using the omega function
which finds ωh and ωt based upon a hierarchical factor analysis.
Alternative functions scoreItems and cluster.cor will also score multiple scales and
report more useful statistics. “Standardized” alpha is calculated from the inter-item corre-
lations and will differ from raw alpha.
Functions for examining the reliability of a single scale or a set of scales include:
alpha Internal consistency measures of reliability range from ωh to α to ωt . The alpha
function reports two estimates: Cronbach’s coefficient α and Guttman’s λ6 . Also
reported are item - whole correlations, α if an item is omitted, and item means and
standard deviations.
guttman Eight alternative estimates of test reliability include the six discussed by Guttman
(1945), four discussed by ten Berge and Zergers (1978) (µ0 . . . µ3 ) as well as β (the
worst split half, Revelle, 1979), the glb (greatest lowest bound) discussed by Bentler
and Woodward (1980), and ωh andωt ((McDonald, 1999; Zinbarg et al., 2005).
omega Calculate McDonald’s omega estimates of general and total factor saturation.
(Revelle and Zinbarg (2009) compare these coefficients with real and artificial data
sets.)
cluster.cor Given a n x c cluster definition matrix of -1s, 0s, and 1s (the keys) , and a n
x n correlation matrix, find the correlations of the composite clusters.
scoreItems Given a matrix or data.frame of k keys for m items (-1, 0, 1), and a matrix
or data.frame of items scores for m items and n people, find the sum scores or av-
erage scores for each person and each scale. If the input is a square matrix, then
it is assumed that correlations or covariances were used, and the raw scores are not
available. In addition, report Cronbach’s alpha, coefficient G6*, the average r, the
scale intercorrelations, and the item by scale correlations (both raw and corrected for
item overlap and scale reliability). Replace missing values with the item median or
mean if desired. Will adjust scores for reverse scored items.
score.multiple.choice Ability tests are typically multiple choice with one right answer.
score.multiple.choice takes a scoring key and a data matrix (or data.frame) and finds
total or average number right for each participant. Basic test statistics (alpha, average
r, item means, item-whole correlations) are also reported.
48
McDonald’s ωh and ωt . Consider a simulated data set, representing 9 items with a hierar-
chical structure and the following correlation matrix. Then using the alpha function, the
α and λ6 estimates of reliability may be found for all 9 items, as well as the if one item is
dropped at a time.
> set.seed(17)
> r9 <- sim.hierarchical(n=500,raw=TRUE)$observed
> round(cor(r9),2)
V1 V2 V3 V4 V5 V6 V7 V8 V9
V1 1.00 0.58 0.59 0.41 0.44 0.30 0.40 0.31 0.25
V2 0.58 1.00 0.48 0.35 0.36 0.22 0.24 0.29 0.19
V3 0.59 0.48 1.00 0.32 0.35 0.23 0.29 0.20 0.17
V4 0.41 0.35 0.32 1.00 0.44 0.36 0.26 0.27 0.22
V5 0.44 0.36 0.35 0.44 1.00 0.32 0.24 0.23 0.20
V6 0.30 0.22 0.23 0.36 0.32 1.00 0.26 0.26 0.12
V7 0.40 0.24 0.29 0.26 0.24 0.26 1.00 0.38 0.25
V8 0.31 0.29 0.20 0.27 0.23 0.26 0.38 1.00 0.25
V9 0.25 0.19 0.17 0.22 0.20 0.12 0.25 0.25 1.00
> alpha(r9)
Reliability analysis
Call: alpha(x = r9)
Item statistics
n raw.r std.r r.cor r.drop mean sd
V1 500 0.77 0.76 0.76 0.67 0.0327 1.02
V2 500 0.67 0.66 0.62 0.55 0.1071 1.00
V3 500 0.65 0.65 0.61 0.53 -0.0462 1.01
V4 500 0.65 0.65 0.59 0.53 -0.0091 0.98
V5 500 0.64 0.64 0.58 0.52 -0.0012 1.03
V6 500 0.55 0.55 0.46 0.41 -0.0499 0.99
V7 500 0.60 0.60 0.52 0.46 0.0481 1.01
V8 500 0.58 0.57 0.49 0.43 0.0526 1.07
V9 500 0.47 0.48 0.36 0.32 0.0164 0.97
Some scales have items that need to be reversed before being scored. Rather than reversing
the items in the raw data, it is more convenient to just specify which items need to be
49
reversed scored. This may be done in alpha by specifying a keys vector of 1s and -1s.
(This concept of keys vector is more useful when scoring multiple scale inventories, see
below.) As an example, consider scoring the 7 attitude items in the attitude data set.
Assume a conceptual mistake in that item 2 is to be scored (incorrectly) negatively.
> keys <- c(1,-1,1,1,1,1,1)
> alpha(attitude,keys)
Reliability analysis
Call: alpha(x = attitude, keys = keys)
Item statistics
n raw.r std.r r.cor r.drop mean sd
rating 30 0.60 0.60 0.60 0.33 65 12.2
complaints- 30 -0.57 -0.58 -0.74 -0.74 50 13.3
privileges 30 0.66 0.65 0.54 0.42 53 12.2
learning 30 0.80 0.79 0.78 0.64 56 11.7
raises 30 0.82 0.83 0.85 0.69 65 10.4
critical 30 0.50 0.53 0.35 0.27 75 9.9
advance 30 0.74 0.75 0.71 0.58 43 10.3
Note how the reliability of the 7 item scales with an incorrectly reversed item is very poor,
but if the item 2 is dropped then the reliability is improved substantially. This suggests
that item 2 was incorrectly scored. Doing the analysis again with item 2 positively scored
produces much more favorable results.
> keys <- c(1,1,1,1,1,1,1)
> alpha(attitude,keys)
Reliability analysis
Call: alpha(x = attitude, keys = keys)
50
rating 0.81 0.81 0.83 0.41 4.2 0.052 0.035 0.45
complaints 0.80 0.80 0.82 0.39 3.9 0.057 0.035 0.43
privileges 0.83 0.82 0.87 0.44 4.7 0.048 0.054 0.53
learning 0.80 0.80 0.84 0.40 4.0 0.054 0.045 0.38
raises 0.80 0.78 0.83 0.38 3.6 0.056 0.048 0.34
critical 0.86 0.86 0.89 0.51 6.3 0.038 0.030 0.56
advance 0.84 0.83 0.86 0.46 5.0 0.043 0.048 0.49
Item statistics
n raw.r std.r r.cor r.drop mean sd
rating 30 0.78 0.76 0.75 0.67 65 12.2
complaints 30 0.84 0.81 0.82 0.74 67 13.3
privileges 30 0.70 0.68 0.60 0.56 53 12.2
learning 30 0.81 0.80 0.78 0.71 56 11.7
raises 30 0.85 0.86 0.85 0.79 65 10.4
critical 30 0.42 0.45 0.31 0.27 75 9.9
advance 30 0.60 0.62 0.56 0.46 43 10.3
It is useful when considering items for a potential scale to examine the item distribution.
This is done in scoreItems as well as in alpha.
> items <- sim.congeneric(N=500,short=FALSE,low=-2,high=2,categorical=TRUE) #500 responses to 4 discrete item
> alpha(items$observed) #item response analysis of congeneric measures
Reliability analysis
Call: alpha(x = items$observed)
Item statistics
n raw.r std.r r.cor r.drop mean sd
V1 500 0.81 0.81 0.74 0.62 0.058 0.97
V2 500 0.74 0.75 0.63 0.52 0.012 0.98
V3 500 0.72 0.72 0.57 0.48 0.056 0.99
V4 500 0.68 0.67 0.48 0.41 0.098 1.02
51
6.2 Using omega to find the reliability of a single scale
Two alternative estimates of reliability that take into account the hierarchical structure of
the inventory are McDonald’s ωh and ωt . These may be found using the omega function
for an exploratory analysis (See Figure 17) or omegaSem for a confirmatory analysis using
the sem based upon the exploratory solution from omega.
McDonald has proposed coefficient omega (hierarchical) (ωh ) as an estimate of the gen-
eral factor saturation of a test. Zinbarg et al. (2005) http://personality-project.
org/revelle/publications/zinbarg.revelle.pmet.05.pdf compare McDonald’s ωh to
Cronbach’s α and Revelle’s β . They conclude that ωh is the best estimate. (See also Zin-
barg et al. (2006) and Revelle and Zinbarg (2009) http://personality-project.org/
revelle/publications/revelle.zinbarg.08.pdf ).
One way to find ωh is to do a factor analysis of the original data set, rotate the factors
obliquely, factor that correlation matrix, do a Schmid-Leiman (schmid) transformation to
find general factor loadings, and then find ωh .
ωh differs slightly as a function of how the factors are estimated. Four options are available,
the default will do a minimum residual factor analysis, fm=“pa” does a principal axes factor
analysis (factor.pa), fm=“mle” uses the factanal function, and fm=“pc” does a principal
components analysis (principal).
For ability items, it is typically the case that all items will have positive loadings on the
general factor. However, for non-cognitive items it is frequently the case that some items
are to be scored positively, and some negatively. Although probably better to specify
which directions the items are to be scored by specifying a key vector, if flip =TRUE
(the default), items will be reversed so that they have positive loadings on the general
factor. The keys are reported so that scores can be found using the scoreItems function.
Arbitrarily reversing items this way can overestimate the general factor. (See the example
with a simulated circumplex).
β , an alternative to ω, is defined as the worst split half reliability. It can be estimated by
using iclust (Item Cluster analysis: a hierarchical clustering algorithm). For a very com-
plimentary review of why the iclust algorithm is useful in scale construction, see Cooksey
and Soutar (2006).
The omega function uses exploratory factor analysis to estimate the ωh coefficient. It is
important to remember that “A recommendation that should be heeded, regardless of the
method chosen to estimate ωh , is to always examine the pattern of the estimated general
factor loadings prior to estimating ωh . Such an examination constitutes an informal test
of the assumption that there is a latent variable common to all of the scale’s indicators
that can be conducted even in the context of EFA. If the loadings were salient for only a
relatively small subset of the indicators, this would suggest that there is no true general
52
factor underlying the covariance matrix. Just such an informal assumption test would have
afforded a great deal of protection against the possibility of misinterpreting the misleading
ωh estimates occasionally produced in the simulations reported here.” (Zinbarg et al., 2006,
p 137).
Although ωh is uniquely defined only for cases where 3 or more subfactors are extracted, it
is sometimes desired to have a two factor solution. By default this is done by forcing the
schmid extraction to treat the two subfactors as having equal loadings.
There are three possible options for this condition: setting the general factor loadings
√
between the two lower order factors to be “equal” which will be the rab where rab is the
oblique correlation between the factors) or to “first” or “second” in which case the general
factor is equated with either the first or second group factor. A message is issued suggesting
that the model is not really well defined. This solution discussed in Zinbarg et al., 2007.
To do this in omega, add the option=“first” or option=“second” to the call.
Although obviously not meaningful for a 1 factor solution, it is of course possible to find
the sum of the loadings on the first (and only) factor, square them, and compare them to
the overall matrix variance. This is done, with appropriate complaints.
In addition to ωh , another of McDonald’s coefficients is ωt . This is an estimate of the total
reliability of a test.
McDonald’s ωt , which is similar to Guttman’s λ6 , (see guttman) uses the estimates of
uniqueness u2 from factor analysis to find e2j . This is based on a decomposition of the
variance of a test score, Vx into four parts: that due to a general factor, g, that due to
a set of group factors, f , (factors common to some but not all of the items), specific
factors, s unique to each item, and e, random error. (Because specific variance can not be
distinguished from random error unless the test is given at least twice, some combine these
both into error).
Letting x = cg + A f + Ds + e then the communality of item j , based upon general as well as
group factors, h2j = c2j + ∑ fi2j and the unique variance for the item u2j = σ 2j (1 − h2j ) may be
used to estimate the test reliability. That is, if h2j is the communality of item j , based upon
general as well as group factors, then for standardized items, e2j = 1 − h2j and
It is important to distinguish here between the two ω coefficients of McDonald, 1978 and
Equation 6.20a of McDonald, 1999, ωt and ωh . While the former is based upon the sum of
squared loadings on all the factors, the latter is based upon the sum of the squared loadings
53
on the general factor.
1cc0 1
ωh =
Vx
Another estimate reported is the omega for an infinite length test with a structure similar
to the observed test. This is found by
1cc0 1
ωinf =
1cc0 1 + 1AA0 10
V1
0.5
0.7 V3
0.4
F1*
0.6 0.4
V2
0.6
V4
0.6 0.4
0.4 0.3
V6
0.5
V7
0.4 0.4
F3*
0.4
0.3 V8
0.2
V9
Figure 17: A bifactor solution for 9 simulated variables with a hierarchical structure.
In the case of these simulated 9 variables, the amount of variance attributable to a general
factor (ωh ) is quite large, and the reliability of the set of 9 items is somewhat greater than
that estimated by α or λ6 .
54
> om.9
9 simulated variables
Call: omega(m = r9, title = "9 simulated variables")
Alpha: 0.8
G.6: 0.8
Omega Hierarchical: 0.69
Omega H asymptotic: 0.83
Omega Total 0.83
Compare this with the adequacy of just a general factor and no group factors
The degrees of freedom for just the general factor are 27 and the fit is 0.33
The number of observations was 500 with Chi Square = 162.29 with prob < 2.9e-21
The root mean square of the residuals is 0.08
The df corrected root mean square of the residuals is 0.09
RMSEA index = 0.1 and the 10 % confidence intervals are 0.086 0.115
BIC = -5.5
55
6.3 Estimating ωh using Confirmatory Factor Analysis
The omegaSem function will do an exploratory analysis and then take the highest loading
items on each factor and do a confirmatory factor analysis using the sem package. These
results can produce slightly different estimates of ωh , primarily because cross loadings are
modeled as part of the general factor.
> omegaSem(r9,n.obs=500)
Compare this with the adequacy of just a general factor and no group factors
The degrees of freedom for just the general factor are 27 and the fit is 0.33
The number of observations was 500 with Chi Square = 162.29 with prob < 2.9e-21
The root mean square of the residuals is 0.08
The df corrected root mean square of the residuals is 0.09
RMSEA index = 0.1 and the 10 % confidence intervals are 0.086 0.115
BIC = -5.5
56
g F1* F2* F3*
Correlation of scores with factors 0.84 0.60 0.53 0.60
Multiple R square of scores with factors 0.71 0.37 0.28 0.37
Minimum correlation of factor score estimates 0.43 -0.27 -0.43 -0.27
The degrees of freedom of the confimatory model are 18 and the fit is 27.89682 with p = 0.0636428
general/max 5.51 max/min = 1.41
mean percent general = 0.68 with sd = 0.1 and cv of 0.15
Explained Common Variance of the general factor = 0.68
To get the standard sem fit statistics, ask for summary on the fitted object
57
> splitHalf(r9)
58
for attenuation. As is the case for most of the psych functions, additional information is
returned as part of the object.
First, create keys matrix using the make.keys function. (The keys matrix could also be
prepared externally using a spreadsheet and then copying it into R). Although not normally
necessary, show the keys to understand what is happening.
Note that the number of items to specify in the make.keys function is the total number of
items in the inventory. That is, if scoring just 5 items from a 25 item inventory, make.keys
should be told that there are 25 items. make.keys just changes a list of items on each
scale to make up a scoring matrix. Because the bfi data set has 25 items as well as 3
demographic items, the number of variables is specified as 28.
> keys <- make.keys(nvars=28,list(Agree=c(-1,2:5),Conscientious=c(6:8,-9,-10),
+ Extraversion=c(-11,-12,13:15),Neuroticism=c(16:20),
+ Openness = c(21,-22,23,24,-25)),
+ item.labels=colnames(bfi))
> keys
The use of multiple key matrices for different inventories is facilitated by using the su-
perMatrix function to combine two or more matrices. This allows convenient scoring of
large data sets combining multiple inventories with keys based upon each individual inven-
tory. Pretend for the moment that the big 5 items were made up of two inventories, one
59
consisting of the first 10 items, the second the last 18 items. (15 personality items + 3
demographic items.) Then the following code would work:
> keys.1<- make.keys(10,list(Agree=c(-1,2:5),Conscientious=c(6:8,-9,-10)))
> keys.2 <- make.keys(15,list(Extraversion=c(-1,-2,3:5),Neuroticism=c(6:10),
+ Openness = c(11,-12,13,14,-15)))
> keys.25 <- superMatrix(list(keys.1,keys.2))
The resulting keys matrix is identical to that found above except that it does not include
the extra 3 demographic items. This is useful when scoring the raw items because the
response frequencies for each category are reported, and for the demographic data,
This use of making multiple key matrices and then combining them into one super matrix
of keys is particularly useful when combining demographic information with items to be
scores. A set of demographic keys can be made and then these can be combined with the
keys for the particular scales.
Now use these keys in combination with the raw data to score the items, calculate basic
reliability and intercorrelations, and find the item-by scale correlations for each item and
each scale. By default, missing data are replaced by the median for that variable.
> scores <- scoreItems(keys,bfi)
> scores
(Unstandardized) Alpha:
Agree Conscientious Extraversion Neuroticism Openness
alpha 0.7 0.72 0.76 0.81 0.6
Guttman 6* reliability:
Agree Conscientious Extraversion Neuroticism Openness
Lambda.6 0.7 0.72 0.76 0.81 0.6
60
Conscientious 0.26 0.72 0.35 -0.305 0.30
Extraversion 0.46 0.26 0.76 -0.284 0.32
Neuroticism -0.18 -0.23 -0.22 0.812 -0.12
Openness 0.15 0.19 0.22 -0.086 0.60
In order to see the item by scale loadings and frequency counts of the data
print with the short option = FALSE
To see the additional information (the raw correlations, the individual scores, etc.), they
may be specified by name. Then, to visualize the correlations between the raw scores, use
the pairs.panels function on the scores values of scores. (See figure 18
To find the correlations of the items with each of the scales (the “structure” matrix) or
the correlations of the items controlling for the other scales (the “pattern” matrix), use the
cluster.loadings function. To do both at once (e.g., the correlations of the scales as well
as the item by scale correlations), it is also possible to just use scoreItems.
61
> png('scores.png')
> pairs.panels(scores$scores,pch='.',jiggle=TRUE)
> dev.off()
quartz
2
Figure 18: A graphic analysis of the Big Five scales found by using the scoreItems function.
The pair.wise plot allows us to see that some participants have reached the ceiling of the
scale for these 5 items scales. Using the pch=’.’ option in pairs.panels is recommended when
plotting many cases. The data points were “jittered” by setting jiggle=TRUE. Jiggling this
way shows the density more clearly. To save space, the figure was done as a png. For a
clearer figure, save as a pdf.
62
and scoring the items as correct or incorrect. Both of these operations may be done using
the score.multiple.choice function. Consider the 16 example items taken from an online
ability test at the Personality Project: http://test.personality-project.org. This is
part of the Synthetic Aperture Personality Assessment (SAPA) study discussed in Revelle
et al. (2011, 2010).
> data(iqitems)
> iq.keys <- c(4,4,4, 6,6,3,4,4, 5,2,2,4, 3,2,6,7)
> score.multiple.choice(iq.keys,iqitems)
(Unstandardized) Alpha:
[1] 0.84
item statistics
key 0 1 2 3 4 5 6 7 8 miss r n mean
reason.4 4 0.05 0.05 0.11 0.10 0.64 0.03 0.02 0.00 0.00 0 0.59 1523 0.64
reason.16 4 0.04 0.06 0.08 0.10 0.70 0.01 0.00 0.00 0.00 0 0.53 1524 0.70
reason.17 4 0.05 0.03 0.05 0.03 0.70 0.03 0.11 0.00 0.00 0 0.59 1523 0.70
reason.19 6 0.04 0.02 0.13 0.03 0.06 0.10 0.62 0.00 0.00 0 0.56 1523 0.62
letter.7 6 0.05 0.01 0.05 0.03 0.11 0.14 0.60 0.00 0.00 0 0.58 1524 0.60
letter.33 3 0.06 0.10 0.13 0.57 0.04 0.09 0.02 0.00 0.00 0 0.56 1523 0.57
letter.34 4 0.04 0.09 0.07 0.11 0.61 0.05 0.02 0.00 0.00 0 0.59 1523 0.61
letter.58 4 0.06 0.14 0.09 0.09 0.44 0.16 0.01 0.00 0.00 0 0.58 1525 0.44
matrix.45 5 0.04 0.01 0.06 0.14 0.18 0.53 0.04 0.00 0.00 0 0.51 1523 0.53
matrix.46 2 0.04 0.12 0.55 0.07 0.11 0.06 0.05 0.00 0.00 0 0.52 1524 0.55
matrix.47 2 0.04 0.05 0.61 0.07 0.11 0.06 0.06 0.00 0.00 0 0.55 1523 0.61
matrix.55 4 0.04 0.02 0.18 0.14 0.37 0.07 0.18 0.00 0.00 0 0.45 1524 0.37
rotate.3 3 0.04 0.03 0.04 0.19 0.22 0.15 0.05 0.12 0.15 0 0.51 1523 0.19
rotate.4 2 0.04 0.03 0.21 0.05 0.18 0.04 0.04 0.25 0.15 0 0.56 1523 0.21
rotate.6 6 0.04 0.22 0.02 0.05 0.14 0.05 0.30 0.04 0.14 0 0.55 1523 0.30
rotate.8 7 0.04 0.03 0.21 0.07 0.16 0.05 0.13 0.19 0.13 0 0.48 1524 0.19
sd
reason.4 0.48
reason.16 0.46
reason.17 0.46
reason.19 0.49
letter.7 0.49
letter.33 0.50
letter.34 0.49
letter.58 0.50
matrix.45 0.50
matrix.46 0.50
matrix.47 0.49
matrix.55 0.48
rotate.3 0.40
rotate.4 0.41
rotate.6 0.46
rotate.8 0.39
63
vars n mean sd median trimmed mad min max range skew kurtosis
reason.4 1 1523 0.64 0.48 1 0.68 0 0 1 1 -0.58 -1.66
reason.16 2 1524 0.70 0.46 1 0.75 0 0 1 1 -0.86 -1.26
reason.17 3 1523 0.70 0.46 1 0.75 0 0 1 1 -0.86 -1.26
reason.19 4 1523 0.62 0.49 1 0.64 0 0 1 1 -0.47 -1.78
letter.7 5 1524 0.60 0.49 1 0.62 0 0 1 1 -0.41 -1.84
letter.33 6 1523 0.57 0.50 1 0.59 0 0 1 1 -0.29 -1.92
letter.34 7 1523 0.61 0.49 1 0.64 0 0 1 1 -0.46 -1.79
letter.58 8 1525 0.44 0.50 0 0.43 0 0 1 1 0.23 -1.95
matrix.45 9 1523 0.53 0.50 1 0.53 0 0 1 1 -0.10 -1.99
matrix.46 10 1524 0.55 0.50 1 0.56 0 0 1 1 -0.20 -1.96
matrix.47 11 1523 0.61 0.49 1 0.64 0 0 1 1 -0.47 -1.78
matrix.55 12 1524 0.37 0.48 0 0.34 0 0 1 1 0.52 -1.73
rotate.3 13 1523 0.19 0.40 0 0.12 0 0 1 1 1.55 0.40
rotate.4 14 1523 0.21 0.41 0 0.14 0 0 1 1 1.40 -0.03
rotate.6 15 1523 0.30 0.46 0 0.25 0 0 1 1 0.88 -1.24
rotate.8 16 1524 0.19 0.39 0 0.11 0 0 1 1 1.62 0.63
se
reason.4 0.01
reason.16 0.01
reason.17 0.01
reason.19 0.01
letter.7 0.01
letter.33 0.01
letter.34 0.01
letter.58 0.01
matrix.45 0.01
matrix.46 0.01
matrix.47 0.01
matrix.55 0.01
rotate.3 0.01
rotate.4 0.01
rotate.6 0.01
rotate.8 0.01
Once the items have been scored as true or false (assigned scores of 1 or 0), they made
then be scored into multiple scales using the normal scoreItems function.
64
examine in terms of the probability of various alternatives being endorsed, it is even better
to examine the pattern of these responses as a function of the underlying latent trait or
just the total score. This may be done by using irt.responses (Figure 19).
Dτ λj
δj = q , αj = q (2)
1 − λ j2 1 − λ j2
where D is a scaling factor used when converting to the parameterization of logistic model
and is 1.702 in that case and 1 in the case of the normal ogive model. Thus, in the case of
the normal model, factor loadings (λ j ) and item thresholds (τ) are just
αj δj
λj = q , τj = q .
1 + α 2j 1 + α 2j
Consider 9 dichotomous items representing one factor but differing in their levels of diffi-
culty
> set.seed(17)
> d9 <- sim.irt(9,1000,-2.,2.,mod="normal") #dichotomous items
> test <- irt.fa(d9$items)
> test
65
> data(iqitems)
> iq.keys <- c(4,4,4, 6,6,3,4,4, 5,2,2,4, 3,2,6,7)
> scores <- score.multiple.choice(iq.keys,iqitems,score=TRUE,short=FALSE)
> #note that for speed we can just do this on simple item counts rather than IRT based scores.
> op <- par(mfrow=c(2,2)) #set this to see the output for multiple items
> irt.responses(scores$scores,iqitems[1:4],breaks=11)
reason.4 reason.16
0 2 4 6 0 2 4 6
0.8
0.8
1 3 5 1 3 5
P(response)
P(response)
0.4
0.4
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
theta theta
reason.17 reason.19
0 2 4 6 0 2 4 6
0.8
0.8
1 3 5 1 3 5
P(response)
P(response)
0.4
0.4
0.0
0.0
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
theta theta
Figure 19: The pattern of responses to multiple choice ability items can show that some
items have poor distractors. This may be done by using the the irt.responses function.
A good distractor is one that is negatively related to ability.
66
Item Response Analysis using Factor Analysis
Factor analysis with Call: fa(r = r, nfactors = nfactors, n.obs = n.obs, rotate = rotate,
fm = fm)
Similar analyses can be done for polytomous items such as those of the bfi extraversion
scale:
> data(bfi)
> e.irt <- irt.fa(bfi[11:15])
> e.irt
67
> op <- par(mfrow=c(3,1))
> plot(test,type="ICC")
> plot(test,type="IIC")
> plot(test,type="test")
> op <- par(mfrow=c(1,1))
0.8
V1 V2 V3 V4 V5 V6 V7 V8 V9
0.4
0.0
−3 −2 −1 0 1 2 3
V5
V7
0.8
V1 V4 V6 V8
V3 V9
V2
0.4
0.0
−3 −2 −1 0 1 2 3
Reliability
0.64
3
2
−0.09
1
0
−3 −2 −1 0 1 2 3
Figure 20: A graphic analysis of 9 dichotomous (simulated) items. The top panel shows
the probability of item endorsement as the value of the latent trait increases. Items differ
in their location (difficulty) and discrimination (slope). The middle panel shows the infor-
mation in each item as a function of latent trait level. An item is most informative when
the probability of endorsement is 50%. The lower panel shows the total test information.
These items form a test that is most informative (most accurate) at the middle range of
the latent trait.
68
Factor analysis with Call: fa(r = r, nfactors = nfactors, n.obs = n.obs, rotate = rotate,
fm = fm)
The item information functions show that not all of items are equally good (Figure 21):
These procedures can be generalized to more than one factor by specifying the number of
factors in irt.fa. The plots can be limited to those items with discriminations greater
than some value of cut. An invisible object is returned when plotting the output from
irt.fa that includes the average information for each item that has loadings greater than
cut.
> print(e.info,sort=TRUE)
More extensive IRT packages include the ltm and eRm and should be used for serious Item
Response Theory analysis.
69
> e.info <- plot(e.irt,type="IIC")
−E2
1.2
E4
1.0
0.8
Item Information
−E1
E3
0.6
E5
0.4
0.2
0.0
−3 −2 −1 0 1 2 3
Figure 21: A graphic analysis of 5 extraversion items from the bfi. The curves represent
the amount of information in the item as a function of the latent score for an individual.
That is, each item is maximally discriminating at a different part of the latent continuum.
Print e.info to see the average information for each item.
70
of one factor (Figure 23) and then use that correlation matrix to do an omega analysis to
show the sub-structure of the ability items .
> iq.irt
Factor analysis with Call: fa(r = r, nfactors = nfactors, n.obs = n.obs, rotate = rotate,
fm = fm)
71
> iq.irt <- irt.fa(iq.tf)
rotate.4
0.8
reason.17
reason.4
rotate.6
letter.34 rotate.3
0.6
letter.7
letter.58
Item Information
reason.19
reason.16 rotate.8
matrix.47
letter.33
0.4
matrix.46
matrix.45
matrix.55
0.2
0.0
−3 −2 −1 0 1 2 3
Figure 22: A graphic analysis of 16 ability items sampled from the SAPA project. The
curves represent the amount of information in the item as a function of the latent score
for an individual. That is, each item is maximally discriminating at a different part of
the latent continuum. Print iq.irt to see the average information for each item. Partly
because this is a power test (it is given on the web) and partly because the items have not
been carefully chosen, the items are not very discriminating at the high end of the ability
dimension.
72
> om <- omega(iq.irt$rho,4)
Omega
rotate.3
rotate.4 0.7
0.6 0.6
rotate.8 0.6
0.6 F1*
0.5
rotate.6
0.5
0.6 letter.34
0.5
0.7 letter.7 0.4
0.6 F2*
letter.33 0.4
0.6
0.3
0.6 letter.58
g
0.6 matrix.47
0.7
reason.17 0.5 F3*
0.6 0.3
0.6 reason.4 0.2
0.2
0.6 reason.16
0.5
reason.19
0.5 F4*
0.4 matrix.45 0.8
0.3
matrix.46
matrix.55
Figure 23: An Omega analysis of 16 ability items sampled from the SAPA project. The
items represent a general factor as well as four lower level factors. The analysis is done
using the tetrachoric correlations found in the previous irt.fa analysis. The four matrix
items have some serious problems, which may be seen later when examine the item response
functions.
73
average of the three (classical test theory) fails to consider that the items might differ
in either discrimination or in difficulty. The scoreIrt function applies basic IRT to this
problem.
Consider 1000 randomly generated subjects with scores on 9 true/false items differing in
difficulty. Selectively drop the hardest items for the 1/3 lowest subjects, and the 4 easiest
items for the 1/3 top subjects (this is a crude example of what tailored testing would do).
Then score these subjects:
> v9 <- sim.irt(9,1000,-2.,2.,mod="normal") #dichotomous items
> items <- v9$items
> test <- irt.fa(items)
8 Multilevel modeling
Correlations between individuals who belong to different natural groups (based upon e.g.,
ethnicity, age, gender, college major, or country) reflect an unknown mixture of the pooled
correlation within each group as well as the correlation of the means of these groups.
These two correlations are independent and do not allow inferences from one level (the
group) to the other level (the individual). When examining data at two levels (e.g., the
individual and by some grouping variable), it is useful to find basic descriptive statistics
(means, sds, ns per group, within group correlations) as well as between group statistics
(over all descriptive statistics, and overall between group correlations). Of particular use
is the ability to decompose a matrix of correlations at the individual level into correlations
within group and correlations between groups.
8.1 Decomposing data into within and between level correlations using
statsBy
There are at least two very powerful packages (nlme and multilevel ) which allow for complex
analysis of hierarchical (multilevel) data structures. statsBy is a much simpler function
to give some of the basic descriptive statistics for two level models.
74
Figure 24: IRT based scoring and total test scores for 1000 simulated subjects. True theta
values are reported and then the IRT and total scoring systems.
> pairs.panels(scores.df,pch='.',gap=0)
> title('Comparing true theta for IRT, Rasch and classically based scoring',li
Comparing true theta for IRT, Rasch and classically based scoring
−2 0 1 2 0.4 0.8 1.2 0.0 0.4 0.8
True theta
2
0.83 0.40 0.05 0.55 0.40 −0.19
0
−3
0 1 2
irt theta
● 0.50 0.07 0.68 0.50 −0.20
−2
total
fit
−0.02 0.10 0.82
0.8
● ● ●
0.4
rasch
0.5
● ● ● ● 0.96 −0.14
−1.0
total
0.0 0.4 0.8
● ● ● ● ● −0.01
1.4
fit
1.0
● ● ● ● ● ●
0.6
75
This follows the decomposition of an observed correlation into the pooled correlation within
groups (rwg) and the weighted correlation of the means between groups which is discussed
by Pedhazur (1997) and by Bliese (2009) in the multilevel package.
where rxy is the normal correlation which may be decomposed into a within group and
between group correlations rxywg and rxybg and η (eta) is the correlation of the data with
the within group values, or the group means.
76
function. Set correlation is
n
R2 = 1 − ∏(1 − λi )
i=1
where λi is the ith eigen value of the eigen value decomposition of the matrix
R = R−1 −1 −1
xx Rxy Rxx Rxy .
Unfortunately, there are several cases where set correlation will give results that are much
too high. This will happen if some variables from the first set are highly related to those
in the second set, even though most are not. In this case, although the set correlation
can be very high, the degree of relationship between the sets is not as high. In this
case, an alternative statistic, based upon the average canonical correlation might be more
appropriate.
setCor has the additional feature that it will calculate multiple and partial correlations
from the correlation or covariance matrix rather than the original data.
Consider the correlations of the 6 variables in the sat.act data set. First do the normal
multiple regression, and then compare it with the results using setCor. Two things to
notice. setCor works on the correlation or covariance or raw data matrix, and thus if
using the correlation matrix, will report standardized β̂ weights. Secondly, it is possible to
do several multiple regressions simultaneously. If the number of observations is specified,
or if the analysis is done on raw data, statistical tests of significance are applied.
For this example, the analysis is done on the correlation matrix rather than the raw
data.
> C <- cov(sat.act,use="pairwise")
> model1 <- lm(ACT~ gender + education + age, data=sat.act)
> summary(model1)
Call:
lm(formula = ACT ~ gender + education + age, data = sat.act)
Residuals:
Min 1Q Median 3Q Max
-25.2458 -3.2133 0.7769 3.5921 9.2630
Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) 27.41706 0.82140 33.378 < 2e-16 ***
gender -0.48606 0.37984 -1.280 0.20110
education 0.47890 0.15235 3.143 0.00174 **
age 0.01623 0.02278 0.712 0.47650
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
77
Compare this with the output from setCor.
DV = ACT
intercept = 109.94
slope se t p lower.ci upper.ci VIF
gender -0.26 0.08 -3.17 0.087 -0.61 0.09 1.55
education 0.56 0.07 7.72 0.016 0.25 0.87 1.22
age -0.64 0.08 -8.28 0.014 -0.97 -0.31 1.38
Multiple Regression
R R2 Ruw R2uw Shrunken R2 SE of R2 overall F df1 df2 p SE residual
ACT 1 0.99 0.98 0.96 0.98 0 76.34 3 2 0.013 0.15
DV = SATV
intercept = 3575.8
slope se t p lower.ci upper.ci VIF
gender 0.10 0.09 1.11 0.380 -0.28 0.48 1.55
education 0.72 0.08 9.21 0.012 0.39 1.06 1.22
age -0.82 0.08 -9.76 0.010 -1.18 -0.46 1.38
Multiple Regression
R R2 Ruw R2uw Shrunken R2 SE of R2 overall F df1 df2 p
SATV 0.99 0.99 0.89 0.79 0.97 0 65.18 3 2 0.0151
SE residual
SATV 0.16
DV = SATQ
intercept = 3682.59
slope se t p lower.ci upper.ci VIF
gender -0.52 0.05 -9.73 0.010 -0.76 -0.29 1.55
education 0.40 0.05 8.32 0.014 0.19 0.60 1.22
age -0.47 0.05 -9.24 0.012 -0.69 -0.25 1.38
Multiple Regression
R R2 Ruw R2uw Shrunken R2 SE of R2 overall F df1 df2 p SE residual
SATQ 1 1 1 0.99 0.99 0 177.11 3 2 0.00562 0.1
Note that the setCor analysis also reports the amount of shared variance between the
predictor set and the criterion (dependent) set. This set correlation is symmetric. That is,
78
the R2 is the same independent of the direction of the relationship.
For a much more detailed discussion of setCor see the mediation, moderation and regres-
sion analysis tutorial.
10 Simulation functions
It is particularly helpful, when trying to understand psychometric concepts, to be able
to generate sample data sets that meet certain specifications. By knowing “truth” it is
possible to see how well various algorithms can capture it. Several of the sim functions
create artificial data sets with known structures.
A number of functions in the psych package will generate simulated data. These functions
include sim for a factor simplex, and sim.simplex for a data simplex, sim.circ for a cir-
cumplex structure, sim.congeneric for a one factor factor congeneric model, sim.dichot
to simulate dichotomous items, sim.hierarchical to create a hierarchical factor model,
sim.item is a more general item simulation, sim.minor to simulate major and minor fac-
tors, sim.omega to test various examples of omega, sim.parallel to compare the efficiency
of various ways of determining the number of factors, sim.rasch to create simulated rasch
data, sim.irt to create general 1 to 4 parameter IRT data by calling sim.npl 1 to 4
parameter logistic IRT or sim.npn 1 to 4 paramater normal IRT, sim.structural a gen-
eral simulation of structural models, and sim.anova for ANOVA and lm simulations, and
sim.vss. Some of these functions are separately documented and are listed here for ease
of the help function. See each function for more detailed help.
sim The default version is to generate a four factor simplex structure over three occasions,
although more general models are possible.
sim.simple Create major and minor factors. The default is for 12 variables with 3 major
factors and 6 minor factors.
sim.structure To combine a measurement and structural model into one data matrix.
Useful for understanding structural equation models.
sim.hierarchical To create data with a hierarchical (bifactor) structure.
sim.congeneric To create congeneric items/tests for demonstrating classical test theory.
This is just a special case of sim.structure.
sim.circ To create data with a circumplex structure.
sim.item To create items that either have a simple structure or a circumplex structure.
sim.dichot Create dichotomous item data with a simple or circumplex structure.
sim.rasch Simulate a 1 parameter logistic (Rasch) model.
79
sim.irt Simulate a 2 parameter logistic (2PL) or 2 parameter Normal model. Will also
do 3 and 4 PL and PN models.
sim.multilevel Simulate data with different within group and between group correla-
tional structures.
Some of these functions are described in more detail in the companion vignette: psych for
sem.
The default values for sim.structure is to generate a 4 factor, 12 variable data set with
a simplex structure between the factors.
Two data structures that are particular challenges to exploratory factor analysis are the
simplex structure and the presence of minor factors. Simplex structures sim.simplex will
typically occur in developmental or learning contexts and have a correlation structure of r
between adjacent variables and rn for variables n apart. Although just one latent variable
(r) needs to be estimated, the structure will have nvar-1 factors.
Many simulations of factor structures assume that except for the major factors, all residuals
are normally distributed around 0. An alternative, and perhaps more realistic situation, is
that the there are a few major (big) factors and many minor (small) factors. The challenge
is thus to identify the major factors. sim.minor generates such structures. The structures
generated can be thought of as having a a major factor structure with some small correlated
residuals.
Although coefficient ωh is a very useful indicator of the general factor saturation of a
unifactorial test (one with perhaps several sub factors), it has problems with the case of
multiple, independent factors. In this situation, one of the factors is labelled as “general”
and the omega estimate is too large. This situation may be explored using the sim.omega
function.
The four irt simulations, sim.rasch, sim.irt, sim.npl and sim.npn, simulate dichoto-
mous items following the Item Response model. sim.irt just calls either sim.npl (for
logistic models) or sim.npn (for normal models) depending upon the specification of the
model.
The logistic model is
ζj −γj
P(x|θi , δ j , γ j , ζ j ) = γ j + . (4)
1 + eα j (δ j −θi
where γ is the lower asymptote or guessing parameter, ζ is the upper asymptote (nor-
mally 1), α j is item discrimination and δ j is item difficulty. For the 1 Paramater Logistic
(Rasch) model, gamma=0, zeta=1, alpha=1 and item difficulty is the only free parameter
to specify.
(Graphics of these may be seen in the demonstrations for the logistic function.)
80
The normal model (irt.npn calculates the probability using pnorm instead of the logistic
function used in irt.npl, but the meaning of the parameters are otherwise the same.
With the a = α parameter = 1.702 in the logiistic model the two models are practically
identical.
11 Graphical Displays
Many of the functions in the psych package include graphic output and examples have
been shown in the previous figures. After running fa, iclust, omega, irt.fa, plotting the
resulting object is done by the plot.psych function as well as specific diagram functions.
e.g., (but not shown)
f3 <- fa(Thurstone,3)
plot(f3)
fa.diagram(f3)
c <- iclust(Thurstone)
plot(c) #a pretty boring plot
iclust.diagram(c) #a better diagram
c3 <- iclust(Thurstone,3)
plot(c3) #a more interesting plot
data(bfi)
e.irt <- irt.fa(bfi[11:15])
plot(e.irt)
ot <- omega(Thurstone)
plot(ot)
omega.diagram(ot)
The ability to show path diagrams to represent factor analytic and structural models is dis-
cussed in somewhat more detail in the accompanying vignette, psych for sem. Basic routines
to draw path diagrams are included in the dia.rect and accompanying functions. These
are used by the fa.diagram, structure.diagram and iclust.diagram functions.
12 Miscellaneous functions
A number of functions have been developed for some very specific problems that don’t fit
into any other category. The following is an incomplete list. Look at the Index for psych
for a list of all of the functions.
block.random Creates a block randomized structure for n independent variables. Useful
for teaching block randomization for experimental design.
df2latex is useful for taking tabular output (such as a correlation matrix or that of de-
scribe and converting it to a LATEX table. May be used when Sweave is not conve-
nient.
cor2latex Will format a correlation matrix in APA style in a LATEX table. See also
fa2latex and irt2latex.
81
> xlim=c(0,10)
> ylim=c(0,10)
> plot(NA,xlim=xlim,ylim=ylim,main="Demontration of dia functions",axes=FALSE,xlab="",ylab="")
> ul <- dia.rect(1,9,labels="upper left",xlim=xlim,ylim=ylim)
> ll <- dia.rect(1,3,labels="lower left",xlim=xlim,ylim=ylim)
> lr <- dia.ellipse(9,3,"lower right",xlim=xlim,ylim=ylim)
> ur <- dia.ellipse(9,9,"upper right",xlim=xlim,ylim=ylim)
> ml <- dia.ellipse(3,6,"middle left",xlim=xlim,ylim=ylim)
> mr <- dia.ellipse(7,6,"middle right",xlim=xlim,ylim=ylim)
> bl <- dia.ellipse(1,1,"bottom left",xlim=xlim,ylim=ylim)
> br <- dia.rect(9,1,"bottom right",xlim=xlim,ylim=ylim)
> dia.arrow(from=lr,to=ul,labels="right to left")
> dia.arrow(from=ul,to=ur,labels="left to right")
> dia.curved.arrow(from=lr,to=ll$right,labels ="right to left")
> dia.curved.arrow(to=ur,from=ul$right,labels ="left to right")
> dia.curve(ll$top,ul$bottom,"double") #for rectangles, specify where to point
> dia.curved.arrow(mr,ur,"up") #but for ellipses, just point to it.
> dia.curve(ml,mr,"across")
> dia.arrow(ur,lr,"top down")
> dia.curved.arrow(br$top,lr$bottom,"up")
> dia.curved.arrow(bl,br,"left to right")
> dia.arrow(bl,ll$bottom)
> dia.curved.arrow(ml,ll$right)
> dia.curved.arrow(mr,lr$top)
left to right
upper left left to right upper right
up
across
middle left right to left middle right top down
double
right to left
lower left lower right
left to right up
82
Figure 25: The basic graphic capabilities of the dia functions are shown in this figure.
cosinor One of several functions for doing circular statistics. This is important when
studying mood effects over the day which show a diurnal pattern. See also circa-
dian.mean, circadian.cor and circadian.linear.cor for finding circular means,
circular correlations, and correlations of circular with linear data.
fisherz Convert a correlation to the corresponding Fisher z score.
geometric.mean also harmonic.mean find the appropriate mean for working with different
kinds of data.
ICC and cohen.kappa are typically used to find the reliability for raters.
headtail combines the head and tail functions to show the first and last lines of a data
set or output.
topBottom Same as headtail. Combines the head and tail functions to show the first and
last lines of a data set or output, but does not add ellipsis between.
mardia calculates univariate or multivariate (Mardia’s test) skew and kurtosis for a vector,
matrix, or data.frame
p.rep finds the probability of replication for an F, t, or r and estimate effect size.
partial.r partials a y set of variables out of an x set and finds the resulting partial
correlations. (See also setCor.)
rangeCorrection will correct correlations for restriction of range.
reverse.code will reverse code specified items. Done more conveniently in most psych
functions, but supplied here as a helper function when using other packages.
superMatrix Takes two or more matrices, e.g., A and B, and combines them into a “Super
matrix” with A on the top left, B on the lower right, and 0s for the other two
quadrants. A useful trick when forming complex keys, or when forming example
problems.
13 Data sets
A number of data sets for demonstrating psychometric techniques are included in the
psych package. These include six data sets showing a hierarchical factor structure (five
cognitive examples, Thurstone, Thurstone.33, Holzinger, Bechtoldt.1, Bechtoldt.2,
and one from health psychology Reise). One of these (Thurstone) is used as an example
in the sem package as well as McDonald (1999). The original data are from Thurstone and
Thurstone (1941) and reanalyzed by Bechtoldt (1961). Personality item data representing
five personality factors on 25 items (bfi) or 13 personality inventory scores (epi.bfi), and
14 multiple choice iq items (iqitems). The vegetables example has paired comparison
83
preferences for 9 vegetables. This is an example of Thurstonian scaling used by Guilford
(1954) and Nunnally (1967). Other data sets include cubits, peas, and heights from
Galton.
Thurstone Holzinger-Swineford (1937) introduced the bifactor model of a general factor
and uncorrelated group factors. The Holzinger correlation matrix is a 14 * 14 matrix
from their paper. The Thurstone correlation matrix is a 9 * 9 matrix of correlations
of ability items. The Reise data set is 16 * 16 correlation matrix of mental health
items. The Bechtholdt data sets are both 17 x 17 correlation matrices of ability tests.
bfi 25 personality self report items taken from the International Personality Item Pool
(ipip.ori.org) were included as part of the Synthetic Aperture Personality Assessment
(SAPA) web based personality assessment project. The data from 2800 subjects are
included here as a demonstration set for scale construction, factor analysis and Item
Response Theory analyses.
sat.act Self reported scores on the SAT Verbal, SAT Quantitative and ACT were col-
lected as part of the Synthetic Aperture Personality Assessment (SAPA) web based
personality assessment project. Age, gender, and education are also reported. The
data from 700 subjects are included here as a demonstration set for correlation and
analysis.
epi.bfi A small data set of 5 scales from the Eysenck Personality Inventory, 5 from a Big 5
inventory, a Beck Depression Inventory, and State and Trait Anxiety measures. Used
for demonstrations of correlations, regressions, graphic displays.
iq 14 multiple choice ability items were included as part of the Synthetic Aperture Person-
ality Assessment (SAPA) web based personality assessment project. The data from
1000 subjects are included here as a demonstration set for scoring multiple choice
inventories and doing basic item statistics.
galton Two of the earliest examples of the correlation coefficient were Francis Galton’s
data sets on the relationship between mid parent and child height and the similarity of
parent generation peas with child peas. galton is the data set for the Galton height.
peas is the data set Francis Galton used to ntroduce the correlation coefficient with
an analysis of the similarities of the parent and child generation of 700 sweet peas.
Dwyer Dwyer (1937) introduced a method for factor extension (see fa.extension that
finds loadings on factors from an original data set for additional (extended) variables.
This data set includes his example.
miscellaneous cities is a matrix of airline distances between 11 US cities and may be
used for demonstrating multiple dimensional scaling. vegetables is a classic data set
for demonstrating Thurstonian scaling and is the preference matrix of 9 vegetables
84
from Guilford (1954). Used by Guilford (1954); Nunnally (1967); ?, this data set
allows for examples of basic scaling techniques.
Although the individual help pages for the psych package are available as part of R and
may be accessed directly (e.g. ?psych) , the full manual for the psych package is also
available as a pdf at http://personality-project.org/r/psych_manual.pdf
News and a history of changes are available in the NEWS and CHANGES files in the source
files. To view the most recent news,
R code
news(Version > "1.2.8",package="psych")
15 Psychometric Theory
The psych package has been developed to help psychologists do basic research. Many of
the functions were developed to supplement a book (http://personality-project.org/
r/book An introduction to Psychometric Theory with Applications in R (Revelle, prep)
More information about the use of some of the functions may be found in the book .
For more extensive discussion of the use of psych in particular and R in general, consult
http://personality-project.org/r/r.guide.html A short guide to R.
16 SessionInfo
This document was prepared using the following settings.
> sessionInfo()
locale:
[1] C
85
other attached packages:
[1] psychTools_1.9.5.26 psych_1.9.5
86
References
Bechtoldt, H. (1961). An empirical study of the factor analysis stability hypothesis. Psy-
chometrika, 26(4):405–432.
Blashfield, R. K. (1980). The growth of cluster analysis: Tryon, Ward, and Johnson.
Multivariate Behavioral Research, 15(4):439 – 458.
Blashfield, R. K. and Aldenderfer, M. S. (1988). The methods and problems of cluster
analysis. In Nesselroade, J. R. and Cattell, R. B., editors, Handbook of multivariate
experimental psychology (2nd ed.), pages 447–473. Plenum Press, New York, NY.
Bliese, P. D. (2009). Multilevel Modeling in R (2.3) A Brief Introduction to R, the multilevel
package and the nlme package.
Cattell, R. B. (1966). The scree test for the number of factors. Multivariate Behavioral
Research, 1(2):245–276.
Cattell, R. B. (1978). The scientific use of factor analysis. Plenum Press, New York.
Cohen, J. (1982). Set correlation as a general multivariate data-analytic method. Multi-
variate Behavioral Research, 17(3).
Cohen, J., Cohen, P., West, S. G., and Aiken, L. S. (2003). Applied multiple regres-
sion/correlation analysis for the behavioral sciences. L. Erlbaum Associates, Mahwah,
N.J., 3rd ed edition.
Cooksey, R. and Soutar, G. (2006). Coefficient beta and hierarchical item clustering - an
analytical procedure for establishing and displaying the dimensionality and homogeneity
of summated scales. Organizational Research Methods, 9:78–98.
Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika,
16:297–334.
Dwyer, P. S. (1937). The determination of the factor loadings of a given test from the
known factor loadings of other tests. Psychometrika, 2(3):173–178.
Everitt, B. (1974). Cluster analysis. John Wiley & Sons, Cluster analysis. 122 pp. Oxford,
England.
Goldberg, L. R. (2006). Doing it all bass-ackwards: The development of hierarchical factor
structures from the top down. Journal of Research in Personality, 40(4):347 – 358.
Grice, J. W. (2001). Computing and evaluating factor scores. Psychological Methods,
6(4):430–450.
Guilford, J. P. (1954). Psychometric Methods. McGraw-Hill, New York, 2nd edition.
87
Guttman, L. (1945). A basis for analyzing test-retest reliability. Psychometrika, 10(4):255–
282.
Hartigan, J. A. (1975). Clustering Algorithms. John Wiley & Sons, Inc., New York, NY,
USA.
Henry, D. B., Tolan, P. H., and Gorman-Smith, D. (2005). Cluster analysis in family
psychology research. Journal of Family Psychology, 19(1):121–132.
Holzinger, K. and Swineford, F. (1937). The bi-factor method. Psychometrika, 2(1):41–54.
Horn, J. L. (1965). A rationale and test for the number of factors in factor analysis.
Psychometrika, 30(2):179–185.
Horn, J. L. and Engstrom, R. (1979). Cattell’s scree test in relation to Bartlett’s chi-square
test and other observations on the number of factors problem. Multivariate Behavioral
Research, 14(3):283–300.
Jennrich, R. and Bentler, P. (2011). Exploratory bi-factor analysis. Psychometrika,
76(4):537–549.
Jensen, A. R. and Weng, L.-J. (1994). What is a good g? Intelligence, 18(3):231–258.
Loevinger, J., Gleser, G., and DuBois, P. (1953). Maximizing the discriminating power of
a multiple-score test. Psychometrika, 18(4):309–317.
MacCallum, R. C., Browne, M. W., and Cai, L. (2007). Factor analysis models as ap-
proximations. In Cudeck, R. and MacCallum, R. C., editors, Factor analysis at 100:
Historical developments and future directions, pages 153–175. Lawrence Erlbaum Asso-
ciates Publishers, Mahwah, NJ.
Martinent, G. and Ferrand, C. (2007). A cluster analysis of precompetitive anxiety: Re-
lationship with perfectionism and trait anxiety. Personality and Individual Differences,
43(7):1676–1686.
McDonald, R. P. (1999). Test theory: A unified treatment. L. Erlbaum Associates, Mahwah,
N.J.
Mun, E. Y., von Eye, A., Bates, M. E., and Vaschillo, E. G. (2008). Finding groups
using model-based cluster analysis: Heterogeneous emotional self-regulatory processes
and heavy alcohol use risk. Developmental Psychology, 44(2):481–495.
Nunnally, J. C. (1967). Psychometric theory. McGraw-Hill, New York,.
Pedhazur, E. (1997). Multiple regression in behavioral research: explanation and prediction.
Harcourt Brace College Publishers.
88
R Core Team (2019). R: A Language and Environment for Statistical Computing. R
Foundation for Statistical Computing, Vienna, Austria.
Revelle, W. (1979). Hierarchical cluster-analysis and the internal structure of tests. Mul-
tivariate Behavioral Research, 14(1):57–74.
Revelle, W. (2019). psych: Procedures for Personality and Psychological Research. North-
western University, Evanston, https://CRAN.r-project.org/package=psych. R package
version 1.9.4.
Revelle, W. (in prep). An introduction to psychometric theory with applications in R.
Springer.
Revelle, W., Condon, D., and Wilt, J. (2011). Methodological advances in differential
psychology. In Chamorro-Premuzic, T., Furnham, A., and von Stumm, S., editors,
Handbook of Individual Differences, chapter 2, pages 39–73. Wiley-Blackwell.
Revelle, W. and Rocklin, T. (1979). Very Simple Structure - alternative procedure for
estimating the optimal number of interpretable factors. Multivariate Behavioral Research,
14(4):403–414.
Revelle, W., Wilt, J., and Rosenthal, A. (2010). Individual differences in cognition: New
methods for examining the personality-cognition link. In Gruszka, A., Matthews, G.,
and Szymura, B., editors, Handbook of Individual Differences in Cognition: Attention,
Memory and Executive Control, chapter 2, pages 27–49. Springer, New York, N.Y.
Revelle, W. and Zinbarg, R. E. (2009). Coefficients alpha, beta, omega and the glb:
comments on Sijtsma. Psychometrika, 74(1):145–154.
Schmid, J. J. and Leiman, J. M. (1957). The development of hierarchical factor solutions.
Psychometrika, 22(1):83–90.
Shrout, P. E. and Fleiss, J. L. (1979). Intraclass correlations: Uses in assessing rater
reliability. Psychological Bulletin, 86(2):420–428.
Sneath, P. H. A. and Sokal, R. R. (1973). Numerical taxonomy: the principles and practice
of numerical classification. A Series of books in biology. W. H. Freeman, San Francisco.
Sokal, R. R. and Sneath, P. H. A. (1963). Principles of numerical taxonomy. A Series of
books in biology. W. H. Freeman, San Francisco.
Spearman, C. (1904). The proof and measurement of association between two things. The
American Journal of Psychology, 15(1):72–101.
ten Berge, J. M., Krijnen, W. P., Wansbeek, T., and Shapiro, A. (1999). Some new
results on correlation-preserving factor scores prediction methods. Linear Algebra and
its Applications, 289(1-3):311 – 318.
89
Thorburn, W. M. (1918). The myth of Occam’s razor. Mind, 27:345–353.
Thurstone, L. L. and Thurstone, T. G. (1941). Factorial studies of intelligence. The
University of Chicago press, Chicago, Ill.
Tryon, R. C. (1935). A theory of psychological components–an alternative to ”mathematical
factors.”. Psychological Review, 42(5):425–454.
Tryon, R. C. (1939). Cluster analysis. Edwards Brothers, Ann Arbor, Michigan.
Velicer, W. (1976). Determining the number of components from the matrix of partial
correlations. Psychometrika, 41(3):321–327.
Zinbarg, R. E., Revelle, W., Yovel, I., and Li, W. (2005). Cronbach’s α, Revelle’s β , and
McDonald’s ωH : Their relations with each other and two alternative conceptualizations
of reliability. Psychometrika, 70(1):123–133.
Zinbarg, R. E., Yovel, I., Revelle, W., and McDonald, R. P. (2006). Estimating gener-
alizability to a latent variable common to all of a scale’s indicators: A comparison of
estimators for ωh . Applied Psychological Measurement, 30(2):121–144.
90