Methods and Stats in I/O: - Science - Research - Data Analysis - Correlation and Regression - Psychometrics
Methods and Stats in I/O: - Science - Research - Data Analysis - Correlation and Regression - Psychometrics
• Science
• Research
• Data Analysis
• Descriptive and Inferential
• Correlation and Regression
• Psychometrics
• Psychometrics
• Study of how people respond to tests and questionnaires
• Measurement
• System of rules of assigning numbers to represent a person’s
standing on some underlying characteristic
• Latent Variable
• Theoretical variable of interest
• Cannot be observed directly
• Measure
• Operational definition of construct
• Imperfect indicator of latent variable
Applications of Psychometrics
• Validity
• Measurement Precision
• Reliability
• Understanding error
• Scale Development
• Item selection
• The attribute must exist and have a causal impact on test scores
• Criterion-Related
• Correlation of test scores (predictor) with job performance (criterion)
• Represented as the validity coefficient (i.e., a correlation)
Criterion-Related Validity
• Predictive Validity
• Predictors correlate with criterion separated by time
• Concurrent Validity
• Predictors correlate with criterion at the same time
Predicting Preference to Work Alone
Content-Related Validity
• The content of the predictor and criterion represent an adequate
sample of important work behaviors and KSAOs defined by the job
analysis
• Construct Validity
• The integration of validity evidence which is important for determining the
meaning of test scores
• Correlation between similar and dissimilar tests should be in the predicted
direction (and sometimes strength)
• Evidence from other sources (literature reviews, studies, theories, etc.)
Multi-Trait Multi-Method Matrix
1 2 3 4 5
1. Verbal Ability Test
2. Interview Rating of .5
Communication
3. Sample Lecture .4 .6
4. Test of I/O Knowledge .2 .1 .1
5. Interview Rating of I/O .1 .3 .1 .4
Knowledge
6. Number of Top Tier Pubs .1 .1 .1 .5 .4
Variable F1 F2
1. Verbal Ability Test .8 .1
2. Interview Rating of .6 .2
Communication
3. Sample Lecture .7 .2
4. Test of I/O Knowledge .1 .8
5. Interview Rating of I/O .4 .5
Knowledge
6. Number of Top Tier Pubs .2 .7
Measurement Theory
• Classical Test Theory
• More precisely,
• It indicates the fraction of observed variance that is systematic, as opposed to
random
• In CTT, reliability is the squared correlation between true and observed scores
Types of Reliability
• Test-Retest
• Equivalent Forms
• Internal Consistency
• Inter-Rater
Test-Retest Reliability
• Calculated by correlating measurements taken at time 1 with
measurements taken at time 2
• Represented as a correlation coefficient
• Higher the correlation, higher the reliability
Equivalent Forms Reliability
• Calculated by correlating measurements from a sample of individuals
who complete two different forms of the same test
Reliable
Neither Valid but not
nor Reliable Valid
• Applications include
• Study of item bias
• Creating equivalent forms
• Computer adaptive testing
Item Information
• Item 1 (a = 2, b=-1)
• Item 2 (a = 2, b=-0.5)
• Item 3 (a = 1, b=1)
• Item 4 (a = 1.5, b=2)
Test Information
• The information provided by a set of items is simply
the sum of the Item Information
𝑇𝐼 𝜃 = 𝐼𝑖 𝜃
𝑖
• Item 1 and 4
Which item adds the most?
• Start with Items 1 & 4
• If we add Item 2
• If we add Item 3
Computerized Adaptive Testing
• CATs are used in many popular tests today
• SAT, ACT, GRE
• In CAT, we choose the next item which we hope in some way supplies
us with the most information about the individual’s trait
Intro To CAT
1. Pick an initial item
2. Based on response estimate θ
3. Using current θ, select item with max I(θ)
4. Base on response, update θ estimate
5. Check stopping rule
• e.g., stop if SE < .3
6. If stopping rule is not met, repeat 3-5
Example: Prior Distribution
Item 1 (a = 1, b= 0)
Item 2 (a = 1.5, b= 2)
Item 3 (a = 2, b= -1)
Item 4 (a = 2, b= 1)
Prior
Probability
0.10
0.00
-4 -2 0 2 4
Theta
Item 1 Correct
Item 1 (a = 1, b= 0)
Item 2 (a = 1.5, b= 2)
Item 3 (a = 2, b= -1)
Item 4 (a = 2, b= 1)
Item 1
Probability
0.10
0.00
-4 -2 0 2 4
Theta
Item 2 Wrong
Item 1 (a = 1, b= 0)
Item 2 (a = 1.5, b= 2)
Item 3 (a = 2, b= -1)
Item 4 (a = 2, b= 1)
Item 2
Probability
0.10
0.00
-4 -2 0 2 4
Theta
Item 3 Correct
Item 1 (a = 1, b= 0)
Item 2 (a = 1.5, b= 2)
Item 3 (a = 2, b= -1)
Item 4 (a = 2, b= 1)
Item 3
Probability
0.10
0.00
-4 -2 0 2 4
Theta
Item 4 Wrong
Item 1 (a = 1, b= 0)
Item 2 (a = 1.5, b= 2)
Item 3 (a = 2, b= -1)
Item 4 (a = 2, b= 1)
Item 4
Probability
0.10
0.00
-4 -2 0 2 4
Theta
Research on PROMIS
• Our team at IIT was able to improve the PROMIS CAT by reducing the
number of items by 50% and making the CAT more efficient in general