Chapter 5 Reliability
Chapter 5 Reliability
Chapter 5 Reliability
RELIABILITY
• Reliability - is a synonym for dependability or
consistency.
• For example, a reliable friend who is always there for us
in a time of need.
RELIABILITY
• Reliability - refers to the consistency in measurement
• It is the ability of the test to give consistent results
• Refers to the proportion of the total variance attributed
to true variance. The greater the proportion of the total
variance attributed to true variance, the more reliable
the test.
What causes
Inconsistency?
CONCEPT OF RELIABILITY
• A score on an ability test is presumed to reflect not
only the testtaker’s true score on the ability being
measured but also error.
• Error refers to the component of the observed test
score that does not have to do with the testtaker’s
ability.
• Observed Score = True Score + Error
CONCEPT OF RELIABILITY
• A statistic useful in describing sources of test score
variability is the variance —the standard deviation
squared.
• Variance from true differences is true variance, and
variance from irrelevant, random sources is error
variance.
• Total variance = True Variance + Error Variance
CONCEPT OF RELIABILITY
• Error variance represents any condition that is
irrelevant to the purpose of the test.
• It is reduced by controlling the test environment,
instructions, time limit, rapport, etc.
• A systematic error source does not change the
variability of the distribution or affect reliability.
• For example, a weighing scale consistently
under weighed everyone who stepped on it by 5
pounds, then the relative standings of the
people would remain unchanged.
SOURCES OF ERROR VARIANCE
Test Construction
• Item sampling or content sampling, or the
variation among items within a test, as well as to
variation among items between tests.
• The extent to which a testtaker’s score is affected
by the content sampled on the test and by the way
the content is sampled (that is, the way in which
the item is constructed) is a source of error
variance.
SOURCES OF ERROR VARIANCE
Test Administration (Testing procedures)
• Test environment
• room temperature, level of lighting, ventilation,
changes in weather, broken pencil point, and noise
• Test-Taker variables
• emotional problems, physical discomfort, lack of
sleep, illness, fatigue, drugs or medications taken,
and worry
• Examiner-related variables
• physical appearance and demeanor, manner of
speaking, emphasis on certain words (unknowingly
providing clues), eye nodding, other nonverbal
gestures.
SOURCES OF ERROR VARIANCE
Test Scoring and Interpretation (Scoring System)
• Hand-scoring versus machine scoring
• Objective versus subjective scoring (projective)
• In some tests of personality, examinees are
asked to supply open-ended responses to
stimuli such as pictures, words, sentences, and
inkblots, and it is the examiner who must then
quantify or qualitatively evaluate responses.
TEST-RETEST RELIABILITY ESTIMATES
• One way of estimating the reliability of a measuring
instrument is by using the same instrument to
measure the same thing at two points in time.
TEST-RETEST RELIABILITY ESTIMATES
• Test-retest reliability is an estimate of reliability
obtained by correlating pairs of scores from the
same people on two different administrations of
the same test.
• The test-retest measure is appropriate when
evaluating the reliability of a test that purports to
measure something that is relatively stable over
time, such as a personality trait.
• Consider possible intervening factors between
test administrations to attain proper conclusions
about the reliability of the measuring
instrument
TEST-RETEST RELIABILITY ESTIMATES
• An estimate of test-retest reliability may be most
appropriate in gauging the reliability of tests that
employ outcome measures such as reaction time
or perceptual judgments (including discriminations
of brightness, loudness, or taste).
• This may still be confounded due to experience,
practice, memory, fatigue, and motivation)
may intervene and confound an obtained
measure of reliability.
PARALLEL-FORMS AND ALTERNATE-FORMS
RELIABILITY ESTIMATES
*Make up Exams
• Parallel forms of a test exist when, for each form of
the test, the means and the variances of observed
test scores are equal.
• More practically, scores obtained on parallel tests
correlate equally with other measures.
PARALLEL-FORMS AND ALTERNATE-FORMS
RELIABILITY ESTIMATES
• Alternate forms are simply different versions of a
test that have been constructed so as to be
parallel.
• Although they do not meet the requirements for
the legitimate designation “parallel,” alternate
forms of a test are typically designed to be
equivalent with respect to variables such as
content and level of difficulty.
PARALLEL-FORMS AND ALTERNATE-FORMS
RELIABILITY ESTIMATES
1. Two test administrations with the same group are
required
2. Test scores may be affected by factors such as
motivation, fatigue, or intervening events such as
practice, learning, or therapy (although not as
much as when the same test is administered
twice).
PARALLEL-FORMS AND ALTERNATE-FORMS
RELIABILITY ESTIMATES
• An additional source of error variance, item
sampling, is inherent in the computation of an
alternate- or parallel-forms reliability coefficient.
• Testtakers may do better or worse on a specific
form of the test not as a function of their true
ability but simply because of the particular items
that were selected for inclusion in the test.
• Developing alternate forms of tests can be time-
consuming and expensive.
PARALLEL-FORMS AND ALTERNATE-FORMS
RELIABILITY ESTIMATES
• An estimate of the reliability of a test can be
obtained without developing an alternate form of
the test and without having to administer the test
twice to the same people.
• internal consistency estimate of reliability or as an
estimate of inter-item consistency.
SPLIT-HALF RELIABILITY ESTIMATES