0% found this document useful (0 votes)
8 views42 pages

Lecture 4 - CH 4

The lecture discusses the fundamentals of psychological testing, focusing on the assumptions that underlie the measurement of psychological traits and states, including their existence, quantifiability, and the limitations of tests. It emphasizes the importance of reliability and validity in creating good tests, as well as the role of norms in evaluating test scores. Additionally, the lecture covers the differences between norm-referenced and criterion-referenced evaluations and the impact of cultural considerations in psychological assessment.

Uploaded by

delil
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views42 pages

Lecture 4 - CH 4

The lecture discusses the fundamentals of psychological testing, focusing on the assumptions that underlie the measurement of psychological traits and states, including their existence, quantifiability, and the limitations of tests. It emphasizes the importance of reliability and validity in creating good tests, as well as the role of norms in evaluating test scores. Additionally, the lecture covers the differences between norm-referenced and criterion-referenced evaluations and the impact of cultural considerations in psychological assessment.

Uploaded by

delil
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 42

PSYC 303:

MEASUREMENT AND
EVALUATION IN
PSYCHOLOGY
LECTURE 4: Of Tests and Testing: Basics and Assumptions
MEYMUNE N. TOPÇU, PhD
Recap from Lecture 2
Measures of Skewness -
Variability Kurtosis

Standard Scores

Measures of
central tendency
Correlation

Levels of
Meta-Analysis
Measurement
 Assumptions about psychological
testing and assessment
 What is a good test?
 Reliability
 Validity

Lecture Plan  Norms


 Sampling to develop norms
 Types of norms
 Fixed reference groups scoring systems
 Norm- vs. criterion-referenced Evaluation
 Culture and Inference
What is a “good test”?
Assumption 1: Psychological traits and states
exist
 Components of stability and change in our
Assumptions behavior
about  Trait: A long-term characteristic of an individual
psychological that shows through their behavior, actions, and
testing and feelings
assessment  Based on a sample of behavior
 Intelligence, cognitive style, interests,
personality
 State: A temporary condition that an individual
is experiencing for a short period of time
 Any examples?
https://youtu.be/n4b_Mj7GndQ?si=RoXx7GIv3Zk29Pix
Why is it important to distinguish
between traits and states in
psychological testing & assessment?
 How do traits exist? Do they have a physical
experience?
 A psychological trait exists as a construct
 Construct: An informed, scientific concept developed or
constructed to describe or explain behavior
 The construct’s existence can be inferred from overt
behavior
 Observable action or the product of an observable action

 Trait is not expected to be manifested 100% of the


time
 What determines if a trait will be manifested or not?
 The strength of the trait & the nature of the situation
 Situation-dependent, E.g., American football -
Playground
 https://www.mindgarden.com/145-state-trait-anxiety-
inventory-for-adults
 Attributions of a trait or state term are relative
 E.g., “Özge is very shy” – an unstated comparison with degree of shyness in average person
 The reference group can greatly influence one’s conclusions or judgments
 Measuring sensation seeking (the need for varied novel, complex sensations)
 Sensation seeking scale vs. performance-based measures
Assumption 2: Psychological traits and states can
be quantified and measured
 If psychological traits and states vary by degree
Assumptions they are quantifiable
about  Defining the trait/state
psychological  The same phenomenon can be defined in different
testing and ways
 E.g., “Aggressive salesperson”, “Aggressive killer”,
assessment “Aggressive waiter”
 How aggressiveness is defined by the test
developer
 “The number of self-reported acts of harming others”
 “The number of observed acts of aggression”
 The test developer should provide a clear
“operational definition”
 After defining a trait/state a test-developer
considers the types of item content
 Components of intelligence in US adults
 If knowledge of American history: “Who was the
second president of the US”
 If social judgment: “Why should guns in the
home always be inaccessible to children?”
 Should all items have equal weight?
 The social judgment item could be given more
weight
 Developing appropriate ways to score/interpret
 Cumulative scoring: A trait is measured by a
series of test items
Assumption 3: Test-related behavior predicts non-
test-related behavior
 The obtained sample of behavior is typically used
to make predictions about future behavior
Assumptions  E.g., Predicting success in life from intelligence scores

about obtained in childhood

psychological  To postdict behavior: Understanding of behavior


that has already taken place. E.g., Criminal’s state
testing and of mind
assessment
Assumptions
about
psychological
testing and
assessment Assumption 4: All tests have limits and
imperfections
 Why?
 Test users should understand the limitations of
tests and how those limitations can be
compensated for by data from other resources
Assumption 5: Various sources of error are part of the
assessment process
 Error: Factors other than what a test attempts to
Assumptions measure will influence performance on test

about  Does an intelligence test score truly reflect


psychological intelligence or factors other than intelligence?

testing and  Error variance: The component of a test score

assessment attributable to scores other than the trait or ability


measured
 Assessees, Assessors, Instruments can all be sources
of error variance
 Random errors: Errors that happen as a matter of
chance
 E.g., the weather on the day of testing

 Error is an element in the process of measurement


Assumption 6: Unfair and biased
assessment procedures can be identified
and reformed
 Sophisticated procedures to identify and
correct test bias and list of ethical
Assumptions guidelines to ensure test fairness
about  Fairness-related questions and problems
psychological can still arise
testing and  E.g., The test is used with a person whose
assessment background/experience is different from
the group the test was intended for
 Tests are tools and they can be used
properly or improperly
Assumption 7: Testing and assessment offer
powerful benefits to society
 Imagine a world without psychological
Assumptions tests/assessments. How would it be?
about  In a world without tests…
psychological  People can easily trick others that they are a
testing and surgeon
assessment  Personnel might be hired on the basis of nepotism
rather than documented merit
 It would be very difficult to offer treatments for
educational difficulties
 The military/business sector would not have a tool
to screen applicants
 We need good tests…
 Psychometric soundness of a test
 A good test/measuring tool is reliable
 In theory, the perfectly reliable measuring
tool consistently measures in the same way
What is a “good
test”? 1 Kg

Reliability

1 Kg 1.3 Kg 1.2 Kg
1 Kg 1.3 Kg 0.9 Kg
1 Kg 1.3 Kg 1 Kg
Why is it more difficult to achieve
perfect reliability for psychological
tests?
How does calculating reliability
differ when you are measuring a
trait vs. a state?
 Psychometric soundness of a test
 A valid test measures what it claims to
measure
 E.g., Intelligence

What is a “good  Items that make up a test adequately sample


test”? the range of areas that must be sampled to
adequately measure the construct

Validity  How are the scores interpreted? How do


scores on this test relate to other scores
measuring the same/opposite construct?
 E.g., A valid test of introversion should be
negatively correlated with a valid test of
extraversion
 A good test is one that trained examiners can
administer, score, and interpret with a
minimum of difficulty

What is a “good  A good test is a useful test, one that yields


actionable results that will ultimately benefit
test”?
individual test takers or society at large
 If the purpose of a test is to compare the
Other Considerations
performance of the test taker with the
performance of other test takers, then a
“good test” is one that contains adequate
norms
Why choose one test over the other?
 What is the objective of using a test? how
does the test meet that objective?
 How is the construct defined?
 Who is the test designed for use with? (age,
gender, reading level etc.)
 How appropriate is it for the targeted test
takers?
 What type of data will be generated from
using this test?
 Will there be a need for other assessment
tools?
 Does the test require an expert test user?
Norms  Norm-referenced testing and assessment: a
method of evaluation and a way of deriving
meaning from test scores by evaluating an
individual test taker’s score and comparing it
to scores of a group of test takers
 Aim: To understand where the test taker
stands among other test takers
 Norm (singular): Behavior that is
usual/average/normal
 Norms (Plural): the test performance data of
a particular group of test takers that are
designed for use as a reference
 Normative sample: The group of people
Norms whose performance on a particular test is
analyzed
 The data may be in raw or converted scores
 To norm (verb): Refers to the process of
deriving norms
 Norming a test is expensive
 User norms: Descriptive statistics based on a
group of test takers in a given period of time
Sampling to  Standardization: The process of administering
develop norms a test to a representative sample
 Population: The complete set of individuals
with at least one common observable
characteristic
 Sample of the population: a portion of the
universe of people deemed to be
representative of the whole population
 Sampling: The process of selecting a
representative group of people
 Subgroups in a population may differ in terms
of certain characteristics. It can be essential
to have these differences proportionately
represented
 Stratified sampling help prevent sampling
bias and aid in the interpretation of results
Sampling to
 Stratified-random sampling: When every
develop norms member of the population had the same
chance of being included in the sample
 Purposive sampling: Arbitrarily select some sample
because we believe it to be representative of the
population
 The prob: The sample may no longer be
representative
 Decision of sampling: Comparing what is ideal and
what is practical
Sampling to  Incidental/Convenience sampling: employ a sample
develop norms that is not necessarily the most appropriate but is
simply the most convenient
 Budgetary or other limitations
 E.g., PSYC 101 samples
 Exclusionary criteria
 People with uncorrected vision impairment
 People taking medicine that can affect performance
 People who are not fluent in English etc.
Recall your own experience as a research
subject. How appropriate was it for the
researcher to use students as convenient
sample?
 Developing norms for a standardized
test
 Administering the test to the sample
 Standard set of instructions
 Recommended settings
 Summarizing the data using
descriptive statistics (?)
 Test developers provide information
to support recommended
interpretations of the results: the
nature of the content,
norms/comparison groups, other
technical evidence
 Percentile norms: the raw data from a test’s
Types of Norms standardization sample converted to percentile form
 Dividing the distribution into 100 equal parts
Percentile  Percentile: an expression of the percentage of
people whose score on a test or measure falls below
a particular raw score.
 Percentage correct: What proportion of the items did
the test taker got correct
 GRE example
 What might be a problem of using percentiles?
 With normally distributed scores real differences between raw scores may be minimized near the ends
of the distribution and exaggerated in the middle of the distribution
 Highest frequency of raw scores are in the middle –even smallest differences will appear large in
percentiles
 For the tails – Differences between raw scores may be great with very small percentile differences
Types of Norms  Age equivalent norms: Average performance of different
samples of test takers who were at various ages at time
of test
Age Norms  Carefully constructed age norms of physical
characteristics is OK.
 For psychological characteristics it is tricky
 Identifying the mental age according to intelligence test.
 Problem with it: E.g., Young Sheldon
 Technical ground: SD can be different for different ages
 Grade Norms: Designed to indicate the average
Types of Norms test performance of test takers in a given school
grade
Grade Norms  The test is administered to a group of
representative samples of children over a range
of consecutive grade levels (1st to 6th)
 The school year is 10 months: a 6th grade
student performing average for the 4th month of
the school year receives 6.4
Types of Norms  Using nationally representative samples to
compare tests that measure the same construct
National Anchor Norms
 Readings tests: BRT & RAT
 The 96th percentile= Raw score of 67 on BRT and
14 on RAT
 The national anchor norms must be obtained by
administering the two tests on the same sample
Types of Norms  Subgroup Norms: Segmenting a normative
sample by a criteria used in initial selection of
Subgroup and Local subjects (age, educational level, ethnicity,
Norms handedness etc.)
 The manual can provide normative info for each
 Local Norms: typically developed by test users
to provide normative info on the local
populations’ performance
Norm vs. Criterion
Referenced Evaluation
 What is the difference?
 Norm-referenced: Evaluating
the test score in relation to
other scores on the same test
 Criterion-referenced: Evaluating
a test score based on whether
some criterion is met
 Criterion: A standard on which
a judgment or decision may be
based
 E.g., Diploma, driver license
etc.
Norm vs. Criterion
Referenced Evaluation

 Criticism: May assess mastery of


basic knowledge, skills, or both,
but has little or no meaningful
application at the upper end of the
knowledge/skill continuum
 These two are not mutually
exclusive a test can be both norm
and criterion referenced
 In a sense all testing is normative
 http://www.edpsycinteractive.org/
topics/measeval/crnmref.html
Can you think of norm vs. criterion
referenced tests you took before?
 Test users should not lose sight of culture as a factor in test administration,
scoring, and interpretation
 Is the test appropriate for the targeted test taker population
 The do’s and don’t regarding culture and psychological tests

You might also like