This document discusses criteria and validity in assessment. It emphasizes that appropriate criteria highlight the most important aspects of the work, not just easy to measure parts. Rubrics are scoring guides that describe varying levels of quality or proficiency along a continuum. There are two types: holistic rubrics provide an overall score while analytic rubrics rate distinct traits separately. Validity refers to making proper inferences from evidence. Rubrics must ensure they assess understanding, not just performance skills. Their criteria should demonstrate understanding is present even if criteria are not fully met, and understanding could be present even if criteria are.
This document discusses criteria and validity in assessment. It emphasizes that appropriate criteria highlight the most important aspects of the work, not just easy to measure parts. Rubrics are scoring guides that describe varying levels of quality or proficiency along a continuum. There are two types: holistic rubrics provide an overall score while analytic rubrics rate distinct traits separately. Validity refers to making proper inferences from evidence. Rubrics must ensure they assess understanding, not just performance skills. Their criteria should demonstrate understanding is present even if criteria are not fully met, and understanding could be present even if criteria are.
This document discusses criteria and validity in assessment. It emphasizes that appropriate criteria highlight the most important aspects of the work, not just easy to measure parts. Rubrics are scoring guides that describe varying levels of quality or proficiency along a continuum. There are two types: holistic rubrics provide an overall score while analytic rubrics rate distinct traits separately. Validity refers to making proper inferences from evidence. Rubrics must ensure they assess understanding, not just performance skills. Their criteria should demonstrate understanding is present even if criteria are not fully met, and understanding could be present even if criteria are.
This document discusses criteria and validity in assessment. It emphasizes that appropriate criteria highlight the most important aspects of the work, not just easy to measure parts. Rubrics are scoring guides that describe varying levels of quality or proficiency along a continuum. There are two types: holistic rubrics provide an overall score while analytic rubrics rate distinct traits separately. Validity refers to making proper inferences from evidence. Rubrics must ensure they assess understanding, not just performance skills. Their criteria should demonstrate understanding is present even if criteria are not fully met, and understanding could be present even if criteria are.
Download as PPT, PDF, TXT or read online from Scribd
Download as ppt, pdf, or txt
You are on page 1of 14
CRITERIA AND VALIDITY
The need for criteria
Appropriate criteria highlight the most
revealing and important aspects of the work (given the goals), not just those parts of the work that are merely easy to see or score
In identifying appropriate criteria, one must
clarify a set of independent variables in the performance that affect the quality of judgment The criteria specify the conditions that any performance must meet to be successful: they define, operationally, the task requirements Rubric
Is a criterion-based scoring guide consisting
of a fixed measurement scale and descriptions of the characteristics for each score point
Describes degrees of quality, proficiency, or
understanding along a continuum Rubrics answer the questions
By what criteria should performance be
judged and discriminated? Where should we look and what should we look for to judge performance success? How should the different levels of quality, proficiency, or understanding be described and distinguished from one another? Two types of rubrics
Holistic - provides an overall impression of
a student’s work - yield a single score or rating for a product or performance Analytic - divides a product or performance into distinct traits or dimensions and judges each separately - rates each of the identified traits independently and scored separately Rubrics to assess understanding
Understanding is a matter of degree on a
continuum. It is not a matter of simple right versus wrong but more or less naïve or sophisticated, more or less superficial or in- depth A rubric for understanding must provide concrete answers to the following key assessment questions: What does understanding look like? What differentiates a sophisticated understanding from a naïve understanding, in practice? What does a range of explanations look like, from the most naïve or simplistic to the most complex and sophisticated? Backward design from criteria and rubrics Backward design suggests that any explicit goal in Stage 1 implies the criteria needed in Stage 2, even before a particular task is designed. Facet- Related Criteria Facet 1 Facet 2 Facet 3 Explanation Interpretation Application •Accurate •Meaningful •Effective •Coherent •Insightful •Efficient •Justified •Significant •Fluent •Systematic •Illustrative •Adaptive •Predictive •Illuminating •Graceful
Facet 4 Facet 5 Facet 6
Perspective Empathy Self-knowledge •Credible •Sensitive •Self-aware •Revealing •Open •Metacognitive •Insightful •Receptive •Self-adjusting •Plausible •Perceptive •Reflective •Unusual •Tactful •Wise Designing and refining rubrics based on student work Step 1: Gather samples of student performance that illustrate the desired understanding or proficiency Step 2: Sort student work into different “stacks” and write down the reasons Step 3: Cluster the reasons into traits or important dimensions of performance Step 4: Write a definition of each trait Step 5: Select samples of student performance that illustrate each score point on each trait Step 6: Continuously refine The challenge of validity
Validity refers to the meaning that can and
cannot properly make of specific evidence, including traditional test-related evidence
The challenge: At what events or data should
we look to obtain the most telling evidence of more general abilities? Validity affects rubrics design. One has to make sure that one employs the right criteria for judging understanding (or any other target), not just what is easy to count or score. In assessing for understanding, one must beware of confusing mere correctness or skill in performance with degree of understanding. Two questions that help self-assess the validity of criteria and rubrics
Could the proposed criteria be met but the
performer still not demonstrate deep understanding? Could the proposed criteria not be met but the performer nonetheless still show understanding?
If the answer is yes, then the proposed
criteria and rubric are not yet ready to provide valid inferences.