Assessment in Learning 2 Module 3
Assessment in Learning 2 Module 3
Assessment in Learning 2 Module 3
Objective: At the end of this module, the students will be able to differentiate Criterion and Norm-
Referenced Assessment. The students will be able to conduct an Item Analysis also.
Sources:
https://www.turnitin.com/blog/what-is-item-analysis-and-other-important-exam-design-
principles#:~:text=Item%20analysis%20is%20the%20act,unconsciously%20on%20a%20regular%20basis.
https://www.turnitin.com/blog/what-is-item-analysis-and-other-important-exam-design
principles#:~:text=Item%20analysis%20is%20the%20act,unconsciously%20on%20a%20regular%20basis.
https://www.edglossary.org/criterion-referenced-
test/#:~:text=Criterion%2Dreferenced%20tests%20and%20assessments,specific%20stage%20of%20their%20education.
https://www.edglossary.org/norm-referenced-test/
_________________________________________________________________________________________________________
But when the process is formalized, item analysis becomes a scientific method
through which tests can be improved, and academic integrity upheld.
• Item Difficulty -- is the exam question (aka “item”) too easy or too hard?
When an item is one that every student either gets wrong or correct, it
decreases an exam’s reliability. If everyone gets a particular answer
correct, there’s less of a way to tell who really understands the material
with deep knowledge. Conversely, if everyone gets a particular answer
incorrect, then there’s no way to differentiate those who’ve learned the
material deeply.
• Item Discrimination -- does the exam question discriminate between
students who understand the material and those who do not? Exam
questions should realize the varying degrees of knowledge students have
on the material, reflected by the percentage correct on exam questions.
Desirable discrimination can be shown by comparing the correct answers
to the total test scores of students--i.e., do students who scored high
overall have a higher rate of correct answers on the item than those who
scored low overall? If you separate top scorers from bottom scorers,
which group is getting which answer correct?
• Item Distractors -- for multiple-choice exams, distractors play a
significant role. Do exam questions effectively distract test takers from the
correct answer? For example, if a multiple-choice question has four
possible answers, are two of the answers obviously incorrect, thereby
rendering the question with a 50/50 percent chance of correct response?
When distractors are ineffective and obviously incorrect as opposed to
being more disguised, then they become ineffective in assessing student
knowledge. An effective distractor will attract test takers with a lower
overall score than those with a higher overall score.
Item analysis entails noting the pattern of student errors to various questions in
all the ways stated above. This analysis can provide distinct feedback on exam
efficacy and support exam design.
Shoring up student learning can be enacted through feedback, but also exam
design. The data from item analysis can drive the way in which you design
future tests. As noted previously, if student knowledge assessment is the bridge
between teaching and learning--then exams ought to measure the student
learning gap as accurately as possible.
Item analysis should bring to light both questions and answers as you revise or
omit items from your test.
In doing so, item analysis can increase the efficacy of your exams by testing
knowledge accurately. And knowing exactly what it is students know and what
they don’t know, helps both student learning and instructor efficacy.
Not only can item analysis drive exam design, but it can also inform course
content and curriculum.
When it comes to item difficulty, it’s important to note whether errors indicate a
misunderstanding of the question or of the concept the item addresses. When a
large number of students answer an item incorrectly, it’s notable. It may be a
matter of fine-tuning a question for clarity; is the wording of the question
confusing? Are the answers clear?
Or it could be that the material may have to be reviewed in class, possibly with
a different learning approach.
Whether you employ item analysis manually or via software, we think data-
driven exams and curricula are a great thing. And we hope this helps you out on
your pedagogical journey.
EXAMPLE
Step 1: Arrange the scores from highest to lowest.
STEP 2: Separate the top 27% and the bottom 27% of the cases.
42 x 0.27 = 11.34 ≈ 11
𝑹𝒖 = 𝟗
𝑹𝒍 = 𝟏
Solution:
9+1
Index of difficulty = x 100
22
Interpretation: The index of difficulty is 45.45 percent (45. 45%). This percentage value is below
fifty percent (50%), hence, Item 18 is Right difficult and must be retained.
STEP 2: Separate the top 27% and the bottom 27% of the cases.
42 x 0.27 = 11.34 ≈ 11
𝑹𝒖 = 𝟗
𝑹𝒍 = 𝟏
Solution:
9−1
Index of discrimination =
11
The discrimination index is 0.727 which means the item is very good and must be retained.
ACTIVITY 2
Find the INDEX of DIFFICULTY and INDEX of DISCRIMINATION. Show the steps and
interpretation.
Result of Item 2 Taken by 36 Grade 8 students in Mathematics Achievement Test Subject for Item
Analysis