Notes
Notes
The teacher needs a way to determine how successful the instruction has been for the class in
general and for the students in particular.
Assessments can show whether or not the newly designed instruction has met its objectives.
EVALUATION: This refers to the interpretation of assessment either in terms of grades (e.g.
A, B, C)
Or qualities (e.g. Good, Fair, Poor). It involves the collection and analysis of data
for the purpose of making decisions.
Test: Refers to any procedure used to assess the performance described in an objective.
Types of Evaluation.
Statistical Considerations.
Anticipated Variability: Is instruction planned for learners who differ considerably in regard to
course content? OR
Does one expect to work with learners who are uniformly uninformed?
Is instruction individualized to the extent of creating variability among the learners? i.e.
Everyone is expected to learn, but better learners are expected to learn and improve more than
those with lesser ability.
When individual differences are present at the initial stages, then we expect variability (i.e.
spreading out). In this case, the instruction is individualized but the time length is fixed. This is
the main set up in Kenya’s education system. Pupils enter class one and are all expected to be in
class eight at the same time to attempt the KCPE. The system does not give room for pupils to
repeat classes on account of being slow in catching what is taught in class.
If one plans to teach each individual until they all reach a predetermined standard of performance
and then stop teaching, then one should expect little variation if any. In this case no individual
differences should be expected in the final assessments. Here, instructional time is
individualized to reduce variability in performance.
Content Validity: This refers to the degree to which the test content matches with instructional
objectives
Reliability: This is the degree to which a measure can be trusted to be stable over multiple
conditions.
Improving Tests.
Item Difficulty Index: This refers to the ability of an item to provide information to us about
whether or not the item helps us tell successful learners from the unsuccessful ones.
Distractor Analysis: This is the process for improvement of multiple-choice test items by
identifying distracters (wrong choices) as being effective or not. The wrong choices given in the
test items should be ones which do not give away the correct alternative. They should be as close
to the correct choice as possible in order for the candidate to think carefully in making their
choices. Ineffective distracters can be modified to salvage an item that might otherwise be
rejected on the basis of the discrimination value.
Criterion-Referenced assessment is important for evaluating both learners, progress and the
quality of instruction itself. The results of criterion-referenced assessment indicate to the
teacher, exactly how well learners were able to achieve each instructional objective. They also
indicate which components of the instruction worked well and which ones need to be revised.
Criterion –referenced assessment enables learners to reflect on their own performance.
Thus, clarity in specifying instructional objectives and the criteria for adequate (acceptable)
performance is necessary as a guide to adequate test construction.
Authentic Assessment: - There is a close relationship between learning contexts and assessment
contexts. It involves many different skill and aspects of knowledge. It may continue over a
relatively long period as teaching and learning proceed. It is non-algorithmic i.e. the path of
action is not fully specified in advance. It can involve both individuals working alone or in small
groups. Students are much more autonomous and more involved in planning their own tasks and
assessment procedures. The focus of control rests with the learner. Test takers are active
participants in assessment activities. It offers real challenges based on real tasks. It often
culminates in learner’s own research product for which content is mastered as a means, rather
than an end. Scoring assessment tasks can be complex and a wide range of scoring techniques
may be required. It is not possible to construct parallel forms and have claims of high predictive
validity.
-increase inter-rater reliability for grading where multiple graders are assessing student work,
5. Placement Tests: These help to identify the starting point for each learner. They can
be used
5. Diagnostic Tests: They are constructed to measure prerequisite skills. They are helpful
for learners who are falling behind in group instruction. Such learners can be helped
through remedial instruction on prerequisite skills.
6. Progress Tests: They are tests which are administered after lessons. They help ensure
students have mastered the lesson’s objectives. They can be used as practice tests over
desired objectives
Assessment Rubrics:- This is a list of criteria against which to compare elements of the
performance. It is essentially a one or two-page document that describes varying levels of
quality from excellent to poor. It is used for a relatively complex assignment. It gives learners
informative feedback about their work in progress. While the formats of rubrics vary, two
features are common. These are (1) A list of criteria (what counts) in a project or assignment
and, (2) Gradations of quality.