Assessment of Student Learning 1
Assessment of Student Learning 1
,,4
t&
4ffi.
37L.26
R261s
'2010
----s
I
/
s
I
Chapter 3 SOCIAL, LEGAL, AND ETHICAL
T IMPLICAIIONS OF TESTS + 28
t
I Criticisms of Testing + 28
Iil
Ethical Testing Principles + 30
E
i
. EthicalTesting Practices + 31
I
a
UnethicalTesting Practices + 32
I
B
{ Reuiew Exercises + 33
*
I
I 4 FACTORS INFLUENCING TEST CONSTRUCTION
Chapter
I
AND TEST PERFORMANCE + 34
{
&
s Factors that Influence Test Construction + 34
q
Exhaneous Factors that Influence Test Performance +37
Reuiew Exercises + 39
ChapteT12GRADINGANDREPoRTINGPRACTICES+138
Nature of Grades/Marks + 139
Functions of Grades/Mark + 139
PurPoses of Grades/Marks + 140
r
Types of Grades/Marks + 140
Advantages of Grades/Mark + 141
Disadvantages of Gradesfir4arks + 141
Common Grading-Related Problems + 142
Averaging Scores to Determine a Grade + 742
Use of Zeroes + 142
Lowering Grades Because of Behavioral Infractions + 142
Guidelines for Effective Grading + 143
Criteria for a Marking-Reporting System + 143
Modes of Computing FinalGrades + 143
Reuiew Exercises + 145
Glossary +L47
References + 151
Index+ 157
1g
OBJECTIVES
ot
At the end of the chapter, the learners are
. identify what constifutes high_quality expected to:
assessments;
' Iist down the productive and unproductive
uses of tests: and
. classifu the various types of tests.
77
,-
I
2. Appropriateness of Assessment Methods
Once the learning targets have been identified, match them w-:
their corresponding methods by considering the strengths of varic",'
methods in measuring different targets.
Table z.t
MATCHING LEARNING TARGETS WITH ASSESSMENT METHODS
ii:::|i:.
SX** ffi{XM*i*;;;:
Knowledge 5 4 3 4 3 2
Reasoning 2 5 4 4 2 2
Skills 1 3 5 2 5 3
Products 1 1 5 2 4 4
Affect 1 2 4 4 4 5
3. Validity
iJ This refers to the extent to which the test seryes its purpose or the
'
efficiency with which it intends to measure. Validity is a characteristic
thatpertains to the appropriateness of the inferences, uses, and resulE
rl,
e/dti
.k \gw assessment of student Learning 1: cognitive Learning
'"8,1
,-.s
ll. criterion-related ualidity determines the relationship
iwith between an assessment and another measure of
the
rrious same trait. It provides validity by relating an assessment
to some valued measure (criterion) thut .un either
provide an estimate of current performance (concurrent
criterion-related evidence) or predict fufu re performance
(predictive criterion-related evidence).
lll. c on str uct-r el ated u al idity determines which assessment
is a meaningful measure of an unobservable hait or
characteristic like intelligence, reading comprehension,
honesty, motivation, attifude, learning style, und
anxiety.
iv. Face ualidify is determined on the basis of the
appearance
of an assessment, whether based on the slperficial
examination of the test, there seems to be a reaionable
measure of the objectives of a domain.
V. Instructional-related ualidity determines to what
extent
the domain of content in the test is taught in
class.
b. TestValidity Enhancers
The following are suggestions for enhancing the ll
:r the varidity of
classroom assessments:
zristic ll
esults i. prepare a table of
specifications (TOS). I
it
ii. Construct appropriate test items.
mely, iii. Formulate directions that are brief, crear, and
concise.
r and iv. consider the reading vocaburary of the examinees.
ms. The test should not be made up oilurgonr.
v. Make the sentence strucfure of your test items
simple.
inent.
vi. Never have an identifiable pattern of answers.
s€ in vii. Anange the test items from easy to difficult.
ration viii. Provide adequate time for student to complete the
assessment.
,rity
Thus, an observed score can be higher or lower than the hue
score, depending on the nature of error. The sources of error are
the
reflectedinTable 2.2.
me Table z.z
rne SOURCES OF ERROR
lnd
ion
uer, Health Directions
rl is Mood Luck
Ing Motivation Item ambiguity
low
Test-taking skills Heat in the room
hod lJ
Anxiety Lighting
Iect
rval Fatigue Sample of items lt
Generalability Observer differences and bias
I
il
ailel Test interpretation and scoring
rihe
ted. c. Test Reliability Enhancers
lner The following should be considered in enhancing the
s for reliability of classroom assessments:
r'cut
i. Use a sufficient number of items or tasks. A longer test is
more reliable.
L3e
lt. Use independent raters or observers who can provide
tnto
similar scores to the same performances.
-25[
-51
iii. Make sure the assessment procedures and scoring are
:S objective.
Continue the assessment untilthe results are consistent.
v. Eliminate or reduce the influence of extraneous events
or factors.
vi. Assess the difficulty level of the test.
vii. Use shorter assessments more frequenfly rather than a
few long assessments.
u / :/,
Chapter 2: Principles of High-Quality Assessment,i
Zlf{
'.s#;*
r 5. Fairness
This pertains to the intent that each question should be made
clear as possible to the examinees and the test is absent of any bia--s
=
An example of a bias in an intelligence test is an item about a person :r
object that has not been part of the cultural and educational contex :n
the test taker. In mathematicaltests for instance, the reading difficr:-
level of an item can be a source of unfairness.'Identified eleme:=
of fairness are the student's knowledge of learning targets befc::
instruction, the opportunity to learn, the attainment of pre-requi_<::
knowledge and skills, unbiased assessment tasks and procedura=
and teachers who avoid stereotypes.
6. Positive Consequences
These enhance the overall qualip of assessment, particularly -:.
effect of assessments on the sfudents' motivation and sfudy habil.
Learning Analysis. Tests are used to identifu the reasons or causes \\: "
students do not learn and the solutions to help them learn. Ideally, d .r€::
should be designed to determine what students do not know so that -r=
teachers can take appropriate actions.
Improuement of Cuticulum. Poor performance in a test may indice:.
that the teacher is not explaining the material effectively, the textbook is :.:"
clear, the students are not properly taught, and the students do not see --:
meaningfulness of the materials. When only a few students have difficuffe,
the teacher can address them separately and extend special help. If the enr
class does poorly, the curriculum needs to be revised or special units need :
be developed for the class to continue.
Improuement of reacher.In a reliable grading system, the class avera::
is the grade the teacher has earned.
r J
those who are already fearfur of tests;
unannounced tests do not give sfude
and'surprise tests do not promote effio
ll:":i-:tltT."_l:,prepare;
learning or higher achievemeni.
Ridiculing. This means using tests tb
deride sfudents.
Tracking. Students are grouped according
to deficiencies as revealed
tests without continuous reevaruution, r
ehx(rcotions ( eesb
Throughout the. years, psychologists
and educators have cooperatir,,pq
::11::"i::.:11 _u;ttei
performance with greater
un"a t",t ,.;;, ffi;;;sure
the ,tua"nl
accuracv. These tests may u" Jr]ll,r"at
to:
":Llil
1. Administration
a' Indiuidual - givln orary and requires the examinees,
cons*
since the m3nn9r.of answering miry
fli"::,^._l be as importanr a
the score. An exampre of this i.-irr"
wl.iri;r;fril'i#;,iffi3
.l: ihree^indivijuarf
iSl": Another
scales. l-h",is a powerpointv administer"a ini"irig"" r,
:l presenta,"r" r."i"tu"j':
performance test in a speech class.
i'
:91
b. Group - for measuringcognitive skills
to measure achievemen:
Most tests in schools are clnsid"r"a group
' i"Jtl *n"re differcr
ii
rr test takers can take the tests as a group.
2. Scoring
a. objectiue - independent scorers agreeon
the number of points
answer shourd receive, e.g., multiple re
choice una tu" or farse.
b' subjectiue - answers can be scored through various
These are then given different values war=
uv;;;:,
performance tests. ";.;;;;;;
3. Sort of Response being Emphasized
a' Power - ailows examinees a generous time rimit to be
answer every able ::
item. The questions are difficurt and
this difficui:.
is what is emphasized.
b. speed - *itfseverely rimited time constraints but
the items a_:-;
easy and only a few examinees are
expected to make errors.
4. Types of Response the Examinees
must Make
a' Performance - requires.students to perform a task. This is usuaji-
administered individuaily so that itre
e*u-in", .un .ount tiu
effors and measure the time the examinee
has performed"r_
each task.
b' Paper-and-pencir
- examinees are asked to write on paper.
...,rt...*%'
Assessment of Student Learning 1:
cognitive Learning
"67,3:X
tnts
ient 5. What is Measured
a. Sample - limited representative test designed to measure the
total behavior of the examinee, although no test can exhaustively
lbv measure all the knowledge of an individual.
b. Sign tesr - diagnostic test designed to obtain diagnostic signs to
suggest that some form of remediation is needed.
6. Nature of the Groups being Compared
a. Teacher-made test - for use within the classroom and contains
the subject being taught by the same teacher who constructed
;ely the test.
rnts'
iing
b. Standardized test - constructed by test specialists working with
curriculum experts and teachers.
.*.. , art
Chapter 2: Principles of High-euality Assessment ,g,'iSY
,r:i:i ..,,. .
72' Non-language tests are administered by means of pantorn::
painting or signs and symbors, e.g., p-grl.rire
Raveri's Matrice.
the Abstract Reasoning Tests.
13' structured tests have very specific, well-defined
instructions
expected outcomes.
74. Projective tests. present ambiguous stimulus
or questions designe_
elicit highly individualized reiponses.
15. Product tests emphasize only the final answer.
16' Process tests focus on how the examinees
attack, sorve, or wor:.
a problem.
77 ' External reports are tests where a ratee is
evaruated by anc:
person.
18. Internal reports are self-evaluation.
19' open book tests depend on one's understanding
express one's ideas and evaluate ' and
concepts.
20' closed book tests depend heavily on the memory
of the examr:,.
i
27. Non-learning format tests determine how
much informatio:.
i
:f*
students know.
22' Learning format tests require the sfudents
to apply previousry rea
'ulx
materials.
23' convergent format tests purposely read the
examinees to one
answer.
24. Divergent format tests read the examinees
to several
answers.
25. scale measurements distribute ratings along a continuum.
26. Test measurements refer to the items being dichotomous
right or wrong, but not both. or €
b..::ing K;ri*ta
I :: ihe 1. Explain why validity implies reliabilip but not the reverse.
2. Generate some other qualities that you believe contribute to making
good assessments.
E:=:e a
3. List down your personal experiences of unfair assessments.
OBJECTIVES
erificisnrs o/ eesting
2.
Relevance of information I
3.
Informed consent
4.
Confidentiality
(7976) derived the following implications for
_^_chase teaching frorn :*r
7974Family Educational Rights and privacy Act (Buckley
Amendment):
1. Teachers cannot post the grades of sfudents.
2. Teachers cannot display the works of their sfudents as an exampi. .r
poor or excellent work.
28
3. Teachers are not allowed to let the shrdents grade or correct any
t other student's paper.
4. Teachers cannot ask sfudents to raise their hands to determine if they
{ answered correctly or incorrectly to any item.
5. Teachers cannot distribute test papers in a manner that will permit
other sfudents to observe the scores of others.
6. Teachers cannot assume that letters of reiommendations requested
by students willbe kept confidential.
.r.r'3:,'
Chapter 3: Social, Legal, and Ethical Implications of Tests;,29:1#
;;,, .
".4;4.
4' Be sure each item has "face
varidity." Items measure some
aspects of life as perceived importa
by the students.
5. Avoid unannounced examinations.
6' Schedule personar conferences
with students as often as possibre
reduce anxiep and redirect
learning.
7. Emphasize more on strengths,
not on deficiencies.
8' Do not emphasize competitive
examinations when some studer
are unable to compete.
9. Treat each student,s grades
confidentially.
10' Allow sfudents to choose among
activities of equal insfuuctionar
varue
Permanent categorization of
students.(Tracking). The notion
measurement instruments are iJ:a
-f,uJ" and their;;;;i_""ce is fixed
infaribre
therefore, unchangeable, fru. anc_
teachers -t il;i,",".o
gn iti o n Jor ::["':rffi:i?,T?liii;ror
Penalizing Bright and creatiue
i nstan c;
studenrs. one or-tr,n most
consiste:r
Sf:'jil::1ff:':,'::l*:1,:::,'::."Tn'y" " 'rr,.,iiu"ialr"n.iure response= ,
,ouf.course'therearesomevasue,"-.,1r,",fl,I[:.":1ffi:t:ltJil:tx,i::':
Discrimination against Minority
studen.b.Many minoritgr sfudents
i take tests because th-ev have do ::n
?* the .kili;; r.rorr"ag", and attitu:;
required to succeed, In this ";il;;"d
case, it is not reasonabie to
'{* criticizetests. Rath;.:
J ro r ra,in g io1"
:lli JI":Iti #: :'lJ :1.?5 ;r'-
" ;;;.
".i,,, a e n ts th e -
M e as ur e m ent of L i m it3 d,an d
up erf ci ar Aspe ct of B eh
.s i
au or . Acom m o L_.
i
ENtS
b. If the students will benefit by talking to other professionals
concerned with a case
c. If the student gives permission for confidential communications
to be reported to others
u€. as such, their
2. Test secunfy. Tests are professional instruments and
to
nat dissemination is restricted to those with the technical competence
^^l
a.lu! ,rrn tnn* properly. No standardlzedtests should be left unsecured.
r- 'O to
3. Test Scores and lnterpretations. These should only be available
individuals who are qualified to use them. Test results should
be
against misuse
t-.:ent interpreted to parents in a manner that will ensure
,f ioc
and misinterPretations.'
ian a
provide a manual
4. Test Publication. standardized tests should
nOt
or technical handbook describing how the test can be used most II
effectively and who may use it. Advertisements about test
rC results A
n:ide persuasive.
should be factual and descriptive but not emotional or
dd
a:her, \rl
aeded
t thical eesting Practica
:l-.only
1. It is both ethical and advantageous to inform students in advance
:aits,
that they are about to take a test and to tell them something about
of
the nature of the test. They should also be told of the advantages
taking the test and where the results would be used'
practice
2. Teachers should explain the mechanics of taking a test and
heavy
r:riner the students on how to fill out an answer sheet (i.e., making
marks, and erasing marks completely). It is however essential
that the
social
testing teacher does not make the question available'
3. It is perfectly proper to try to motivate students to do as well as they
s Who .u1 u, long-ai they are not threatened or made anxious about their
I asthe performance.
and 4. It is essential that all testing materials and results be kept secured
before, during, and after testing.
It is ethicalto combine classes for testing as long as there are
adequate
5.
p-.torc to safeguard the test and make sure that the students are
iollowing instructions. The ideal ratio is one proctor to a maximum
F., of 30 students.
72 -Y1C
d-'ed
E.*gh
E:e
F€5.
fE::al
[t
ill'e
,d
OBJECTIVES
,i. There are several factors that influence the type of test a teacher has to
i construct. These are:
l|t
1. The Function of the Test in the Instructional Process
,l*,1;
ffi&
f;ffiffis&sffi
strffixx
W
xffiffi *,itt$i[g*$ :rPlA&iiiiei:iii fiaririi*s[&.4 stit3ll*ii&trffi
Focus of Prerequisite Course Predefined Most Course or unit
measurement entry skills cr unit segment of common
objectives instruction learning
errors
Nature of Limited Broad Limited Limited Broad sample of
sample sample of sample of all sample of sample of all objectives
selected skills obiectives selected skills selected skills
34
(Table 4.t Basic Types of Clossroom Tests continued)
rii:.::i:lsxss,. E',xllil
'*:*"EU$ffi",'.^l
:{I{5f,&&}CXfi }li[:ill ffis*ffii
?st ruxcmox !ii*€a&i&t :,,ptiidi{iii€iiti &ii*llt&ii& &ffi$*tm ill$ii&*rii*i I
Item difficulty Typically has Typically has Varies Typicaliy has Typically has wide
low level of wide range with the low level of range of difficulty
?ST
difficulty of difficulty segment of difHculty
instruction
tc€ Time of Beginning Beginning Periodically As needed End of course or
administration of course or of course or during during unit
unit unit instruction instruction
Use of results Remedy Instructional lmprove Remedy Assign
entry planning and and direct errors grades, certify
deficiencies advanced learning related to accomplishment,
or placement through persistent evaluate teaching
assignment ongoing learning
'r to learning feedback difficulties
group
(Airasian & Madaus, 1972, as cited in Linn and Gronlund,
2000)
2. Testing Frequency
Testing frequency depends on the function the test is to serve.
NtO In most cases, single testing can suffice for selectiori, placement,
and summative evaluation decisions. For diagnosis, formative
t
evaluation, and motivation, frequent testing is recommended and Ji
necessary. If the purpose is to provide feedback, frequent testing
I
e-ifu tl
can help the teacher guide and direct the students to thl right track]
ESES
particularly during the early stage of learning. Testing fLqrnn.y
n .}.re
t--nn helps the student acquire knowledge as efficiently as plssiutl. rne
tests thus serve as learning drills. Frequent testing can also motivate
b".,el
the learners, particularly those with lower abilities. Although it has
some disadvantages, the foremost of which is that it takes much
of
instruction time, the decision still lies with the teacher as to which is
to be prioritized.
3. The Use of open-Book and closed-Book Examinations
An open-book test permits the sfudents to use books or notes
during examinations. This is advisable when the teacher emphasizes
some of the higher objectives in Bloom's Taxonomy. Rather than
spend
the bulk of the time memorizing, it can be devotei instead to appiying
specific formula on information. The values of this type of test incfuA!
having sfudents apply rather than memorize information and use skills
in utilizing reference materials for a limited time. on the other hand, a
closed-book test is advisable and should be implemented when a test
emphasizes recall of information.
rt:'rl. .. .
r--
4. The Use of Take-Home Examination
Take-home examinations are extensions of open-book e
Its advantage is that the sfudent can work at his/hei leisure
ani
il
his/her most comfortable rate of speed using *nit"r* refers.
I
I t;
is available. It is most useful when the sfudent needs to resor
Il references not found in the classroom or even in the school.
Howe
it has at least two disadvantages, namely, difficulty of scoring ::
yuo$s objectivelv,
3!d the possibilitv that the studlnt might no:
his/her own work. However, if the ieacher emphasizes the
exe:
or teaching device to indicate student strengtirs, weaknesses. a
method of improvement rather than gradi"g, tn. use of take-r-,:.r
tests are maximized.
*t *f*'
Attessment of Student Learning 1: Cognitive Learning
dSO _€
,Y&;'&"'
I
€vtraneous iartors thnt lnfluence eest ?erforwauce
tests.
Maximum performance tests (achievement, intelligence, and aptitude)
and at
require the student to obtain the highest possible scores while typical
erence performance tests (attitude, interest, and personality inventory) call for
sort to
ihe students to obtain scores without exerting much effort. The goal of
!1:ever.
cognitive measurement is to obtain an examinee's best and highest level of
E long pn.for*urce. The purpose of affective assessment is to assess an examinee's
not do usual, representative, and typical behavior (Hopkins & stanley, 1981).
xercise
An examinee's performance in any test is influenced by personal traits,
s. and knowledge, or proficiency. Extraneous factors, oftentimes unrecognized,
r-home
likewise cause equal or stronger influences. These include the following:
.,e.r#J'
Chapter 4: Factors Influencing Test Construction and Test Peformaor",# 37 _.{
.4" ,r1;tt
r k. Uses relevant and extraneous cues to help identifu the cor:,
option (the correct answer is more likely to be qualifiec
longer, represents a different degree of generalization. cr
composed of textbook or stereotyped phraseology)
l. Tries to uncover the test constructor's tendencies to -
certain response positions or to include disproport
between true-or-faise answers
m. Recognizes the use of specific determiners and
from grammatical consfuuction, €.g., articles, agreemen:"
parallelisms
2. Practice
Several sfudies show that there is an improvement in scorei
when an examinee takes a subsequent test on a particular test or :=
parallel form. Practice testing helps the examinee improve hisrl:e
scores the next time he/she takes the same test or any of its parala
forms. The following are the generalizations on the effect of practic;
on cognitive tests:
a. The effects of practice are more pronounced in people wiil
a limited educationalbackground or experience with tests.
,t
b.
,
3. Coaching
The effect of coaching is smaller compared to the effectbrought
about by practice as long as the objective questions are reasonablg
straightforward.
..:.:.1 ,.1..r,._
to use a. Speed us. Accuracy Set - Some examinees work slowly but
IOnl0n accurately while some work quickly but less cautiously. The
former may consume much time in just a few items but the
chances of getting the answer correct, for each item is greater.
clues
The latter, on the other hand, may cover almost all items
ilents,
but the accuracy of answers is less. Ideally, a test should be
timed in such a way that 90% of examinees complete the
test.
SCOTES b. Acquiescence Set - Cronbach (1984) found that in case of
r or its uncertainty between a true or false item, most examinees
i is her answer true. Metfessel found that most instructors have
:arallel the tendency to include more true items on the test (in
:actice Cronbach, 1984). The acquiescence set allows more people
to get credits, usually undeserved, for true items than false
ones.
le rvith
c. Positional-Pret'erence Set - Most students who are ignorant
of the answer to an item do not answer in a random manner,
Ji
but base their answer on some patterns.
,.arallel
d. Option Length Set - Most examinees choose longer options, J,}
:.;pical
Kroai't:writ i
1. Cite the arguments for and against testing frequency.
rrought
nnably
2. Discuss the distinct feafures of a take-home examination.
3. Explain why the effects of practice in student's learning are more
pronounced than those of coaching.
not. A
:o poor
resting,
ng. the
ects on
€stablishing
,,, {oorfiiilg Cargett
oBJECTTVES I
&
A good classroom assessment has clear, appropriate learning targets.
i Defining learning targetsloutcomes is the first step in good teaching. It is also
hi essential in the assessment of student learning.
ttt
Purpoxs of lustructional 6oals and Obiectiues
Instructional goals and objectives provide direction for the instructional
process by clarifying the intended learning outcomes. They also convey
instructionalintent to stakeholders such as sfudents, parents, other personnel,
and the public. Most importanfly, they provide a basis for assessing student
learning by describing the performance to be measured. Instructional goals
and objectives are sometimes stated in terms of actions to be taken. Thus, there
is a statement saying: "demonstrate to students how to operate computers."
This statement clearly indicates the teaching activity. However, it does not
clearly specify the intended learning outcomes and does not point very
explicitly to the type of student assessment that would be most appropriate.
Educational goals are general statements of what sfudents will know
and be able to do. Goals are written to cover most of the instructional time,
such as a unit, semester, or year, and indicate in broad terms what will be
emphasized during that time period. Goals are essentialbecause they reflect
educational philosophies about what is important. They also provide a
starting point for more specific learning objectives. By beginning with goals,
the general outline can be validated by parents, teachers, and school officials.
40
An example of a goal is "knou how to think uitically, be good followers of
school policies, and work with peers."
Instructional objectives are sometimes stated in terms of actions to be
taken. Thus, statements like these abound: "the students will demonstrate
how to use the microscope" and "the student will underline the subject and
predicate of each sentence."
Statements, such as the ones mentioned, direct attention to the students
and the types of performance they are expected to exhibit as a result of
instruction. Thus, the focus shifts from the teacher to the sfudents and from
the learning experiences to the learning outcomes.
In stating instructional objectives, it is important to keep in mind that the
concern is on the product of learning rather than with the process of learning.
This is not meant to suggest that the process is unimportant. Remember, the
long-term instructional objective concerns the product.
Instructional objectives are usually relatively specific statements of the
studentperformance thatshould be demonskated atthe end of an instructional
unit. They are meant as intended learning outcomes. They should be stated
in terms of specific, observable, and measurable sfudent response. These
types of objectives are characterized by the use of action verbs, such as list,
outline, distinguish, categorize, apply, subtract, synthesize, underline, and
gets. define. Highly precise behavioral objectives include the following criteria:
T
also 1. Audience. This is a description of the students who are expected to
perform or demonstrate the behavior of a specific grade level. J.I
{earning eargets
Terms associated with learning targets are goal, objectives, competency,
outcome, standard, and expectation. Learning target is defined as a statement
of sfudent performance that includes both a description of what students
should know or be able to do at the end of a unit of instruction and the criteria
for judging the level of performances demonskated.
The word learning is used to convey that targets' emphasis on the
importance of how students will change. The learning targets are composed
Lu of content and criteria. Content is what sfudents should know and be able to
1i, do. On the other hand, criteria are dimensions of student performance used
for judging attainment.
It is necessary that the criteria for judging levels of performance be included
in the target. The target "students will know the cities in the Philippines"
means something different if the students have to commit the major chartered
cities to memory than if the students can correctly match half of the names
of major chartered cities with provinces or regions. The criterion for judging
levels of performance needs to be communicated to the students prio-r to
instruction. Below is an example of a learning target:
students will be able to explain why changing the gouernment
from
presidential to parliamentary form is (un)necessary by writing an essay
that indicates the fauorable or unfauorable conditions the chinge wouli
bring to the economy. The papers wilt be graded holistically bf, looking
for euidence of reasons and knowledge of the forms of gouLrnment anZ
organization.
,:..r ..rar,.,
Cha pter 5 : Esta bl ishing Learni n g fargets..;,ai'if;
.::,,.,,,f ,.,j;..
The cognitiue domain includes objectives that deal with the recall or
organization of knowledge and the development of intellectual abilities
and skills. The cognitive taxonomy contains six major classes of objectives
arranged in hierarchical order on the basis of the complexity of a task.
Bloom's Taxonomy enables teachers and educators to use exact and
varied terminologies for stating specific learning outcomes. These terms are
identified in the table below:
Table 5.t
ffi
FORMULATING COGN ITIVE LEARN I NG TARGETS
&
. er'fo"
,.{447$ Assessment of Student Learning 1: Cognitive Learning
{&*'
:all or (Table 5.r. F ormuloting Cognitive Learning T argets continued)
rilities
rctives
:t and
ns are
ANALYSIS. lt refers to the Recognizes unstated Breaks down;
ability to break down material assumptions; diagrams;
into its component parts recognizes logical fallacies differentiates;
so that its organizational in reasoning; discriminates;
structure may be understood. distinguishes between distinguishes;
This may include the facts and opinion/ identifies; illustrates;
identification of the parts, inferences; infers; outlines; points
and recognition of the evaluates the relevance of outl relates; selects;
organizational principles data; separates; subdivides
involved. Learning outcomes analyzes the organizational
here represent a higher structure of a work
intellectual level than the
comprehension of both the
content and structural form of
the material.
SYNTHESIS. lt refers to the Writes a well organized Categorizes;
ability to put parts together theme; combines; compiles;
to form a new whole. This gives a well organized composes; creates;
may involve the production speech devises; designs;
of a unique communication writes a creative short explains; generates;
/l
(theme or speech), a plan story, poem, or music; modifies; plans; J}
of operations (research proposes a plan for an organizes; rearranges; q
proposal), or a set of experimentl reconstructs; relates; I
abstract relations (scheme integrates learning from reorganizes; revisesl
for classifying information). different areas into a plan rewrites; summarizesl
Learning outcomes in this for solving a problem; tells; writes
area stress creative behaviors,formulates a new scheme
with maior emphasis on the for classifying objects,
formulation of new patterns or events or ideas
structures.
EVALUATION. Evaluation is Judges the logical Appraises; compares;
concerned with the ability to consistency of written concludesl contrasts;
judge the value of a material material; criticizesl describes;
(statement, novel, poem, judges the adequacy with discriminates; explains;
research, and report) for a which conclusions are justifies; interprets;
given purpose. The judgments supported by data; relates; summarizes;
are to be based on definite judges the value of work supports
EteS;
criteria or be given to them. (art, music, and writing) by
;
Learning outcomes in this area use of internal criteria;
;
are highest in the cognitive judges the value of work
lates;
hierarchy because they contain (art, music, and writing) by
elements of all the categories, use of external standards
plus conscious value judgments of excellence
based on clearly defined
criteria.
, *)-
Chapter 5: Establishing Learning Targets u 45 .'
.r;. ' r
r The at'fectiue domain is concerned with changes in interests, attitudes, lt:
more specific moral stance. Therefore, the only way to assess the attitudes :
:€
or values people have is by observing what they do or say. Affective domain j-.
,t :l
:-
Receiving (Attending) Phenomena: Examples: Listens to others with respect; ::
Awareness, willingness to hear, selected listens for and remembers the name of
attention newly introduced people
This is the lowest category in the affective Key Words: asks; chooses; describes;
domain. At this level, the student is aware follows; gives; holds; identifies; locates;
i
of the existence of a condition or problem names; points to; selects; sits; erects;
*n and is willing to at least listen attentively replies; uses
1*,
to what others have to say about it. The
element of commitment is not present, and
the behavior is somewhat analogous to
"sitting on the fence." The student is aware
of an issue, but has not yet made a decision lnt
about it. Ha
be
Responding to Phenomena: Active Examples: Participates in class discussion; pr(
participation on the part of the learners; gives a presentation; questions new ideals, ch;
attends and reacts to a particular concepts, models, etc. to fully understand ob
phenomenon. them; knows the safety rules and practices do
Learning outcomes may emphasize them SO(
compliance in responding, willingness to Key Words: answers; assists; aids; Th
respond, or satisfaction in responding complies; conforms; discusses; greets; do
(motivation). At this level, the student is helps; labels; per{orrns; practices; dl'lr
willing to go along with an idea or a value, presents; readsl recites; reports; selects; th;
such as being willing to follow school rules, tells; writes per
actively volunteers to respond, and takes or
satisfaction in the response. The level of
commitment is minimal and the behavior lev
is analogous to iumping off the fence, but par
holding on to it and being ready to iump val
back at any moment. de,
wil
hig
-
drudes, (Table 5.2. F ormulating Aff ectiv e Ledr ning Targets continued)
livided
nent.
s. and
W
Valuing: The worth or value a person attaches Examples: Demonstrates belief in the
Ete or to a particular object, phenomenon, or behavior democratic process; is sensitive toward
individual and cultural differences (values
red to This ranges from simple acceptance to
diversity); shows the ability to solve
the more complex state of commitment.
:iiity- Valuing is based on the internalization of problems; proposes a plan for social
llex or a set of specified values. Clues to these improvement and follows through with
values are expressed in the learner's overt commitmenU informs management on
titudes
behavior and are often identifiable. Here, the matters that one feels strongly about
omain student demonstrates that an attitude has Key Words: completes; demonstratesl
t,,,r'ohl, been accepted and is constantly preferred differentiates; explains; follows; forms;
mplest over competing attitudes and values. The initiates; invites; joins; justifies; proposes;
commitment is clear. The student has walked readsl reports; selects; shares; studies;
away from the fence and is willing to be works
identified as someone holding the attitude
or value.
Organization: Organizes values into priorities Examples: Recognizes the need for
by contrasting different values, resolving balance between freedom and responsible
conflicts between them, and creating behavior; accepts responsibility for one's
a unique value system; emphasis is on behavior; explains the role of systematic
comparing, relating, and synthesizing values planning in solving problems; accepts
c- I
The students eventually recognize that professional ethical standards; creates a
conflicts between values arise and must life plan in harmony with abilities, interests
= be resolved by setting priorities on values. and beliefs; prioritizes time effectively to
:).1 I
To do so, students should use higher-level meet the needs of the organization, family,
and self
lI
cognitive thinking, which will enable them
to resolve value conflicts in a logical and Key Words: adheres; arranges; combinesl
,I
defensible manner. They will then have
greater confidence in their decisions. This
compares; completes; defends; explains;
formulates; generalizes; identifies; I
level is a direct Iink between the cognitive integrates; modifies; ordersl organizes;
and the affective domains. ; relates; synthesizes
I lnternalizing values (Characterization): Examples: Shows self-reliance when
Has a value system that controls behavior; working independently; cooperates in
,.'l I
and internalized a value system to the extent
that those values are clearly reflected in the
person's behavior. When we think of a miser
infl uences; Iistens; modifies; performs;
practicesl proposes; qualifies; questions;
revisesl serves; solves; verifies
or a spendthrift, we are thinking of someone
who has reached the characterization
Ievel. That person has reason for holding
particular values and is satisfied with those
:,
,f'4g ,*- n.tessment of Student Learning 1: Cognitive Learning
otor (Table 5 3. F or mulati ng P sy chomotor Ledrning Tdrgets continued)
ften
i,;lllr:iri!14'tati:axi*&t;i*-i1il';r1!;;r**ijnl*s
but - :'1. - l!-qfd*rr,: *!u&16'"-1pii.,
;ical ,$, i
ajor
ion, 5. Naturalization - response is automatic. Arranges; combines; composes; constructs;
The individual begins to experiment, creates; designs; refines; originates;
ind, creating new motor acts or ways transcends
?or of manipulating materials out of
and understandings, abilities and skills
StOr developed. One acts "without thinking."
ion,
K,;ni*.'t''bi ,
lI
l]
!I
t<.
ds
Preporotion
of 1lossroou,t Assesswent
OBJECTIVES
Good tests do not just happen. They require adequate and extensive
planning so that the goals of instruction (objectives), the teaching strategy
to be employed, the textual material, and the evaluative procedures are all
related in some meaningful fashion. Most teachers recognize lhe importance
of having some systematic procedure for ascertaining the extent to which
the instructional objectives have been realized by their students. Yet, some
teachers still prepare their classroom tests with inadequate planning, or worse,
without planning at all.
In class, planning for each lesson and its accompanying evaluation starts
as early as the conceptualization of the curriculum. This practice has been
reiterated by authorities in the field of education, particularly Ralph Tyler
who, until now, is considered the "Father of Educational Evaluation" (oliva,
2001).
50
Compare \
//. pertormance data\
,tion with behaviorally
/ Establish broad \ \q"o obiectiv2/l / collect \
goals and performance
uent \ obiectives.
[ .ri:} \'1 I
'-r'' Laut^.)
/ Planning \ ---\
/ Deve.lop
Classify goals , Stage
or select
\
and objectives. \) measurement
\ ,;;;ii,"",.- ,l
Define \ /rina situatiorttn
in \ / which achievernent\
o bjectives
>ehavioral fr \ of objectives Can /
^--J
aitL-l
_terms. f \*" shown
*
.Z-t Figure 5.1.TyIER,S EVALUATlON FRAMEWORK (Otiva, 2001)
Before a classroom teacher sits down to write the test items, he/she must ll
ask himself/herself a series of questions. what do I want to do? what is the
J]
best way in which I can accomplish my goal? These are the two most general
questions the classroom teacher must consider. The following qulstions ?
should be asked by the classroom teacher in the test-planning stager
. What skills, knowledge, attifudes, etc. do I want to measure?
enslve
rategy
' Have I clearly defined my instructional objectives in terms of student
behavior?
are all . Have I prepared a table of specifications (TOS)?
nance . What kind of test (item format) do I want to use? Why?
rvhich . How long should the test be?
some . What should be the discrimination level of my test items?
L'OrSe,
. How will I arrange the various item formats?
starts ' How long will I arrange the items within each item format?
been ' what do I need to do to prepare the students in taking the test?
Tyler ' How are the students going to record their answers to the objective
Cliva, items?
. How willthe objective portion be graded?
' For objective items, should guessing instructions be given? should a
correction for guessing be applied?
GOAL
lmproved
learning and
instruction
i
r4l
4. Preparing relevant test items
i
tu 3. Selecting appropriate test types
During the stage of thinking about the test, the teacher must consider
the relationships among the objectives, teaching, and testing. The following
checklist should assist the test constructor:
t/.rf."
of Student Learning 1: Cognitive LearninE
AS+f Arr"ssment
.{**'
negative a monitoring agent and can help keep the teacher from straying off
ty- equal the
inskuctional track.
:ts of the
fiich the once the course content and instructional objectives have been
specified, the teacher is ready to integrate them in some meaningful
fashion so that the test, when completed, will be a valid measure oflhe
sfudent's knowledge.
One could delineate the course contents into finer subdivisions.
er-made whether this needs to be done depends on the nature of the content and
est were the manner in which the course content has been ouflined and taught
by
rtee that the teacher. A good rule-of-thumb to follow in determining how
dJailei
ilidity of the content area should be is to have a sufficient number of iubdivisions
to
ensure an adequate, detailed coverage. The more detailed the blueprints
kills and is, the easier it is to get ideas for test items.
cailed a Table 6.1 contains numbers in certain cells under level of complexities
such as knowledge (K), comprehension (c), application (Ap), analysis
rcloping (An), synthesis (S) and evaluation (E). The totiiof the last
column will
amount give the desired total number of items for each levelof complexities.
The
rcsed. It number 50 is the desired total number of items appropriated to different
easy to levels of complexities. The computed values (fijures) in each cell
in a
tuucting certain level suggest the number of items that should be constructed
on
a specified topic.
\
ll
Hence, the five knowledge questions must be taken from each
strucfure of a topic sentence, writing a journal and editorial, methods ,1
before
rt and
, more
.rdents
ves as
tA \o
z o
o
o o-
r{
J
\J ,ri
v
rn
U 3
=
ltJ
o-
o
TU
ra
lr P
fiJ
_o
\d LtJ
c
AJ
i+l *d
,!- <
.L
(U
cr
(U
I
\Lr
'
o
L
t )
;.: L
b U
o
o (tr
G'
=
UJ (!
L
qJ
o-
E.
=
tA
ci
(U
.=
r!
#
o
-o
]J
CJ
P
.e
E
-o
l
c
.9
P
(!
U
IE
U
trro U
o- o
hg +
C?
o
qJ
=u
EP
nJC -o
fE
LOJ #
U(,
,' .: ..;.,::.:'
I
To compute for the Functional Items (FI):
+
PI x number of items allocated to each level of
complexities/l00
Example: 11.25 x5: 56.251100 : .562b
o
N
o-
J
U
f
= 1. Based on the sample two-way Tos, identify where the items in the
t! different levels of complexities should come from.
C 2. compare the objectives of a lesson plan with the items in a test paper
U
E of a teacher in an elementary or secondary school.
AJ
,L
l 3. Interview teachers on how they prepare a test for their students.
q
AJ l]
o.r
q
II
cU
1
o
=o
q
d)
o
0)
.=
ag
?
c
r!
A)
F)
l^l
rl
:l3l
'hl
pl
GI
Jl
.l
=l
9t
+l
nl
I
-l
JI
-l
-l
-t
/l
Eeuelopwent
of Clossroou,t Assesswent
OBJECTIVES
,,Vultipb-ehoice eest
The multiple choice type of test is a form of assessment in which students
are asked to select the correct or best answer out of the choices from a list. In
this kind of test, an item consists of two parts: the stem and a set of options or
alternatives. It requires the students to select from the options that will make
the stem complete and correct. All incorrect or less appropriate responses are
called distracters or foils.
Oftentimes, multiple choice tests include a stimulus material where the
item or question is drawn. A stimulus material, or an introductory material, is
added information in the form of chart, graph, stanza of a poem, or a novel
pictorial.
The stem is the beginning part of the item that presents the item as a
problem to be solved, a question asked of the students, or an incomplete
58
statement to be completed. It can be presented in three ways: a direct question,
an incomplete statement, or a mathematical equation. If it is an incomplete
statement, allthe options or the last one ends with a period. For elementury
0t/4€/tt sfudenh, it is advisable to use a direct question.
F
r
7. The effects of guessing are largely reduced since
options.
ther e are more
5. Contained Options
Example:
Identifu the error in the sentence.
My parents was in Manila to assist my sister enroll in College.
abcd
No error
e
6. Correct Answer
Example:
What is the summer capital city of the Philippines?
a. Baguio City C. Davao City
b. Cebu City d. Olongapo Cip
7. Group Options
Example:
fJ Write-
"lr
A if the item is a simple sentence.
B if the item is a compound sentence.
C if the item is a complex sentence.
D if the item is a phrase.
E if the item is a clause.
8. Morse Variety
Example:
Write-
A if W affects X but X affects Y but Y aff.ectsZ.
B if W does not affect X but X does not affect Y but Y does
not aff.eclZ.
C if W affects X but X does not affect Y but Y affectsZ.
D if W does not affect X but X affects Y and Y does not
aff.ectZ.
**,8,1'
.&ff$ att"ssment of Student Learning 1: Cognitive Learning
'WY'
Table 7.r
CHECKLIST FOR WRITING MULTIPLE-CHOICE ITEMS
servlng
P,
1. Are the item and the main probrem in the stem crearry presented?
Has the item been cast so that there is no repetition of key
words or
phrases for each option?
t/ u1.
1: Cognitive Learning
ffi5zy'essessment of Student Learning
.Ftx'
rJy agree
4. Limit each statement to the point that is being
tested. Avoid equivocal
items.
rip of the
r appear 5. Avoid excess use of negative words and phrases.
in print, 6' Approximately half of the statements should be farse. Because it is
easier to construct hue item, teachers inadvertenfly
include -o*
statements that are true. The chance of getting
correct answers by
guessing is higher since students who
urZ in ioubt would t"na io
mark the item as true.
rasis on
7. Avoid qualitative terms like best, some, many, and
several.
dge for
rticular Modification of True or False Tests
ent can C orrections/or G uessing
p€r se. sfudents can be penarized for guessing since guessing
raJysis, reflect learning or hue performan.". A.gr.Ints does not
rations for guessing include, among others, the fillowing,
--' of corrections
in- fu'uo,
implifu
eacher a. Equate the scores of students who guess with those who work
,mllex more carefufly under restricted time rimits. Under
severe time
limits, some stu.dents may work rapidly
"ri *;;. items, whire
some may work more slowly and deliberately.
on
gs are
b. Discourage shrdents from guessing, thus facilitating rearning. on n
moral grounds, guessing is tantamount to becoming
es are
r false
and unjustsince sfudentswho guess take
dishonest
advantage of the nature I
of r-F tests and murtiple-choiie tests. on p"augogicar
guess grounds, 1
chance scores may reinforce guessing, so
lf any sfudents may get some
items right even without studying.
c. [mprove the extent to which tests are capable of predicting criteria.
corrected scores correlate highly with ihe.iit"ii"
y.an of scoring than
incorrectness.
t/ers.
Arguments Against C orrections
>e to /or Guessing
1. Sfudents may be discouraged from attempting
to answer even though
they have some information.
2' students who make effors whether they guessed
or not can be
ents
penalized.
sht- 3. The corrections for guessing are laborious
to use.
4. The differences in scores may still go unnoticed
even if the test items
sto are increased.
ays,
Reducing the Effects of Guexing Using Other Ways
1. Encourage the sfudents to guess. If the
students are tord that scores
will be conected for guessing, some may not
attempt to answer some
1. SimpleTrue or False
write TRUE if the statement is correct and FALSE if otherwise.
Example:
'I Baguio City is the summer capital of the philippines.
.-Jtl
...,4t.e{.a'
.9t Assessment of Student Learning
1: Cognitive Learning
,r,ir6..8
,.r
lcontinue C if bcth statements are TRUE.
age them D if bcth statemenis are FALSE.
dagogical (t) Ree ognition tests require the examinee to choose the right answer
frorn the given opiions. (z) An example of a recognition test is a
cornpletion test.
r T-F test
pice-tests 6. Fact ar Opinian
time limit Example:
ldentify if the staten:ent is a FACT or an OPINION.
'i. There are 12 months in a year.
udents to
reducing
L During the month of March, it never rains.
nainty in Cirrle the r,vord(s) in any part of the paragraphs that make(s) the
statement(s) \.vrong.
To a large extent, the gracie a student gets may not truly reflect the
autneniic learning outccrrie due to the flaws in the test construction and
adr-ninistration. Ti-:ere are certain quaiities a good test must possess.
Reiiaoiiit,v imolies ,yaiidity but noi reverse. The latter refers to the
efficiency with lvhich the test intends to measure. One of the factors that
infl,-rence validiilr is directions, 'which inform the teachers on how to score
students' responses to the items. The arrangement of the items must begin
from ijiflicuit ic easy. hle cessarily, test is administered with complexity, clarity,
and uniformitii" I
statement Table 7.2 1
CFIECKTIST FOR WRIT'NG TRUE OR FALSE ITEMS I
Factcrs ..,'llY,A3ir
1. Was each item expressec in clear, simple language?
z. Was lifting statements verbatim from the text avoided?
statement
3. Have negative statements L:een avoided where possible?
4. Have specific determiners, such as ail, may, and sometimes been avoided?
5. Have double-barreled items (part true and part false) been avoided?
6. Have trick quesiions been removed?
7.' ls each item clearly true or false?
8. ls there approxirnately the same number of true and faise items?
,,l4atdrirg-Cyg eest
The matching-type test is similar to the multipie-choice test. In this kind
of test, the examinee associates an item in one column with a choice in the
second coiumn.
ther
1. Luneta Park a. Albay
also 2. Mines Mew Park b. Manila
also 3. Chocolate Hills c. Banaue
4. Camp John Hay d. Bohol
e. Pangasinan
5. Intramuros f. Baguio
g. Palawan
Chapter 7: Development of Classroom Assessmen ,
*i'r#"-
'ffs
/
-
3.SequencingMatchirigrequiresiheexarnineestoarrangethings,
steps, or events in ciironclogicai order'
Arran ge th e st"t', - ti h istc ricai rese arch'
:, J;;'TJ;"d
2. Gathering of source materials
o
I
J. Problem formulation
4. Criticizing soul'ce materials
( Inteq:reting hist,crical data
4.MultipteMatchingrequirestheexamineestomatchtheitemsin C
B to colurnn
colurnn a to g, ittZn match the answers from column
and further match answers frorn colun:rn C to coiumn
D'
towns
Match ihe provinces listed in coiumn A with their capital
inColumneanawiththetouristspotstheyareknownfor.
Colurnn A Colurnn ts Colurnn C
1. Bohol a. TagaYtaY CitY I' Underground
2. Camarines Sur b. Tagbilaran Ci$ River
3. Batangas c. Puertc Frincesa itr' Taal Volcano
4. Palawan d. Piii ill. Water SPorts
Batangas ei$
'rrJ ru. 8iH"',"J- n*,
1,'1 MaYon Volcano
L"i Table 7.3
.l ,-i i;
hr
CHECKLIST FOR WRIT!NG MATCh!ING-TYPE TEST
Yes
4.
i
i
I
i c.
i
(,
--"] l
I
a
9.
1r\
l" u.
of items, restrict the number of
i1. answers should be equal in length.
19
LL,
Types of ComPletion Tests
1. ldentificationTest
giving
It refers to the process of summing up the results of tests,
2. Enumeration
List down the three branches of Philippine government.
1.
2.
3.
3. Fillingthe Blanlcs
Bayang Magiliw
Perlas ng
Alab ng puso
Sa mo'y buhay.
hinirang
DuYan ka ng
r,
J Sa marnlulupig
'Di ka
\J
"
L
"-, -;
4. Analogy
Father: Son, Mother:
1loze Cest
consisting
cloze or cloze deletion test is an exercise, test, or assessment
and the students
of a portion of a texi with certain words removed lclozetext)
requires the ability to
are asked to replace the missing words. The cloze test
the correct words or
understand context and vocabulary to be able to identify
type of words that belong in the deleted passages of the text.
-'- -Word.
may be dele-ted from the text in question either mechanically
to
(every nth word) or selectively, depending on what aspect the test intends
give emphasis to.
Example:
Today I went to the and bought some milk and eggs'
and
I knew it was going to rain, but I forgot to take my
ended uP getting wet on the waY
-'
of student Learning 1: cognitive Learning
6;.ioa$'o*.ssment
{I
€xoy eest
This type of test differs from the completion test in degree rather than
sts, giving
in kind. Essays usually allow greater freedom of response to questions and
require more writing.
it
Orn{ Qusstion
Bii
Advantages
ii.r,: from
1. Both perrnit the exarniner tc rietenain* h*w weii the st'i-r.cient can
synthesize and organize hisiher idsas and express himself/herseif.
anci use
2. Both are not depende;"1t, as ihe n-luitipie-choice tes.t, on the ability
of the pupil to r*c*gnize tFre ecrrect answer: boih require that thl
students know and are able ic suppiy th* correet *nr*n..
3. Both perrnit free responses by th* sturients"
Lirnitations
1. Both provide for a variet-v of limited sarnpling content.
2. Both have iower rate reliability.
ICNS, I
€ tests
,
Setting Criteria
ls and
The foilowing sr,rggesiicl-ls are heipf..i! in developing rubrics for
! day,
essay tests:
'ei's to
r iinal I " i he descriptions rnust focus on the important aspects of an essay
i- use rssponse.
o
L The type of rating {holistic or anaiy'tic} musi match the purpose
;ently of the assessment.
::tent
J, The descripiions of the criteria n:urst be directly observable.
i: rha A
t+ Ensure that the criteria are understcod by ihe students. parenrs,
s anci and others.
5r. The characteriEiics aiid iraits used in the scaie shouid be clearlg
and specificallg defineC.
5. Mininiize €rrors in scoring. These errors may he generosity errors?
centrai tei'rdency errors? and severit r errors.
7. Make the scorinE system feasibie.
'-.r-:.tr--ri:r I l.,r:r,.:i,:i:,,r,.r;,1 i i i.:i*:.aii,-r,,.ia Arr-:i:::5iii.ii.t!
e.1
c:1
Setting Performance Levels ,Table 7.
Table7,4
SAMPLE RUBRIC FOR PERSUASIVE ESSAY TESTS
Worr
ffi
Claim was I made a claim I made a lmade a claim I did not make a
made. and explained claim but did but it was claim.
why it was not explain confusing or
.J controversial. why it was unclear.
controversial. Senl
U,
',;i'.
Reasons I gave clear, I gave reasons I gave one or I did not give fluer
were given accurate reasons in support of two reasons convincing
L1 in support of in support of the the claim, but which did not reasons in
the claim. claim. overlooked support the support of the
important claim well. I claim.
reasons. gave irrelevant Con
or confusing
reasons,
Reasons Ithoroughly I discussed lacknowledged I did not give
were discussed the reasons against that there were reasons against
considered reasons against the claim, reasons against the claim.
against the the claim. but left out the claim but
claim. important did not explain
reasons and/or them.
I did not explain
why the claim
still stands.
)lTnance
egree to
ffiW
r"Jlr ;'*i
'iri{P.ffi** ^,a**,"i
:ord and
Organization My writing was My writing had a My writing was My writing was
wellorganized; clear beginning, mostly organized aimless and
ntitative had a compelling middle, and but got off disorganized.
50na opening; had end. I used an topic at times.
e,Ather a strong, appropriate It had several
informative
s ?nt
body; had
paragraph
format.
paragraph
format errors.
a satisfying
conclusion;
and had an
appropriate
paragraph
format.
l
ffi Word choice The words I used I mostly used My words I used the same
were striking, routine words. were dull and words over and
natural, varied, uninspired and over. some
and vivid. they sounded like words were a bit
lwas tryingtoo confusing.
hard to impress.
Sentence My sentences I wrote well- My sentences I had many run- 1
fluency were clear, constructed were often flat ons, fragments
complete and but routine or awkward. 1
and awkward
of different sentences. There were also phrasing, making
lengths. some run-ons my essay h.:,'d to
,
and fragments. read.
Conventions I used the first My spelling Frequent errors My errors in
person. I used was correct on were distracting grammar,
correct sentence common words. to the reader capitalization,
structure, There were but did not spelling and
grammar, some errors in interfere with punctuations
punctuations grammar and the meaning of made my paper
and spelling. punctuations. my paper. hard to read.
Table 7.5
cHECKLtST FOR WRtTtNG ESSAY quEsTtoNs
rhe student?
4. Are the questions novel? Da they challenge
Kru;rw €xsrtisrs
typcs of test:
1. Forrnuiate sampie iterns {or the foliowing
a. Muitiple Choice with Stimulus Material
b. Ctruster True or False
c. Sequenc.ingMatching
d. Filling the Bianks ComPletion
.lr e. Essay
your Field Study
Ask for or borrow sample tesi qu*stions from
I
\J r-l r)
L.
',. i"".t","r, and anal-vze how ihe questi':ns were $orrnulated'
r! il
Lear'ning
g4 :: nssessrnent of Sir:Cent Learning 1: Cognitive
Yes.r
Itew Anolysii
* !Xgffiirjtri;;::::.,:u,rii - l' .,,.,,.::..:r,.:rli.,:.::;i&
..,ij4?ffiifrtrl*
OBJECTIVES
LJ,
8. Add the number of individuals who responded correcfly to the item
(upper 27To,lower 277o, and middle 46%) and enter the data in the
;il"-.'
L+ eighth column.
9. Divide the value in column 8 by N, the total number of examinees, and
enter the value in the last column. This is the proportion of sfudenb who
responded correcfly. The quotient is the index of discrimination.
Below is an example of an item analysis for item number one with five
options. l*lter B was the correct answer, and 75 sfudenb took the examination.
Table 8.t
SAMPLE OF AN ITEM ANALYSIS
ffi*;* &ffiffiffiffi#ffiffi
1 A
*B
2 3 3
13 7 25 45
C 3 9 2
D 1 o 4
E 1 1 1
d/,{/'
gg5T{assessment of Student Learning 1: Cognitive Learning
,f,;8t8'r
Table 8.2
TABLE FOR !NTERPRETING INDEX OF DISCRIMINATION
Ieaving the (o rneulAR vALUEs)
nd bottom
s into two
etween the Questionable ltem
her groups -o.59 - -o.2o
-o.21 -
pr.ded to O.20 Moderately Discriminating
m .or the o.21 - 0.60
o selected
tumber of
rfie value Table 8.3
TABLE FOR INTERPRETING INDEX OF DIFFICULTY
umber of (e rneumR vALUEs)
ned is the
nade the
n.
the item o.41 - 0.60 Moderately Difficult ltem I
lta in the
o.6r - o.8o I
rees, and o.8r and above Very Easy ltem I
enb who
'I. eorupatotton o/ the D talue (lnder of Disaiwinatun)
rJ :9
20
= 0.3
. The cornputed D value is 0.3, which is interpreted as a
discriminating item.
=
. e',
Chapter B: Item AnalVsis .f g7ff
.;.{S' -:i.rd
.:,,..{ t "
r
1owputatiou of the P talue (lnder of Di/fnulfu) f - - .
1. Determine the total number of students who got the correct answer on
the item from the UPPER and LOWER2770 and from the MIDDLE
]- -
467".
2. Divide the sum of the total number of students who got the corect
answer from the total number of students who took the examination.
,
ty vv'sv
ualue:
R where R is the total number of students who got
-N the right answer
Table 8.4
DECISION TABLE
"J lmprobable-Discard
Difficult Not Discriminating
Moderately Discriminating May need revision
{nJ'
Discriminating Accept
*i;"
L+ Moderately Difficult Not Discriminating Needs revision
Moderately Discriminating May need revision
Discriminating Accept
Easy Not Discriminating Discard
Moderately Discriminating Needs revision
Diicriminating
30
ge item. TG=-xS?+50
/5
= 0.4x S0 + S0
:70
Another way of interpreting test results is through ranking. Ranking is
considered the first step in test score interpretation. It is the ur.lung".n"ni
the scores in order of magnifude or size. Determining the rank simpl-y involves
of I
listing of the scores from the highest to the lowest,*.". I
Example: ,
Scores Rank
(1) 78 1
(2) 74 2
(3) 73 3.5
I finding (4) 73 3.5
re called
(5) 68 5
he pupil
,rTnance (6) 66 6
,' in the (7)
element 65 8
ormula, (8) 65 8
nring is
(e) 65 8
o weigh
may be (10) 57 10
est by a
The ranks assigned to the given scores are on the second column. The
highest score, 78, received a ranking of 1, and the next is 74, ranked 2. The
next two scores, 73, ranked 3.5, in which case the average of the ranks that
would be assigned to them (enclosed in parentheses) if they were not tied,
would be given to each. If unequal, their ranks would be 3 and 4; tied, the
average of the sum of 3 and 4 becomes their ranks. Similar to the scores
ranked 8, the average of their supposEd ranks if they were not tied, will be
given to them. The last score; 57 ranks 10 or N.
Ranking has its uses and limitations. As a means of cornparing scores
within a group, ranking is a simple, readily obtained measure that has some
value. It is useful in checking a group's performance in a test. On the other
hand, many educators feel that the use of ranking tends to overemphasize
individual competition to a greater extent than the practice of assigning letter
marks. Moreover, ranking fails to indicate the extent or amount of diff"rurc"
in the achievement of the students being compared.
Graphing or tabulating data aids in easing the interpretation of test scores.
At times, teachers want to present scores in tabular or graphic forms. Tabular
or graphic forms give a fairly clear picfure of how the students performed.
Frequency distributions, histograms (bar graphs), frequency polygons,
or cumulative frequencies or percent curves can make data interpretable.
Tabulating, however, may result in an appreciable sacrifice of accuracy.
-l' Percentile and percentile ranks may also be employed in interpreting
.{r t;
tri test scores. Percentile is defined as a point, which a certain percent of the
scores fall. A percentile rank gives a person's relative position or the percent
'r, ' 'l
of the students' scores falling below his/her obtained score. To illustrate the
computation of percentiles, consider the data in Table 8.5.
Percentiles should not be confused with percentage. Percentiles have the
advantage of being easy to compute and interpret. In explaining a national
norrn percentile rank to a sfudent, one would say, for example, "your
percentile rank is 80, meaning you have obtained a score higher than gO out
of. every 100 sfudents in a representative sample of fourth year sfudents in
the nation." If it is an 85 percentile rank, one can interpret that the sfudent
belongs to the upper 207o of those who took the examination.
The most powerful means of interpreting test scores is by statistical
analysis, such as measures of central tendency, measures of variability and
measures of correlation.
Table 8.5
DATA ILLUSTRATING THE COMPUTATION OF PER€ENTILES
"dr#lAssessment
ziWtf
'
r'*$,ra
of Student Learning 1: Cognitive Learning
rt
g .res 't2 5o 4 38 72
as some 11 55 5 34 63
he other 10 5o 7 29 51
:rphasize 6 22 38
9 45
rng letter 't6 27
8 4o 5
dlerence
11 19
7 35 3
6 z 8 14
il scores. 3o
6 10
Tabular 5 25 2
rr\med. 4 20 1 4 7
oiygons, 3 15 1 3 5
pretable. 10 2 3
2 1
ac!. 1
,t
1 I
1 5
zrpreting
o
nr of the o o o 1
I
: percent I
n-,ate the
have the
national
With a complete set of test papers with multiple choice items that has
been administered to the students, find the index of difficulty and index of
e" "Your
discrimination of each item and the entire test.
out
u', 80
uients in
e *udent
stical
h, rn , and
Fs
F
E *el,.f,.);
Chapter B: Item AnalVsisgf 91agf
,f&tv'
7
i
Sroding aud Keportiug
?roctrces
*{$*}kX$ *tt*tt*irattteit
'is:*I.:}*ii3*ll:1}**1*:td*q*f
OBJECTIVES
""il1t
Examinations and marks have always been closely related in both
the
teachers' and sfudents' minds. Periodic and monthly examinations
usually
ri have_ a major part to play in the students, final grades.
students, parents, teachers, administratorq prosp ective employers,
student admission officers all need information from t-he schoolto
and
aisist ihem
in decision making. sfudents, primarily should receive such information
through daily interaction with their teachers, although formalperiodic
reports
can also help students when making decisions.
Many educators and sfudents believe that recording of grades motivates
sfudents to learn things they would not otherwise learn.-
Marks rcfer to those systems that use summary symbols of some type.
If only a single symbol is assigned, it should ,"pro"nt achievement in the
subject matter, not attitude and status rather ihun gro*th. It is also
an
objective judgment of one person (student) by another"(teacher).
Below are
some questions regarding r4arking:
1. Are marks an effective conveyor of information about the sfudent,s
achievement?
2. Can anyone achieve the mark helshe wishes if helshe tries hard
enough?
3. Are marks the means or an end to a student's achievement?
4. Is there any correlation between school marks received at
one level
of education and marks received at another?
138
5. Do marks bear any relation to success in life?
6. Do marking practices provide a justifiable introduction to competitive
adult life?
"ti/t{ ;,
Grading methods communicate the teachers' evaluative appraisals of the
sfudents' academic achievement and performance. In the process of grading,
teachers convert different types of descriptive information and various
measures of the students' academic performance into a single grade or mark
that summarizes their assessment of the sfudents' accomplishments (Guskey
& Bailey, 2007).
Grading, on the other hand, is the process by which a teacher assesses
sfudent learning through classroom tests and assignments, the context in which
good teachers establish that process, and the dialogue that surrounds grades
and defines their meaning to various audiences (Walwood & Anderson, 1998
how
p.1).
ages
,Nature ( Smdes/J4arks
Grades are the teacher's judgment on the performance of students based
on certain criteria. Although objective, grades can be subjective from time
to time and are also relative from one school to another school, from one
oth the teacher to another teacher, and from one sfudent to another student. Several
usually variables, such as periodical examinations, class standing, and projects, are
considered in grading the students.
rs. and
st them Tunctions o/ 6 rndes/rVorks
mation
reports The school can never escape relative judgments about sfudents. Grades
or marks have certain functions to perform and these functions are served
)tivates best by an unbiased grade. Such functions are as follows:
1. To help guide the students and the parents with respect to future
,e tgpe.
educational plans;
i in the
ilso an 2. To help the school decide upon a student's readiness to enroll in
certain selective programs or courses;
low are
3. To help higher educationallevels appraise an applicant's acceptability
for the program being offered; and
udent's
4. To help a potential employer decide on the suitability of the student
for certain jobs that depend on academic skills.
s hard
Grades or marks are necessary for guiding the sfudent in his/her school
work, understanding his/her personal trials and tribulations, helping him/
re level her plan his/her educational and occupational fufure, and cooperating with
Pu ryoses of 0 rodes/,,4harks
Grades may serve the following purposes:
1. Administration admission, selection or grouping, promotion,
retention, dismissal, fit for graduation
2. Guidance - diagnostic-readiness, prediction of success, remediation,
validation, career-guidance, psychorogicar assessment
3. Motiuation - skill mastery, goal-setting, positive mobilizer
C{pes o/ 6 rades/,,4,torks
1. Percentage system (Ts-100).lt is often used as it is easily
and
universally understood. It implies precision of judgment that is
[ardly
attainable by most measuring instruments.
2' Poss or Foi/. This is good for survey subjects or vocational
courses
and higher level courses in exact disciplines like math and physics.
The most common justification for the P-F system is that it encouiages
sfudents to take the courses they would otherwise not take
b".urr"
of a fear of lowering their grade point average (GpA) or general
average. This system also reduces student anxiet5r, gives sludents
greater control over the allocation of their sfudy time,
and shifts the
sfudenh' efforts from grade-getting to leaming.
3. Fiue Point Multiple Scale. This enables one to categorize students.
Examples of^this grading type are A, B+, B, c+ u'i c f,-i.is,
1.5, 1.75, 2.0, etc. ",
..*,;r&:t
S l4Oynssessment of Student L'earning 1: Cognitive Learning
suitably be 4. Duol sysf em. (Any combination of the previous three types of
ormulating grades/marks). It may be a letter gr {e or percentage system for
academic subjects and P-F for n .academic and vocational
courses.
5. Checklist and Rating Sco/es. These rating scales or checklists should
include the major cognitive (psychomotor) for each subject matter
area. Checklists or rating scales on affective objectives should also be
romotion, developed. This type of grading is appropriate for early elementary
grades.
nediation,
,r4duan tages o/ 6 roda /,,4,40 rks
Some of the advantages of marks are as follows:
follows: 1. Marks are the least time consuming and most efficient method of
s to their reporting.
2. Symbols can be converted to numbers. Thus, generalaverage grades
s for self- can be computed. Generalaverage grades are usefulin many types of
selection, placement, and classification. They are the best predictors
of success in fufure education.
3. Marks relate not only to chances of obtaining good grades in future
courses; they also relate somewhat to achievements beyond school.
d
4. Marks serve as an overall summary index. students want and need
rpropriate
to know how they did on each separate subjects, as wellas how they
performed on the whole.
f
Awragirg Scores to Deterwine n Qrade
If the purpose of grading is to provide an accurate description of whar
students have learned, then averaging scores from past assessments with
measures of current performance is inappropriate. Relying on data from past
assessments can give the wrong information regarding the student's progress
in the learning process.
q
Below are some guidelines for deciding what evidences or combination
of evidence represents the truest, most appropriate summary of the students'
achievements and performance. i_
r. 1. Give priority to the most recent evidence. Scores from assessments at
#l the end of the marking period are typically more indicative of what
,. l{
the students have learned than those gathered from the beginning. c'litci
}i
rl 2. Give priority to the most comprehensive evidence. If certain sources l
rJ of evidence represent cumulative summaries of the knowledge and
skills the students have acquired, these should hold the greatest
weight in determining the students' grades.
3. Give priority to evidence related to the most important learning goals
or standards. Rank the evidence gathered in terms of its importance
to the course's learning goals or standards.
l.lse of Zeroes
(
. ,Yi,'
There are two ways of computing the final grade: the averaging and
cumulative grading systems.
The aueraging grading system treats the students' performance in each
ent or grading period independently. Table 11.1 shows a student,s academic
performance per quarter in English. With 78, 86,82, and g4 in each of the
four quarters, his final grade based on averaging is 82.5.
.,.,. .,{):
Chapter 12: crading and Reporting eractices_
fljY
{'''
,"'*
Table lt.t
ILLUSTRATION OF THE AVERAGING GRADING SYSTEM
English tB
ffi ffi
86 8z
srl:<$*;eYJ*s
S**Wra'*
a":;. .r* :q!
i!i-*i";&t-q-in
8+ 82.5
83 88 8r
Math l6 77
Science 79 8S 93 8t 86
Filipino 8+ 88 88 91 8l.ts
'88.r
MAKABAYAN 86 89 86 93
85.r5
General Average
#l
t Table tt.z
l lt ILLUSTRATIoNoFTHEcUMULATIVEGRADINGSYSTEM
hl
3l
rd
English l8 (83) 8z (zg) 8o (83) 8z
ffi
8z
General Average
The.tentative grade for the second quarter is 83 and the final second
grade for the second quarter is 82. To compute for 82:
78 x .30 :23.4
83 x.70 :
58.1
23.4 + 58.1= 81.5 or 82
*','
..e,/;;fd
XiftYntt"ssment of Student Learning 1: Cognitive Learning
,,,';df"rp