Classroom Principles
Classroom Principles
Classroom Principles
Two important principles in test construction are validity and reliability According to Gronlund
(2006) “assessment in the teaching and learning context are the methods used to determine
A valid test means that the test measures what it claims to measure (Hughes, 1989 and
Brown 2001). As such when a test is clearly linked to the stated objectives it would be considered
a highly valid test. For example, a form 4 accounting test on the steps of the accounting cycle. The
test should look at the steps of the accounting cycle, source documents, recording in journals,
posting to ledger, extraction of trial balance and financial statements of a business. All the above
areas in the test would make it highly valid whilst any other areas on stock valuation or ratio
The validity of a test can be estimated by utilizing face validity, content validity, concurrent
For content validity the test has to cover all relevant areas of the subject taught and student
behavior involved in tasks. For example a test to form 4 POA students on methods of stock
valuation FIFO, LIFO and AVCO. The test must cover all methods taught and if any method is
left out the validity of the test maybe compromised. Also, if the test includes other areas than
methods of stock valuation such as cash book and bank reconciliation statements the test would
not be valid. A table of specification would ensure that all content will be tested. Construct validity
ensures that the test measures the concept it is intended to measure (Weir, 1990).Face validity
refers to if the test looks valid “on the face of it”. For example a test on updated cash book and
bank reconciliation statements where questions show this. Another teacher not in accounting who
is untrained would expect the test would measure what the accounts teacher claims it to
be.Criterion validity is where the students’ results match with a different test of the same concepts.
According to Bachman and Palmer (2006) the reliability of a test has to do with the
consistency of the test, its reproducibility and the students’ performance on the test. For example,
a test with high reliability administered to a student on two separate occasions would ensure that
the students’ performance remained the same. If this does not occur, it would mean that the student
scored different marks across two test administration making the test poorly reliable.
The methods of computing test reliability can be via test – retest, interrater parallel, decision or
internal consistency.
The test- retest is where the same test is given to the same group of students twice with a
time interval (not more than 2 weeks) between the two administrations of the tests(Wood et al ,
2017) The test scores are correlated and this ensures the test has reliability. A high correlation
questionnaire produces similar participants’ response to all items in the questionnaire. Parallel
reliability occurs where the teacher creates two different but similar tests that measures the same
construct.
References:
Bachman, L.F. & Palmer, A. (1996) .Language Testing in Practice. Oxford University Press.
Brown, H.D. (2001) .Principles of Language Learning. 4th ed. White Plains, New York: Longman.
Gronlund, N. E. (2006). Assessment of student achievement. 8th Edition, Pearson Education Inc.:
Boston.
Wood, D., Harms, P. D., Lowman, G. H., & DeSimone, J. A. (2017). Response speed and