Lit Review - Effectiveness of Computer Based Testing
Lit Review - Effectiveness of Computer Based Testing
Mia Ayres
Steven M. Bradbury
Sara K. Maglov
Matthew T. Sheehan
Introduction
is on the rise in many educational settings. Over time, the number of students enrolled in schools
has increased and as a result the need to be able to assess more efficiently has risen as well
(Karay, Schauber, Stosch & Schuttpelz-Brauns, 2015). Simultaneously, technology has evolved
to provide teachers and course administrators with more options to meet the needs of a growing
The use of computer based assessments has clear cut advantages for educators. The use
of computer based assessments provides teachers with a more timely way to administer, review
and grade assessments (Karay, Schauber, Stosch & Schuttpelz-Brauns, 2015). However, the
benefits for students may not be as straightforward. While today’s students have an increased
ability to utilize technology for a variety of purposes, the impact on students and their
for debate. There is also a need to explore how students feel or perceive these changes to
Problem Statement
There is a growing trend in the education community to use technology in the classroom
for a variety of purposes. The use of computer-based assessments is one trend that provides a
less expensive and more time efficient way to assess student understanding (Baig, et. al, 2018).
EFFECTIVENESS OF COMPUTER BASED TESTING 3
This literature review aims to consider the impact the use of these types of assessments have on
students’ performance, specifically their achievement scores. In addition, this review explores
how students perceive the use of computer based assessments in place of more traditional
achievement measures. Furthermore, relevant literature related to the benefits and the limitations
Review Questions
This literature review will focus on the ways in which computer based testing impacts
student performance, defined as test scores, as well as student perceptions. This review will
address the following questions: (1) In what ways is student performance impacted by the use of
computer based assessments? (2) How do computerized tests positively affect student
performance on standardized tests? (3) In what ways can computerized assessments be limiting
for a test taker? (4) What are student perceptions regarding computer based assessments?
This paper will consider the results from researching the effectiveness of computer based
tests as compared to paper-pencil tests. The effects that the different types of tests have on the
outcomes are called the “mode effects” (Oz & Ozturan, 2018). Many in the world of education
are concerned with the mode effects because the consequences impact not only the students, but
the school staff and community (Backes & Cowan, 2018). With the growing trend to increase
the use of technology in the classroom, many assessments and high stakes tests are now
administered online. Questions are surfacing about how a computer-based assessment might
impact the results of these assessments. The comparison is frequently made to paper-pencil
EFFECTIVENESS OF COMPUTER BASED TESTING 4
assessments which often preceded most of the current computer-based versions. Results are
mixed, especially when considering different content (Oz & Ozturan, 2018). The review of
literature will center on the development of three specific stances that emerge from the mixed
results: computer-based tests are more beneficial, computer-based tests are detrimental, or there
is no significant difference when comparing the results from computer-based and paper-pencil
tests.
This Literature Review will explore three arguments related to the use of computer-based
assessments. Some of the literature discussed will present data and information that shows that
performance and is viewed positively by student participants. Other articles reviewed in this
paper will present information that shows the use of these types of assessments are detrimental to
student performance and are perceived as negative by students. Finally, we will review articles
that show there is no significant benefit or harm to administering assessments via technology on
either student performance or student perception. The conclusion of this literature review will
synthesize the information gathered and provide answers to the review questions presented.
Literature Review
Background of Issue
Modern day education is changing drastically in an attempt to keep up with the rapid
pace of technological advances. Now, more than ever, society is pushing to integrate technology
in the classroom as much as possible. At the same time, standardized tests remain a prominent
factor in education. Many schools around the world are administering computerized versions of
EFFECTIVENESS OF COMPUTER BASED TESTING 5
these tests. Computer Based Tests (CBTs) are predicted to become the primary testing mode of
the future (Ghaderi, 2014). As of 2016, two dozen states administered tests online while most
other states are work toward transitioning to this mode (Backes & Cowan, 2018). While digital
assessments offer many advantages, such as faster score return, not much research has been
comparison to paper and pencil based tests. The purpose of this paper is to summarize all
arguments on this topic that currently prevail in scholarly literature. Opinions on the topic vary
drastically. Professionals have argued that CBTs are immensely beneficial to students while
others seem to think CBTs can be detrimental to student success. At the same time, other studies
have shown there are no significant consequences to administering CBTs versus traditional
testing methods.
One way to prove that technology based assessments have a helpful impact on
individuals is that anxiety levels found among students was found to be less than those
individuals who took a paper and pencil examination (Baig, 2018). One may think that trying
something new, such as computer based assessments, may cause stress and anxiety among those
testing. This however, does not seem to be the case. Test scores in some cases, have actually
improved when taken online. In one Florida state assessment (FCAT), students performed higher
on the Math and Reading Assessments taken online versus those who used paper and pencil
(Martindale, 2005). Students from the deaf and hard of hearing population have also strived in
being able to access online assessments along with paper and pencil (Cawthon, 2015). A final
way online assessments are beneficial is virtual contexts help students in answering multiple
EFFECTIVENESS OF COMPUTER BASED TESTING 6
choice questions correctly. It also helps them better understand their own knowledge and
learning process (Schifter, 2014). These online assessments have a way of connecting to the
students needs and wants. These are the many reasons that show online assessments are
comparison to their counterpart, traditional paper-based tests. Many factors contribute to the
creation of a valid and reliable assessment. It is crucial that all tests, regardless of format, are
fair and equitable for all test-takers. Some studies have indicated that online assessments are not
always the best way to evaluate student performance. Factors such as test design, comfortability,
literacy, and question type can result in lower test scores. The following information outlines
Studies have shown that online assessments can result in significantly lower scores
particularly in Mathematics and English Language Arts (Backes & Cowan, 2018). Reading
comprehension tests administered online can result in lower scores than traditional assessments.
In a recent study, it was found that students who took a computer based test scored 10% of a
standard deviation lower than students who took the test on paper (Backes & Cowan, 2018).
Scores on reading assessments can be lower on computerized tests than paper and pencil tests
because lengthy text can be cumbersome to read and comprehend on a screen (Kim & Huynh,
2015).
EFFECTIVENESS OF COMPUTER BASED TESTING 7
The design of computerized tests have a large impact on its reliability. Question type on
computerized tests can be narrow. Multiple choice prompts are easier to create and grade. With
that said, computer based tests may not offer as many application problems because they require
more background knowledge from students and training from instructors (Zhao, 2012). Limited
space for scratch work on computerized tests can also be restrictive. Some students have trouble
transferring their scratch work from paper to the testing response box on computerized tests
(Backes & Cowan, 2018). Students must be skillful in editing and entering data on the testing
Different versions of testing questions in each mode also raise a concern. One study
compared the differences of the same question offered on a paper test and a computerized test.
The paper test offered a passage of a reading while the computer based test presented the entire
text. Students taking the CBT were expected to know that they needed to scroll through the text
in order to read the passage. If a student did not comprehend that they only needed to read a
portion of the article they likely spent a lot more time on the question than students taking the
concern for many, is how to finance these tests. Who will cover the cost of paying someone to
grade free response questions? Who will fund the money for supplying computers to schools
that are currently without? This presents a question of ethics. Is it fair to test some children
online while others in lower socioeconomic areas are not equipped with the same resources
(Zhao, 2012)? This idea leads us to the most alarming drawback of administering tests online.
EFFECTIVENESS OF COMPUTER BASED TESTING 8
The largest flaw in computer-based tests is that a technology gap exists. Students who
are not exposed to computers regularly are presented with a gross disadvantage than students
who have access to technology all of the time. Low income and minority families are not as
likely to have computers and high speed internet readily available at home (Backes & Cowan,
2018). Urban schools are also less likely to provide digital resources. Students must be exposed
to the technology for practice in order to be successful and feel comfortable using the tools to
take an assessment. Rather than being assessed on content knowledge, students are penalized for
Even when students are adequately exposed to technology, it is possible that not all
educators have been appropriately trained to use digital tools in the classrooms. Teachers may
not have the time or expertise to determine which software or web-based application is
appropriate and compatible (Martindale, 2015). Student success can be hindered if the teacher
has not provided sufficient practice with technology in preparation for the CBT. This factor can
Negative views concerning online assessments are typically associated with test design and
comfortability levels with testing platforms. A major factor contributing to student perception is
the opportunity to review testing materials. CBTs that do not allow students to return to previous
questions are not deemed favorably by examinees (Trisnawati, 2015). Walker (2016) identifies
the rationale behind examinees who are critical of online tests. The elements of CBTs that are
EFFECTIVENESS OF COMPUTER BASED TESTING 9
considered disadvantageous to the students involved in the study include reading from a screen,
sequencing and ordering of questions, noise management of keyboards, lack of familiarity with
program, errors from typing rapidly, and not being able to see the exam as a whole (Walker,
2016).
After reviewing the literature, some research indicates that there is no significant
difference in student achievement when comparing computer-based tests (CBT) and paper-pencil
based tests (PBT). Stating that there is no significant difference does not indicate that all scores
were the exact same no matter what mode was used for testing. Instead, no significant difference
recognizes that there could have been slight differences, but those differences are slight or
insignificant. When determining that there is no significant difference between the modes,
researchers are stating that the performance from one mode could be used as a replacement for
the performance of another mode (Piaw, 2012). Other researchers summarize it by stating that
the CBT could be used as an alternative form of the traditional paper-pencil test (Oz & Ozturan,
2018).
Oz and Ozturan (2018) cite several studies from different grades and subject areas with
the same results: there is no significant difference between student scores when comparing CBT
and PBT. The subject areas that were cited are reading comprehension, English proficiency, and
math, all with the same result of no statistical difference in achievement when comparing CBT
and PBT. Within the same articles they cite other research which shows the contrary.
Ultimately their conclusion is that little research has been studied to show the reliability and
EFFECTIVENESS OF COMPUTER BASED TESTING 10
validity factors when testing students on computers and with paper-pencil (Oz & Ozturan, 2018).
The researchers sought to determine if there was reliability and validity in the scores obtained
from CBT and PBT. Their findings were that both results were reliable and had strong internal
In similar fashion, Kim and Huynh (2008) reported findings that indicate scores from
CBT and PBT are comparable. Their own study showed that the PBT scores were higher than
those from CBT, but the “magnitude of the difference was small” (Kim & Huynh, 2008, p. 567).
Hardcastle, Herrmann-Abell; & DeBoer (2017) also found some results showing differences in
early grades, but as test modes were researched in high school, no significant difference was
found between CBT and PBT. In fact, their data shows that the gap between the differences in
student performances shrinks as students move from elementary to middle school, and then
finally to high school (Hardcastle, Herrmann-Abell; & DeBoer, 2017). An earlier study by
Bennett (2003) yields similar results: the differences between modes is present, but deemed
trivial and minimal. Way, Lin and Kong (2008) cite multiple studies which also conclude that
there is “minimal mode difference.” In their own study, the researchers noted that multiple
choice tests displayed comparable results between modes, while reading and English language
arts tests where “more complicated to interpret” (Way, Lin, & Kong, 2008). Overall, their
conclusion when comparing mode effects is stated as “small or nonexistent” (Way, Lin, & Kong,
2008).
Conclusions
such as paper-pencil tasks has proven to be a more time efficient way to determine student
learning. There is a clear benefit to the use of technology in providing quick assessment
When deciding if the use of computer-based assessments is also beneficial for students, it
is important to consider all sides of the issue. When considering student perceptions,
computer-based assessments have proven effective in reducing test taking anxiety thus making
students more successful (Baig, 2018). When arguing for the use of computer-based assessments
it is also important to consider accessibility. Articles reviewed suggest students with disabilities
or that require special accommodations, such as members of the Deaf or Hard of Hearing
(Schifter, 2014).
Research that is not in support of this trend suggests inexperience with technology may
cause discrepancies in test scores (Tempel, 2012). There is also research that suggests certain
subject areas are less conducive to computer-based testing and considering this when planning
course materials can impact student success rates. (Backes and Cowan, 2018).
student performance when comparing test scores obtained through computer-based assessments
versus paper-pencil assessments. Thus arguing the more efficient computer-based assessments
Student performance has increased dramatically with the use of computer based testing.
In Florida, 4th and 5th grade students results were significantly higher than those who did not
take the computer based FCAT test. Middle school students stayed the same which shows that it
does not have a negative impact (Baig, 2018). During the Pennsylvania System of School
Assessment (PSSA), all 45 students performed at a below basic level during baseline and a
proficient level on the posttest, that was taken online (Greene, 2014). Students are less stressed,
which causes low anxiety levels and students are able to perform successfully on these online
assessments. There has been increased meaningful dialogue among students who use computer
based games to prepare for online assessments. The online learning environment has a positive
effect on learning outcomes (Martindale, 2005). States and schools should take this approach of
online testing because it will ease the anxiety levels of those who experience high levels, which
Computer-based tests can, at times, be harmful to test-takers. There are two main factors
that lead to a restrictive CBT. The first is if the assessment is not properly designed. The
second, is if the assessment is administered to students who are not computer literate due to lack
of opportunities to use technology. If either of these influences are present on a CBT the
assessment could be detrimental to student performance and confidence. Student scores will be
impacted negatively, engagement will drop, and student perception will indicate a rejection for
digital assessments.
EFFECTIVENESS OF COMPUTER BASED TESTING 13
The results for this argument imply that it is crucial that the design of computer based
tests acknowledges the diverse needs of its users. During the construction of assessments,
educators must consider the background knowledge of students and more importantly their prior
experiences with technology. Students will not be successful with computer based tests if they
are not regularly exposed to programs similar to that of the testing platform. Both instructors
and students need to be trained on how to navigate the chosen testing program.
If CBTs become the only form of evaluating student progress than school systems must
make it a priority to supply all students with access to technology that connects to high-speed
internet. This is the only way to insure that online assessments will be fair and equitable for all
users, even those that come from low income families. Unless every student is equipped with the
same technological resources, computer based tests will lack validity and reliability.
While there is evidence of the positive and negative effects of computer-based testing,
many are associated with perception and data collection methods. When comparing the actual
scores in many assessments, data appears to suggest that there is no significant difference in test
scores for students who took the same CBT as compared to the PBT. The lack of significant
difference appears to be most prevalent in the older grades, but also appears to be connected to
the subject matter and form of the test. Most studies and the literature reviews present within the
studies indicate some discrepancies with results from the research. The discrepancies
encourages the researchers to prescribe more research founded in reliable and valid data to better
This literature review aimed to determine what, if any, impact the use of computer-based
testing had on student performance as well as to determine how students perceive this rising
trend. The first three review questions addressed the ways in which computer-based testing
impacted and in what ways might these assessments limit students’ performance. In this
literature review, “student performance” was defined as test scores. A review of the literature
demonstrates that test scores are on the rise in certain states and parts of the country for those
who take online assessments (Martindale, 2005). However, in instances where test scores
decline, it appears poor test creation is to blame. The literature also showed that test scores
varied depending on the different age range of the test taskers. One study suggests that younger
students may struggle to achieve on computer based assessments, but as they get older and more
familiar, scores may increase (Baig, 2018). However, some literature reviewed in this paper
indicate there is little difference in student performance when taking computer-based tests in
The final review question aimed to gather information regarding student perceptions of
computer-based assessments. There is overwhelming research that demonstrates the benefits for
educators in replacing traditional paper-based assessments with technology driven options but
less on what students feel about this rising trend. Literature reviewed in this paper suggests
students are more relaxed when taking online assessments as they report lower anxiety and stress
EFFECTIVENESS OF COMPUTER BASED TESTING 15
levels (Martindale, 2005). One could argue that if students are less stressed, they will be more
engaged in the assessment and perhaps show success. Students who are showing high levels of
anxiety may score much lower. Some literature demonstrates students with little to no
technology experience or school-based technology experience feel less successful when taking
tests online (Schifter, 2014). However, students with more experience have taken the changes in
stride.
performance as well as exploring student perceptions on the topic has provided valuable
information which can be used by educators moving forward. Generally, by providing students
well as ensuring tests are well-designed, student performance can be positively impacted, or
remain relatively the same, when using these tests as a replacement for more traditional
measures. Educators should understand that comfortability with technology will play a role in
how students perceive online assessments and the more opportunities to use the computer in the
classroom in this manner the better. As we move towards developing 21st-century learners, the
use of technology in the classroom could have immediate as well as many long-term benefits for
References
Backes, B., & Cowan, J. (2018). Is the Pen Mightier Than the Keyboard? The Effect of Online
https://files.eric.ed.gov/fulltext/ED583636.pdf
Baig, W. A., Al-Zahrani, E. M., Al-Jubran, K. M., Chaudhry, T., & Qadri, A. A. (2018).
Evaluation of Test Anxiety Levels among Preparatory Year Students of PSMCHS During
https://www.ijmrhs.com/medical-research/evaluation-of-test-anxiety-levels-among-prepa
ratory-year-students-of-psmchs-during-computerbased-versus-paperandpen-exam.pdf.
Bennett, R. E. (2003). Online assessment and the comparability of score meaning ( ETS
https://www.ets.org/Media/Research/pdf/RM-03-05-Bennett.pdf
Cawthon, S. (2015). From the Margins to the Spotlight: Diverse Deaf and Hard of Hearing
http://www.jstor.org.proxy-ln.researchport.umd.edu/stable/26235229
Ghaderi, M., Mogholi, M., & Soori, A. (2014). Comparing Between Computer based Tests and
Greene, A. (2015). Passing Standardized Assessments With Fading Prompts. Journal of Special
Hardcastle, J., Herrmann-Abell, C. F., & DeBoer, G. E. (2017). Comparing Student Performance
https://files.eric.ed.gov/fulltext/ED574099.pdf
Karay, Y., Schauber, S. K., Stosch, C., & Schuttpelz-Brauns, K. (2015). Computer Versus
Kim, D., & Huynh, H. (2008). Computer-Based and Paper-and-Pencil Administration Mode
Martindale, T., Pearson, C., Curda, L. K., & Pilcher, J. (2005). Effects of an Online Instructional
https://files.eric.ed.gov/fulltext/EJ690977.pdf
Oz, H., & Ozturan, T. (2018). Computer-based and paper-based testing: Does the test
administration mode influence the reliability and validity of achievement tests? Journal
https://files.eric.ed.gov/fulltext/EJ1175793.pdf
Are we doing wrong? Social and Behavioral Sciences, (64), 655-664. Retrieved from
https://www.sciencedirect.com/science/article/pii/S1877042812050549
Schifter, C. C., & Carey, M. (2014). Addressing Standardized Testing through a Novel
EFFECTIVENESS OF COMPUTER BASED TESTING 18
http://rethinkingschools.aidcvt.com/archive/26_03/26_03_tempel.shtml
http://dx.doi.org/10.22373/ej.v2i2.345
Walker, R., Handley, Z. (2016). Designing for Learner Engagement with Computer-based
Way, W. D., Lin, C. H., & Kong, J. (2008). Maintaining score equivalences as tests transition
online: issues, approaches and trends. Paper presented at the National Council on
http://images.pearsonassessments.com/images/tmrs/Maintaining_Score_Equivalence_as_
Tests_Transition_Online.pdf
Zhao, Emmeline. (2012). Online Standardized Tests are New, but are they Revolutionary?
https://www.huffingtonpost.com/2012/10/10/online-standardized-tests_n_1954569.html