A Conceptual Framework For Detecting Cheating in O
A Conceptual Framework For Detecting Cheating in O
EMPIRICAL RESEARCH
Denise V. Siegfeldt
Florida Institute of Technology, Nathan M. Bisk College of Business, Department of Extended
Studies, Hampton Roads Site, Fort Eustis, VA 23604, e-mail: [email protected]
ABSTRACT
Selecting the right methodology to use for detecting cheating in online exams requires
considerable time and effort due to a wide variety of scholarly publications on academic
dishonesty in online education. This article offers a cheating detection framework that
can serve as a guideline for conducting cheating studies. The necessary theories and
related statistical models are arranged into three phases/sections within the framework
to allow cheating studies to be completed in a sufficiently quick and precise manner.
This cheating detection framework includes commonly used models in each phase and
addresses the collection and analysis of the needed data. The model’s level of com-
plexity ascends progressively from a graphical representation of data and descriptive
statistical models to more advanced inferential statistics, correlation analysis, regres-
sion analysis, and the optional comparison method and the Goldfeld-Quandt Test for
heteroscedasticity. An instructor receiving positive results on the possibility of cheating
in Phases 1 or 2 can avoid using more advanced models in Phase 3. Tests conducted
on sample courses showed that models in Phases 1 and 2 of the proposed framework
provided results effectively for over 70% of the test groups, saving users further time and
effort. High-tech systems and low-cost recommendations that can mitigate cheating are
discussed. This framework will be beneficial in guiding instructors who are converting
from the traditionally proctored in-class exam to a take-home or online exam without
authentication or proctoring. In addition, it can serve as a powerful deterrent that will
alleviate the concerns that an institution’s stakeholders might have about the reliability
of their programs.
Subject Areas: Cheating Detection Framework, Cheating Mitigation, De-
scriptive and Inferential Statistical Modeling, Goldfeld-Quandt Test, Online
Exam Cheating, Take-Home Exam Cheating.
† Corresponding Author.
This is an open access article under the terms of the Creative Commons Attribution-NonCommercial
License, which permits use, distribution and reproduction in any medium, provided the original work is
properly cited and is not used for commercial purposes.
1
2 Conceptual Framework for Detecting Cheating in Exams
methods to study the nature and incidence of cheating (Fask, Englander, & Wang,
2015; Harmon & Lambrinos, 2008; Hollister & Berenson, 2009).
Cheating in online exams is usually detected by using self-reporting surveys,
empirical methods, and a combination of quantitative and qualitative methods.
However, results from earlier studies are mixed. Some studies have reported cheat-
ing (Fask et al., 2015; Harmon & Lambrinos, 2008; King et al., 2009) while others
registered no cheating (Greenberg et al., 2009; Hollister & Berenson, 2009; Peng,
2007; Werhner, 2010). Shen, Chung, Challis, and Cheung (2007) found no sig-
nificant performance difference between the traditional classes and online classes,
although the traditional class performed slightly better than the online class.
Academic dishonesty has been studied under different exam-course settings,
including: (1) unproctored online (OL) exams in online (OL) courses (Hollister
& Berenson, 2009; King et al., 2009); (2) proctored in-class (IC) exams in OL
courses (Hollister & Berenson, 2009; Werhner, 2010); (3) proctored IC exams in
traditional IC courses (Shon, 2006); (4) unproctored take-home (TH) exams in IC
courses (Andrada & Linden, 1993; Marsh, 1984); and (5) OL exams or quizzes in
IC courses (Fask et al., 2015; Peng, 2007). This study consisted of OL exams in
IC courses and TH exams in IC courses.
can conclude that cheating may have taken place. Due to relatively small sample
sizes, sometimes less than 30, these researchers have used the student t-test for the
hypothesis test.
More advanced correlational and regression analysis models have also been
applied to detect cheating. An approach presented by Harmon and Lambrinos
(2008) utilized correlation and regression analysis as well as variance analysis to
detect cheating in an online macroeconomics exam. A regression model was used
to predict their summer 2004 and summer 2005 final exam scores from student-
related explanatory variables. According to Harmon and Lambrinos (2008, p. 119),
“the courses, although offered a year apart, were almost identical in structure and
content.” Among several variables that were considered by Harmon and Lambrinos
(2008), GPA was the only significant variable that reflected a student’s ability.
Moreover, they reasoned that the independent “human capital variable such as
GPA” has a high relationship (coefficient of determination, R2 ) to the dependent
variable (exam scores) due to the student’s high performance in the proctored in-
class exam, rather than to cheating (Harmon & Lambrinos, 2008, p. 121). Cheating
would result in a low value for R2 . Similarly, Fask et al. (2015, p. 2) introduce
student GPA and class attendance as “mastery (independent) variables” that could
reasonably increase the performance of the student and correspondingly detect if
cheating has occurred.
The OPM 450 and OPM 350 courses offered in academic years 2014-2015
and 2015-2016 were a part of this study. Missed class days (excused or unexcused)
and percentage of assignments submitted up to the exam date are considered as in-
dependent variables that are responsible for the students’ high in-class exam scores
(dependent variable). Due to privacy concerns, student GPA was not included as
an independent variable, although it is highly related to student performance and
was considered by other researchers (e.g., Fask et al., 2015; Harmon & Lambrinos,
2008).
The GQT proposed by Goldfeld and Quandt (1965) for testing homoscedas-
ticity was adapted to test for heteroskedasticity (unequal variance). It is used by
researchers to detect online cheating by examining the variance (σ 2 ) of the residual
errors (ԑi ) from the regression models (Fask et al., 2015; Harmon & Lambrinos,
2008). The limitations of the GQT are discussed in more detail in the Conclusions
and Limitations section. Furthermore, comparison of the predicted mean values
with the observed values obtained from the regression model is an additional
method to detect cheating (Harmon & Lambrinos, 2008).
These statistical models can indicate the possibility of cheating or no cheating
in an objective exam. It is difficult to identify and prove that specific students have
cheated based on statistical models (Marx & Longer, 1986). Alerted instructors
can increase proctoring and can keep a closer observation of the students, change
the exam schedule, modify the exam format, and use other interventions provided
in the Interventions for Mitigating Cheating section of this article.
Proposed Study
A large number of researchers drawn to this field have generated a wide variety of
scholarly efforts on the academic dishonesty occurring in online exams. Therefore,
D’Souza and Siegfeldt 5
an instructor who is planning a cheating study will spend considerable time and
effort searching for the most relevant methodology. There is an urgent need for
a preplanned and well-organized procedure for instructors and researchers to use
in detecting cheating in online or on other unproctored exams (e.g., take-home
exams).
This article provides a conceptual framework for planning and organizing a
sustained cheating detection and mitigating program. The framework emphasizes
the standard means of cheating detection using related concepts from the auditing
profession. Cheating in academics is similar to financial fraud that is detected
during the audit of a company’s financial statements by trained auditors. The “fraud
triangle” discussed by Ramos (2003, p. 28) is the conceptual framework used as a
guide in detecting financial fraud that satisfies one or more of the three conditions
(incentive/pressure, opportunity, and rationalization/attitude) when fraud occurs.
Subsequently, Becker, Connolly, Lentz, and Morrison (2011) matched the three
sides of the “fraud triangle” (Incentive to cheat, an opportunity to cheat, and
rationalization to cheat) with cheating in academics, producing the business fraud
triangle that they used as a framework to uncover academic dishonesty among
business students. The tested business fraud triangle (Becker et al., 2011) has been
adopted as the conceptual framework for the cheating detection framework (CDF).
The conceptualization of the framework occurred during the investigations
into alleged cheating complaints about online exams at a mid-size university in
the Commonwealth of Virginia and included an extensive literature review. The
necessary assumptions, theories, and related graphical and statistical models are
arranged in three phases/sections within the main body of the framework, allowing
cheating studies to be completed in a sufficiently quick and precise manner. The
framework offers instructors the choice of commonly used graphical charts and
statistical models to select in each phase and addresses the collection and analysis of
necessary data. Tests conducted during this study on sample business management
courses showed that the proposed CDF was able to detect cheating in Phases 1 and
2 for over 70% of the test groups, thus saving users further time and effort.
100
90
80
70
Exam 1 Scores
60
50
40
30
EXAM I IC
20
Linear (EXAM I IC)
10
0
0 1 2 3
Missed Class Days
to extend their office hours for advising students and to pursue their scholarly
activities.
The rationale for the conceptual framework offered in the present study
was the sudden rise in exam scores for the OPM 450 course when these exams
were transformed from proctored in-class exams to online exams given no proctors.
According to Siegmann, Moore, and Aquino (2014), higher scores in online exams
could be due to more time being available for instructor-student interactions,
a less stress-filled exam environment, flexible timing, and students doing more
reading for unknown answers. These findings were supported by results from the
anonymous self-reporting survey addressed in the Introduction and Background
section, based on students who completed the OPM 450 and OPM 350 online
exams. Among those students that participated in the survey, 66% expressed a
preference for online exams because of the flexible timing to complete their exams
(77%), and their ability to take the exam when fully prepared (95%).
In this study, the instructor suspected that the higher online exam scores for
OPM 450 were due to cheating because students with poor attendance records were
getting abnormally high exam scores. This outcome contradicted results of Dobkin
et al. (2010), which reported that attendance at class meetings has a positive impact
on exam performance. The attendance plots graphed in Figures 1 and 2 for OPM
450 proctored in-class exams and unproctored online exams, respectively, visi-
bly demonstrate the impact of attendance on the scores for both exams. Figure 1
shows a general pattern of proctored in-class exam scores that realistically de-
creases almost identically with the standard attendance line (negative slope) when
class days missed increases, hence supporting the results of Dobkin et al. (2010).
Figure 2 shows the pattern of online exam scores in comparison to a standard
attendance line. The exam scores increase (positive slope) unrealistically as the
D’Souza and Siegfeldt 7
100
90
80
70
Exam 1 Scores
60
%EXAM 1
50
Linear (%EXAM 1)
40
30
20
10
0
0 1 2 3 4
MIssed Class Days
Study Design
The development and testing of the cheating detection framework (CDF) were
conducted in two parts. In Part 1, each phase of the CDF was assigned appropriate
graphical representation and statistical models in ascending order of complexity. In
Part 2, the functionality of the framework was tested on groups of sample business
courses. A group consisted of an in-class (IC) or take-home (TH) exam given in
the fall 2014 or spring 2015 semester of an academic year and an online (OL) exam
given during the same semester of the following academic year (i.e., fall 2015 and
spring 2016).
A total of seven groups were used in Parts 1 and 2 of this study. Group 1
was used in Part 1 to illustrate the composition and working of the framework, and
Groups 2-7 were similarly organized and were used to test the framework in Part 2.
A detailed composition of all seven groupings is provided in later sections (Table 3).
The equivalence requirements for assignments, quizzes, and exams suggested
by Fask et al. (2014) were strongly enforced within the time, costs, and legal re-
straints. The seven groups, each containing a set of two examinations, had a mix of
students from across campus. The OPM 450 course was a Management major re-
quirement, but it was popular in other departments across campus. Every semester,
the study population consisted of primarily senior level undergraduate students
with a management major or minor from departments including architecture, avi-
ation, journalism and mass media, psychology, and others. Similarly, OPM 350
was populated by juniors from management, aviation, and other students seeking
a major or minor in management. Random registration for these courses across
campus resulted in a class of students with approximately similar characteristics
every semester. The OPM 450 and 350 courses were taught by the same professor
using an identical syllabus, textbook and notes, PowerPoint presentations, exam
8 Conceptual Framework for Detecting Cheating in Exams
format (with different questions), and assignments in each course. The OL and
IC exams consisted of multiple-choice, true/false, and fill-in-the-blank questions,
which could be graded by Blackboard Course Management Systems. The difficulty
level of questions was maintained the same in both exams but the wording and
parameters were slightly modified to get different answers.
To maintain a consistent computer grading system, there is a need for
objective-type questions consisting of a mix of multiple-choice, true/false, and
fill-in-the-blank questions. The smaller descriptive quantitative and qualitative
questions can be included by breaking the question into subquestions whose an-
swers are sequentially entered into the course website in precise single or multiple
numerical value(s) or word(s). For the larger qualitative essay questions, it is highly
recommended that instructors use an anti-plagiarism tool to detect cheating.
NO
YES
p ≤ 0.05 Stop. Cheating is likely. Refer
Reject H0
to interventions to mitigate
cheating.
NO
10 Factor A = 19%
Frequency
(cheating zone).
8
Test of Hypothesis:
6 REJECT H0, (p =
0.002). Strong
4 chance of cheating.
0 FREQ_F15-OL
0 20 40 60 80 100 120
FREQ_F14-IC
Exam 1 Scores
the .05 value by default. The t-test for two sample means with unequal variance is
run on the data and the computed t-value is subject to the following decision rule
(α: = .05, two-tailed):
If t < tc or –t < –tc or if p < .05, the null is rejected, and the alternate
hypothesis is supported. (The tc value in the criterion corresponds to the critical
value associated with the degrees of freedom and the chosen α level.) In the Group 1
case, t = –3.37, t.05, 35 = –2.03, p = .0019. Therefore, the null hypothesis is rejected,
and we conclude that the mean scores in OL and IC exams were significantly
different. With the mean OL exam score being higher, we conclude that cheating
0.2 Factor A% =
0.18 [(|µ IC- µOL|)/µ IC] *100
0.16
0.14 Base Factor = 15%
A < 15% - No
p-values
0.12
Cheating. A >
0.1
15% - Cheating
0.08
0.06 No Cheating Cheating
P-VALUE
0.04
0.02 Linear (P-
0 VALUE)
0% 5% 10% 15% 20% 25% 30%
Factor A
D’Souza and Siegfeldt 11
took place in the OL exams. Based on the framework, it is recommended that the
study is stopped and the instructor plan strategies for mitigating cheating. Although
unnecessary, some instructors may opt to proceed to Phase 3 to reconfirm the results
from Phase 2.
The summarized results of the linear regression analysis are shown in Table 1.
The fitted linear regression equation (5) for IC exams had one significant
12 Conceptual Framework for Detecting Cheating in Exams
Table 1: Results of the multiple regression Equations (3) and (4) (Group 1).
Significant
Multiple Regression Intercept Variable (S) R R2 F Ratio F Sig N
independent variable while the linear regression for OL exams had no significant
variables.
LN Yi,IC = 4.6655 − 0.3611 ABSi,IC . (5)
The value of R2 is low (1.8%) for the OL exam scores and is high (48.1%)
for the IC exam scores. Attendance (ABS) was the only significant variable for
the IC exam scores (equation (4)), accounting for approximately 48% of the
variation, while the other independent variables were not significant. Equation (3)
had no significant variables. Therefore, none of the independent variables in the
regression model for equation (3) contributed toward the approximate 2% of the
variation. This large difference in the value of R2 between OL and IC is because
the performance associated variable ABS contributed to this increase in R2 in
the proctored IC exam, while a low value in the OL exam showed the same
variable as having a low contribution (Fask et al., 2015; Harmon & Lambrinos,
2008). Hence, the possibility of cheating in the OL exams was significantly higher
than cheating in the proctored IC exams, which were under strict supervision.
Optional test
The CDF contains two optional tests for cheating detection. They are not a require-
ment but can be used by instructors desiring additional confirmation of the outputs
from Phases 1, 2, and 3, or who may have skipped some of the phases in favor of
the other options.
The first is the GQT for homoscedasticity, used to test for heteroskedasticity
or the presence of unequal variance. The homoscedasticity assumption for the
regression model requires that the disturbances or errors (ԑi ) are constant and
randomly distributed, producing constant variance (σ 2 ). This concept has been
D’Souza and Siegfeldt 13
used to study disturbances in two independent samples, for example, the OL and
IC exam scores. The GQT tested the multiple regression equations (3) and (4)
for differences in variance, which is termed as the test of heteroskedasticity. The
question, however, is: Do the error terms ԑi for equation (3) and (4) have a different
variance for OL and IC exam scores?
The following steps in applying the GQT are suggested by Murray (2006):
H0 : = σ 2 OL = σ 2 IC ,
H1 = σ 2 OL = σ 2 IC .
iii. Regress the natural log-linear regression for equations (3) and (4).
iv. Compute the F ratio of the sum of squares of the residuals (SSR) for OL
and IC (Table 2).
v. Compare the F ratio with the critical value Fc at α = .05. If F ratio > Fc ,
the null hypothesis is rejected.
The computed values are F ratio = 1.640 and F17, 23 = 2.091. Since, F ratio <
F17, 23 , there is not enough statistical evidence to reject the null hypothesis, and
we can conclude that there is no difference in the variance between the OL exams
and the proctored IC exams. This result satisfies the homoscedasticity assumption
for regression models and concludes that it is unlikely cheating is taking place.
However, this result from this optional test differs from the results from Phases 1,
2, and 3.
The second optional test is a comparison test that was conducted to corrob-
orate the Phase 1, 2, and 3 results. The observed performance associated variable,
ABSi,j values for OL and IC, were input into linear regression equation (5) to com-
pute the predicted exam scores, which were then compared with the observed exam
scores for OL and IC students. The difference between the predicted OL (70.5)
and observed (86.1) scores was highly significant (p = 1.57E-07). In contrast, the
difference between the predicted IC (75.1) and observed (71.2) scores was not
significant (p = .3897). Hence, results of a comparison test confirm the presence
of cheating in the OL exams.
Although cheating could have been detected in the earlier Phases 1 and 2, all
the Phases 1, 2, and 3 along with the optional models were applied to demonstrate
the different models that instructors can use, the data needed for testing, and
interpretation of the results.
Due to space limitations, only the testing results for Group 3 (OPM 450) and Group
6 (OPM 350) are documented here, in detail.
Group 3 was subjected to testing by graphs and descriptive statistical models
in Phase 1. The frequency polygons indicate the highest point was the IC exam
conducted in spring of 2015. The computed mean scores for OL (75) and IC (71)
exams indicated a marginal difference of 6% (factor A) which, according to the
cheating indicator shown in Figure 5, falls in the “no cheating” zone. In the case
of Group 3, one concludes that there was no cheating in the OL exams. However,
the attendance plots for the OL exams shows a positive slope which can be due
to cheating. Therefore the instructor is advised to move to Phase 2 to confirm that
no cheating is taking place. The hypothesis test did not reject the null hypothesis
(p = .290). There is no significant difference in exam scores which leads one to
conclude that there is no strong evidence to prove that cheating is taking place. No
cheating was detected in Phase 2, so further testing in Phase 3 may be discontinued.
The CDF prepares a test report card for the instructor of Group 3 providing details
of tests conducted in Phases 1 and 2. A sample of this test report card is shown in
Figure 6 for Group 6, which is discussed in the following paragraphs.
In Group 6, the CDF’s results were inconsistent for the phases. In Phase 1,
the highest peak of the frequency polygon was for the OL exam (Figure 6). The
attendance plots for both exam scores had a positive slope. The attendance plot
for TH exam that is displayed in Figure 6 had a greater positive slope than the
OL exam. The mean exam scores of 78 in the OL exams and 85 in the TH exam
were close with factor A = 9%. For this value of Factor A, the cheating indicator
(Figure 5) classified Group 3 under the “no cheating” zone. In Phase 2, the test of
the hypothesis was modified to include the TH exam in place of the IC exam, as
follows:
H0 : μOL = μTH ,
H1 : μOL = μTH .
The computed p = .0502 was marginally greater than .0500. An instructor
will be unclear about these borderline outcomes in Phases 1 and 2 and needs to
proceed to Phase 3. In Phase 3, the multiple regressions equation (4) for IC exam
were replaced by the equation (6) for the TH exam as follows:
LN Yi,TH = β 0,TH + β 1,TH GENDERi,TH
+ β 2,TH ABSi,TH + β 3,TH ASSIGNi,ITH . (6)
Equations (3) and (6) are applied to the OL and TH exams in Group 6. The
multiple regression models’ R2 value for the TH exam scores of approximately 31%
is higher than the R2 value for the OL exam scores of approximately 11%, but there
were no significant variables in either equation. Since none of the performance-
associated independent variables in the regression model for equations (3) and (6)
contributed toward the 31% and 11% of R2 , there were other non-explanatory
factors that influenced the dependent variable, one of which was cheating. With
TH having a higher R2 than OL, it can be concluded that cheating took place in
OL and TH exams, but cheating is higher in TH exams. Finally, the GQT provides
D’Souza and Siegfeldt 15
Table 3: (Continued)
Exam Cheating
Gr Course Sem/Exam No. Type Test Results (Y/N/U)
an F ratio = 1.03 and F7, 11 = 3.01. The null hypothesis was not rejected, meaning
there was no difference in the variations and hence, no cheating. The provision of
inconsistent results indicates that the CDF has a limitation in such narrow or close
cases and the instructor should make the final decision after a review of all the test
results documented in the test report card (Figure 6)
The CDF could provide sufficient test results to assist an instructor in de-
termining if cheating possibly took place or not for over 85% of the test groups.
Cheating was detected in 67% of these test groups, while no cheating was detected
in 33% of these test groups. Tests conducted on the groups showed that models
in Phases 1 and 2 of the proposed framework provided results effectively for over
70% of the test groups, saving users further time and effort.
Mean TH = 85. 80
7
Frequency
6 Mean OL = 78.
60
5 Factor A = 9%
4 40
3 FREQ S'15 TH
2 FREQ S'16 OL 20
1
0 0
0 50 100 150 0 0.5 1 1.5 2 2.5
Exam Scores Missed Class Days
Phase 2: Test of Hyppothesis: Decision Rule (α: = 0.05, ttwo-tailed). If t < tc or -tt < - tc the nuull is rejected
and the alteernate hold trrue. t = 2.01
19, tc = 2.09
93, p = 0.05002 Borderlinne Case
P
Phase 3: Optional
O Go
oldberg-Qua or Heteroskkedasticity foor Equation
andt Test Fo ns 3 and 6
MMultiple Semesster/ Measure
M of dfi SSR SSR/df F Ratio Fc
RRegression Exam
m Variance
V α = 0.05
EEquation 3 F7, 11 = 3.01
((OL) Sprin
ng’16/1 Residual
R 11 0.152 0..0139 = 0.0143/ Do Not
0.0139 Reject H0
EEquation 6
Spring
g’ ‘15/1 Residual
R 7 0.100 0.0143 = 1.03
((TH)
P
Phase 3: Optional
O com
mparison teest For pred
dicted and ob
bserved Exaam 1 scoress.
Comparisson Test Noot Required
GQT Tesst: Yes/No. Results: _GQT test of heteroskedast
h ticity – equaal variances______________
Prepared
d By: ______
___________ ______. Ap pproved By:: _____________________. Date: ________
Action Taken:
T _____
___________ __________ ____________________________________________
18 Conceptual Framework for Detecting Cheating in Exams
The framework experienced problems during the test for Group 6 due to
very narrow differences in the mean exam scores. In such borderline cases, the
CDF does provide decisive results, thereby shifting the decision making back to
the instructor. Even with all of the limitations discussed earlier, the CDF is still a
powerful deterrent that can mitigate the concerns that an institution’s stakeholders
might have about the reliability of their programs.
REFERENCES
Andrada, G. N., & Linden, K. W. (1993) Effects of two testing conditions on
classroom achievement: Traditional in-class versus experimental take-home
conditions. Report ED 360 329, Purdue University. Presented at the Annual
Meeting of the American Educational Research Association, Atlanta, GA,
April 2–16.
Barron, J., & Crooks, S. M. (2005) Academic integrity in web-based distance
education. Tech Trends Linking Research and Practice to Improve Learning,
49(2), 40–45.
Becker, D., Connolly, J., Lentz, P., & Morrison, J (2011) Using the business fraud
triangle to predict academic dishonesty among business students. Academy
of Educational Leadership Journal, 10(1), 37–54.
Bellezza, F. S. & Bellezza, S. F (1989) Detection of cheating on multiple-choice
tests by using error, similarity, analysis. Teaching of Psychology, 16(3), 151–
155.
Berkey, D., & Halfond, J. (2015) Cheating, student authentication and proctoring
in online programs. New England Board of Higher Education, July 20,
2015 http://www.nebhe.org/thejournal/cheating-student-authentication-and-
proctoring-in-online-programs/
Bowers, W. J. (1964) Student dishonesty and its control in college. New York
Bureau of Applied Social Research. New York, NY: Columbia University.
Cluskey Jr., G. R., Ehlen, C. R., & Raiborn, M. H. (2011) Thwarting online exam
cheating without proctor supervision. Journal of Academic and Business
Ethics, 4, 1–7.
College of Liberal Arts and Science (CLAS). (2016) Definition of aca-
demic dishonesty. University of Colorado Denver, Denver, 1–2 (www.
ucdenver.edu/academics/colleges/CLAS/faculty-).
deLeeuw, E. D., Hox, J. J., & Dillman, D. A. (2008) The cornerstones of survey
research. International handbook of survey methodology. New York: Psy-
chology Press, Taylor & Francis Group. E. D. deLeeuw, J. J. Hox, & D. A.
Dillman (Eds.) 1–17.
Dobkin, C., Gil, R., & Marion, J. (2010) Skipping class in college and exam per-
formance: evidence from a regression discontinuity classroom experiment.
Economics of Education Review, 29, 566–575.
D’Souza and Siegfeldt 21
Fask, A., Englander, E., & Wang, W. (2014) Online testing and its implications in
business education. Proceedings of the Northeast Business Economic Associ-
ation 41st Annual Meeting, Monmouth University, NJ 07764 (www.nbea.us).
Fask, A., Englander, E., & Wang, Z (2015) On the integrity of online testing for
introductory statistics courses: A latent variable approach. Practical Assess-
ment, Research, and Evaluation, 20(10), 1–12.
Faurer, J. C. (2013) Grade validity of online quantitative courses. Contemporary
Issues in Education Research, 6(1), 93–96.
Goldfeld, S. M., & Quandt, R. E. (1965) Some tests for homoscedasticity. Journal
of the American Statistical Association, 60(310), 539–547.
Greenberg, K., Lester, J. N., Evans, K., Williams, M., Hacker, C., & Halic,
O. (2009) Student learning with performance–based, in-class and learner-
centered, online exams. International Journal of Teaching and Learning in
Higher Education, 20(3), 383–393.
Harmon, O. R., & Lambrinos, J. (2008) Are online exams an invitation to cheat?
Journal of Economic Education, 39(2), 116–125.
Hemming, A. (2010) Online tests and exams: Lower standards or improved learn-
ing? The Law Teacher, 44 (3), 283–308.
Holland, P. W. (1996) Assessing unusual agreement between the incorrect answers
of two examinees using the K-Index: statistical theory and empirical support.
Program Statistics Research Technical Report No. 96-4, Educational Testing
Services (ETS), Princeton, NJ, 08541.
Hollister, K. K., & Berenson, M. L. (2009) Proctored versus un-proctored online
exams: Studying the impact of exam environment on student performance.
Decision Sciences Journal of Innovative Education, 7(1), 271–294.
Kennedy, K., Nowak, S., Raghuraman, R., Thomas, J., & Davis, S. F. (2000) Aca-
demic dishonesty and distance learning: Student and faculty views. College
Student Journal, 34(2), 309–314.
King, C. G., Guyette, R. W. Jr., & Piotrowski, C. (2009) Online exams and
cheating: An empirical analysis of business students’ views. The Journal of
Educators Online, 6(1), 1–11.
Lanier, M. M. (2006) Academic integrity and distance learning. Journal of Crim-
inal Justice Education, 17(2), 244–261.
Marsh, R. (1984) A comparison of take-home versus in-class exams. The Journal
of Educational Research, 78(2), 111–113.
Marx, D. B. & Longer, D. E. (1986) Cheating on multiple choice exams is difficult
to assess quantitatively. North American Colleges and Teachers of Agricul-
ture Journal, March 1986, 23–26.
McCabe, D. L., Treviño, L. K., & Butterfield, K. D. (2001) Cheating in aca-
demic institutions: A decade of research. Ethics and Behavior, 11(3), 219–
232.
Murray, M. P. (2006) Econometrics: A modern introduction. Boston, MA 02116:
Pearson Education, Inc.
22 Conceptual Framework for Detecting Cheating in Exams
Olt, M. R. (2002) Ethics and distance education: Strategies for minimizing aca-
demic dishonesty in online assessment. Online Journal of Distance Learning
Administration, V(III), 1–9.
Peng, Z. (2007) Giving online quizzes in corporate finance and investments for a
better use of seat time. The Journal of Educators Online, 4(2), 1–18.
Pett, M. A. (2016) Nonparametric statistics for health care research: Statistics
for small samples and unusual distributions (2nd ed.). Los Angeles, CA:
Sage.
Ramos, M. ((2003) Auditors’ responsibility for fraud detection. Journal of Ac-
countancy, 195(1), 28–36.
Rogers, C. F. (2006) Faculty perceptions about e-cheating during online testing.
Journal of Computing Sciences in Colleges, 22(2), 206–212.
Shen, Q., Chung, J. K. H, Challis, D. I., & Cheung, R. C. T. (2007) A comparative
study of student performance in traditional mode and online mode of learning.
Computer Applications in Engineering Education, 15(1), 30–40.
Shon, P. C. H. (2006) How students cheat on in-class examinations: Creativity,
strain, and techniques of innovation. Ann Arbor, MI: MPublishing, Univer-
sity of Michigan Library.
Siegmann, S. M., Moore, D., & Aquino, C. P. (2014) Exam performance in a
hybrid course: A model for assessing online and in class delivery modes. The
Online Journal of Distance Education and e-Learning, 2(4), 70–80.
Werhner, M. J. (2010) A Comparison of the performance of online versus tra-
ditional on-campus Earth Science students on identical exams. Journal of
Geoscience Education, 58(5), 310–312.