Research On Abnormal Behavior Detection of Online Examination Based On Image Information
Research On Abnormal Behavior Detection of Online Examination Based On Image Information
Research On Abnormal Behavior Detection of Online Examination Based On Image Information
89
Figure 3. Convolutional neural network diagram of head pose estimation
feature points. The results are shown in Table 3.
IV. EXPERIMENTAL RESULTS
D. Mouth state detection
According to this paper, there is currently no public dataset
During the process of examination, the face detection of that can be used for monitoring abnormal behavioral analysis.
the input image is performed in this paper. In the detected face Therefore, for the purpose of our experiment, a dataset
area, dlib is used to locate the feature points of the face, and containing 30 independent videos is created. The average
related feature points of the mouth are selected, as shown in duration of each video was 2 minutes. Each video in the data
Fig. 4. set simulated 6-10 cheating at random time.
A. Head pose estimation
In the course of the examination, the angle change of the
examinee's head pose in the yaw angle can obviously reflect
Figure 4. Representation of mouth feature point the abnormal behavior of the examinee. The turning range of
According to the formula the yaw angle is [0°, 90°] and [0°, - 90°]. When the examinee
ܯൌ ݉݅ ݊ሺͷͳǤ ݕǡ ͷ͵Ǥ ݕሻ is in the frontal stable state, the yaw angle is almost 0°. This
ۓ
ۖ ܺ ൌ ͶͻǤ ݔെ ሺͳǤ ݔെ ͶͻǤ ݔሻ state is the normal answer state of the examinees. After a
ܻ ൌ ܯെ ሺʹǤ ݕെ ͷͳǤ ݕሻ large number of experiments and, this paper determines that
ܹ۔ൌ ͷͷǤ ݔെ ͶͻǤ ݔ ʹሺͳǤ ݔെ ͶͻǤ ݔሻ when the rotation amplitude of the yaw angle of the examinee
ۖ
ܪ ەൌ ͷͺǤ ݕെ ܯ ʹሺʹǤ ݕെ ͷͳǤ ݕሻ is greater than 15 degrees, it is abnormal behavior according
The positioning of the mouth area is completed, and the S.Prathish’s work. However, due to the jittering of video and
image of the mouth area is converted into a grayscale image. human physiological factors, we need to ignore some
This paper finally determines the binarization threshold is momentary angle changes. They cannot be judged as
0.4, getting the binarization image of the mouth according to abnormal behavior.
the distribution characteristics of the mouth area grayscale.
Consecutive area markers are used for the binarization image
to compare the area size of each area. The area with the
largest area is selected as the mouth connectivity area. The
area is represented by the number of white pixels in the area.
The sobel edge detection algorithm is used to extract the
edges of the connected components, whose perimeter of the
minimum bounding ellipse perimeter is calculated. By (a) 19.519e (b) -42.337e
calculating the circularity e of the mouth's binary region, and
through statistics and summary of experimental data, a
threshold is obtained to judge whether the mouth is open or
closed. The higher the roundness e is, the greater the degree
of mouth opening is.
E. Inference System
The above head pose information and mouth state
(c) -0.129e
information are inputted into the determination system at the
Figure 5. Examinee's Head Pose Estimation
same time. In the decision system, if the examinee's yaw When the face is guaranteed to face the computer
angle is greater than 15° in successive 20 frames or the yaw forward, the yaw angle is displayed as -0.129°, which is
angle changes frequently and widely, it is determined that almost 0°. When the examinee's head rotates left and right,
there is abnormal behavior; At the same time, if the the yaw angle changes positively and negatively. The
examinee's mouth is frequently open and closed, it is measurement results are shown in Fig. 5. Therefore, the head
determined that there is speech behavior, which is abnormal. pose estimate clearly shows the examinee's behavior. In this
If any behavior that meets the judgment of abnormal behavior paper, through a large number of experiments, the examinee's
occurs within any time range, it can be regarded as a yaw angle change threshold is set to 15°. Therefore, when the
possibility of cheating.
90
examinee’s yaw angle is greater than 15° in the successive 20 detected, and Y indicates that abnormal behavior is
frames, it is determined as an abnormal behavior. successfully detected.
100 Through comprehensive experiments, the system has a
false alarm rate of only 5% on the basis of abnormal behavior
monitoring and recognition rate of 90%. In summary, the
0
method proposed in this paper can more accurately monitor
1
79
157
235
313
391
469
547
625
703
781
859
937
1015
1093
1171
1249
the occurrence of abnormal behavior during the online
-100 examination process.
Figure 6. Examinee yaw angle changes in the video sequence
V. CONCLUSION
It is obviously observed in Fig. 6 from the 199th to 397th
frames, 463th to 660th frames, 991th to 1090th frames, and This paper designs a set of solutions for abnormal
1200th to 1280th frames that the examinee's head pose rotates behavior monitoring of online examination based on image
for a long time (more than 20 frames) and exceeds a threshold information. Through the use of head pose estimation based
of 15°. It is determined that the examinee is abnormal during on convolutional neural network and threshold-based mouth
this period. state judgment, and the combination of certain decision rules,
the monitoring of abnormal behavior such as turning heads
B. Mouth State Detection and speaking during the online examination is completed. In
the future work, we will continue to study the monitoring of
subtle abnormal behavior during the examination process,
such as deviating from the screen to achieve a higher
(a) closed state (b) binary image (c) external ellipse recognition rate and reduce the false alarm rate.
REFERENCES
[1] Nowcoder, https://www.nowcoder.com
(d) open state (e) binary image (f) external ellipse [2] ProctorU:Real People Real Proctor, http://www.proctoru.com
Figure 7. The mouth area Binary image External ellipse [3] S. Prathish and K. Bijlani, “An intelligent system for online exam
The roundness e of opening and closing mouth is monitoring,” Proc. Information Science (ICIS), International
Conference on, IEEE, 2016, pp. 138-143.
calculated according to the formula e = 4ʌA/P2. Where eę
[4] X. Zhu and D. Ramanan, “Face detection, pose estimation, and
[0,1], A and P indicate the area and perimeter respectively. landmark localization in the wild,” Proc. Computer Vision and Pattern
The experimental results show that when e < 0.5, the mouth Recognition (CVPR), 2012 IEEE Conference on, IEEE, 2012, pp.
is in the closed state; when 0.5 > e, the mouth is in the open 2879-2886.
state (such as speaking and so on). [5] X. Geng and Y. Xia, “Head pose estimation based on multivariate label
distribution,” Proc. Computer Vision and Pattern Recognition (CVPR),
The examinee’s mouth is open and closed respectively in 2014 IEEE Conference on, IEEE, 2014, pp. 1837-1842.
Fig. 7(a), (d). According to frequent changes in the opening [6] Y. LeCun, et al., “Handwritten digit recognition with a back-
and closing of the mouth within a certain number of frames, propagation network,” Proc. Advances in neural information
we can judge the examinee's speaking behavior and judge processing systems, 1990, pp. 396-404.
abnormal behavior. [7] M. Koestinger, et al., “Annotated facial landmarks in the wild: A large-
scale, real-world database for facial landmark localization,” Proc.
C. Comprehensive Experiments Computer Vision Workshops (ICCV Workshops), 2011 IEEE
International Conference on, IEEE, 2011, pp. 2144-2151.
For the same video with abnormal behavior, we conduct
[8] M.D. Zeiler, “ADADELTA: An Adaptive Learning Rate Method,”
abnormal behavior monitoring experiment through Computer Science, 2012.
professional invigilators, Swathi [4], and our method. Some [9] J. Duchi, et al., “Adaptive subgradient methods for online learning and
records are as follows: stochastic optimization,” Journal of Machine Learning Research, vol.
TABLE IV. Monitoring comparison of monitoring programs 12, no. Jul, 2011, pp. 2121-2159.
Time Video [10] D.P. Kingma and J. Ba, “Adam: A method for stochastic optimization,”
Invigilators Swathi Our method arXiv preprint arXiv:1412.6980, 2014.
0-20* [11] A. Bulat and G. Tzimiropoulos, “How far are we from solving the 2d
21-40 N N Y & 3d face alignment problem?(and a dataset of 230,000 3d facial
41-60* landmarks),” Proc. International Conference on Computer Vision,
61-80 N N N 2017, pp. 8.
81-100 N N Y [12] V. Kazemi and S. Josephine, “One millisecond face alignment with an
101-120 N N Y ensemble of regression trees,” Proc. 27th IEEE Conference on
121-140* Y Computer Vision and Pattern Recognition, CVPR 2014, Columbus,
The marked * in Table 4 indicates no abnormal behavior United States, 23 June 2014 through 28 June 2014, IEEE Computer
Society, 2014, pp. 1867-1874.
during this period, and N indicates no abnormal behavior is
91