Building A Face Expression Recognizer and A Face Expression Database For An Intelligent Tutoring System
Building A Face Expression Recognizer and A Face Expression Database For An Intelligent Tutoring System
Abstract—this work presents the building and validating of a a sporadic and in a short time. All therefore the above
face expression database and a face expression recognizer. The indicates the real importance of understanding those
face expression recognizer uses a geometric-based technique emotions [4]. It is significantly important to understand and
that measures distances between the central point on the face
and other 68 facial landmark points. These measures are
detect emotion in the area of ITS.
transformed into features to train a support vector machine.
II. RELATED WORKS
The database was built inside an educational context while
students program in Java code. The tests validate the accuracy A. Face expression databases
of the recognizer applying a ten-fold cross-validation.
A face expression database is a dataset of images
Keywords—Intelligent-tutoring-system; artificial-intelligence; containing people shows expressions which represent human
face-expression-recognition; face-expression-database emotions. The most important databases are explained below.
Cohn-Kanade (CK) [5] is a dataset of 486 sequences from
I. INTRODUCTION 97 positioned persons. The database seeks to represent the
Nowadays, no one can deny the relevance and the image sequence inside the Facial Action Coding System
importance of artificial intelligence. We can find the use of (FACS). In this way, each expression receive an emotion
artificial intelligence methods in subjects such as social label. The CK includes basic emotions as well as Action
networks, smart devices, games, industrial management Units (AU) annotations. Cohn-Kanade Plus (CK+) [6] is an
control processes, and education. About the educational extension of original CK (version 2). This version includes
issue, the artificial intelligence has been used to individualize posed expressions, non-posed (spontaneous) expressions
the students learning process, e.g. identifying learning styles recording from 84 novel subjects and more metadata
or recognizing personal features on students [1]. information.
Intelligent tutoring systems (ITS) are the answer of Radboud Faces Database (RaFD) [7] contains an image set
researchers to integrate the computational tutors with of 49 models in two subsets: 39 Caucasian Dutch adults and
artificial intelligence to get an improved personalization 10 Caucasian Dutch children. All model expressed eight
using artificial intelligence [2]. The ITSs are systems facial expressions (anger, happiness, fear, sadness, contempt,
designed to handle the learning processes of students. An ITS surprise, disgust, and neutral) with three gaze directions and
can help students to know which topics have not been studied five camera angles. They obtained a total of 120 images per
yet. Also, an ITS can recognize affective states of students to model which had to comply various requisites like wearing a
design and perform the best pedagogical strategies for type of shirt or no hair on the face.
students.
B. Face Expression Recognition
ITSs usually use the emotional Ekman theory, which
describes how human faces move different face parts to Expression recognition deals about how computers get the
express an emotion, either moving a mouth corner or raising ability to understand and process human expressions. Next,
an eyebrow. These emotions are known as basic emotions. an analysis of the main techniques to recognize expressions
is presented below.
Basic emotions have as characteristic being spontaneous and
Geometric-based techniques [8] are techniques that
expressive [3]. However, there are other emotions which
represent faces as a facial points representation. These
emerge when students are working in their learning tasks.
representation is describe by a concatenation of X and Y
These emotions are known as learning-centered emotions. coordinates of fiducial points. The points are used as a feature
Learning-centered emotions relate to educational context and input in a machine-learning and predicting an expression
they emerge during deep learning activities such as thought, system, to represent a face in these techniques there are two
reflection or problem-solving. Some research found a relation types of representations of models, free model, and model-
to emotions as frustration, boredom or engagement during based.
complex cognitive processes. Those emotions emerge during Local Binary Pattern (LBP) [9] is a technique wherein the
a prolonged time unlike basic emotions which appear during original purpose was to identify textures on different surfaces
392
V. EXPERIMENTS, RESULTS, AND DISCUSSIONS [3] P. Ekman, "An argument for basic emotions,"
The validation test of our database using our recognizer Cognition & emotion, vol. 6, pp. 169-200, 1992.
consists in training the classifier using 90% of the database [4] S. D’Mello and A. Graesser, "Dynamics of affective
and 10% remaining for testing. The test is a ten-fold states during complex learning," Learning and
validation cross. The recognizer processed a total of 655 Instruction, vol. 22, pp. 145-157, 2012.
photographs in training time. Table 1 shows the distribution [5] T. Kanade, J. F. Cohn, and Y. Tian,
of classes, the accuracy obtained for this work, and the "Comprehensive database for facial expression
accuracy for another work explained below. analysis," in Automatic Face and Gesture
Recognition, 2000. Proceedings. Fourth IEEE
TABLE 1. DISTRIBUTION OF CLASSES International Conference on, 2000, pp. 46-53.
Emotion Number of Accuracy of this Accuracy of [6] P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z.
classes work work [19] Ambadar, and I. Matthews, "The extended cohn-
Boredom 17 64% 76% kanade dataset (ck+): A complete dataset for action
Engagement 430 64% 91%
Excitement 114 83% 68%
unit and emotion-specified expression," in 2010
Frustration 94 62% 55% IEEE Computer Society Conference on Computer
Vision and Pattern Recognition-Workshops, 2010,
The number of classes presented in Table 1 are the classes pp. 94-101.
for this work and not match with classes used in the work. [7] O. Langner, R. Dotsch, G. Bijlstra, D. H.
The recognizer got a good result in engagement and boredom Wigboldus, S. T. Hawk, and A. van Knippenberg,
emotions. It is evident that excitement and frustration did not "Presentation and validation of the Radboud Faces
get such high scores. To find out if our results were low we Database," Cognition and Emotion, vol. 24, pp.
needed to compare it against other similar works. We decided 1377-1388, 2010.
to compare the work against a single research wherein [8] K.-C. Huang, S.-Y. Huang, and Y.-H. Kuo,
authors used the same emotions of this work [14]. The only "Emotion recognition based on a novel triangular
difference found in the comparison work is the excitement is facial feature extraction method," in Neural
named as delighted, but its description is same. After a Networks (IJCNN), The 2010 International Joint
review, we concluded that results were satisfactory until now, Conference on, 2010, pp. 1-6.
two emotions obtained a result higher than the compared [9] S. Happy and A. Routray, "Automatic facial
research. expression recognition using features of salient
facial patches," IEEE transactions on Affective
VI. CONCLUSIONS Computing, vol. 6, pp. 1-12, 2015.
Detecting emotions is a complicated task even for human [10] Y.-I. Tian, J. F. Cohn, and T. Kanade, "Evaluation
beings, using the face only to detect emotions is a challenge of Gabor-wavelet-based facial action unit
which can become insufficient. It is a research in continuous recognition in image sequences of increasing
development. This work presents the building of a face complexity," presented at the Automatic Face and
expression database focused on learning-centered emotions, Gesture Recognition, 2002. Fifth IEEE International
the database has relation to educational context. Also, the Conference on, 2002.
work presents the building of a geometric-based recognizer [11] (2016). Emotiv EPOC. Available:
that calculates coordinates, distances, and angles in different http://emotiv.com/
faces. The tests validate that this work has successful results [12] D. E. King, "Dlib-ml: A machine learning toolkit,"
in two emotions considering that this work is a beginning to Journal of Machine Learning Research, vol. 10, pp.
recognize centered-learning emotions. This paper can be used 1755-1758, 2009.
as a method to build a face expression database with an
[13] F. Pedregosa, G. Varoquaux, A. Gramfort, V.
educational approach. The database can be found through the
Michel, B. Thirion, O. Grisel, et al., "Scikit-learn:
next URL: ftp://javasensei.ddns.net/corpus.rar
Machine learning in Python," Journal of Machine
REFERENCES Learning Research, vol. 12, pp. 2825-2830, 2011.
[14] N. Bosch, S. D'Mello, R. S. Baker, J. Ocumpaugh,
[1] B. M. McLaren, O. Scheuer, and J. Mikšátko,
V. Shute, M. Ventura, et al., "Automatic Detection
"Supporting collaborative learning and e-
of Learning-Centered Affective States in the Wild,"
discussions using artificial intelligence techniques,"
in IUI '15 Proceedings of the 20th International
International Journal of Artificial Intelligence in
Conference on Intelligent User Interfaces, 2015, pp.
Education, vol. 20, pp. 1-46, 2010.
379-388.
[2] B. P. Woolf, Building Intelligent Interactive Tutors.
Amherst, Massachusetts, Estados Unidos: Morgan
Kaufmann Publishers, 2009.
393