Expert Systems With Applications: John Atkinson, Daniel Campos
Expert Systems With Applications: John Atkinson, Daniel Campos
Expert Systems With Applications: John Atkinson, Daniel Campos
a r t i c l e i n f o a b s t r a c t
Keywords: Current emotion recognition computational techniques have been successful on associating the emotional
Emotion recognition changes with the EEG signals, and so they can be identified and classified from EEG signals if appropriate
Brain–Computer Interfaces
stimuli are applied. However, automatic recognition is usually restricted to a small number of emotions
EEG
classes mainly due to signal’s features and noise, EEG constraints and subject-dependent issues. In order
Feature selection
Emotion classification to address these issues, in this paper a novel feature-based emotion recognition model is proposed for EEG-
based Brain–Computer Interfaces. Unlike other approaches, our method explores a wider set of emotion types
and incorporates additional features which are relevant for signal pre-processing and recognition classifica-
tion tasks, based on a dimensional model of emotions: Valence and Arousal. It aims to improve the accuracy of
the emotion classification task by combining mutual information based feature selection methods and ker-
nel classifiers. Experiments using our approach for emotion classification which combines efficient feature
selection methods and efficient kernel-based classifiers on standard EEG datasets show the promise of the
approach when compared with state-of-the-art computational methods.
© 2015 Elsevier Ltd. All rights reserved.
http://dx.doi.org/10.1016/j.eswa.2015.10.049
0957-4174/© 2015 Elsevier Ltd. All rights reserved.
36 J. Atkinson, D. Campos / Expert Systems With Applications 47 (2016) 35–41
parietal cortex. A popular physiological signal that is highly adopted signal based on its oscillatory processes) Brunner, an C. Vi-
for human emotion assessment is the EEG, etc. Unlike other physio- daurre, and Neuper (2011), Hjorth parameters (i.e., EEG sig-
logical signals, EEG is a non-invasive technique with good temporal nals described by activity, mobility and complexity) Zhang et al.
and acceptable spatial resolution. Thus, EEG might play a major role (2008), wavelet transform (i.e., decomposition of the EEG sig-
on detecting an emotion directly from the brain at higher spatial and nal) Petrantonakis and Hadjileontiadis (2010), fractal dimen-
temporal resolution (Yisi, Sourina, & Minh, 2010). sion (i.e., complexity of the fundamental patterns hidden in a
A major problem with recognizing emotions is that people have signal) Zhang et al. (2008).
different subjective emotional experiences as responses to the same (3) Feature selection: one little used technique of feature se-
stimuli (Wright, 2010; Yisi et al., 2010). Accordingly, emotions can be lection for emotions recognition combines a metaheuristic
classified into two taxonomy models: method known as Genetic Algorithms (GA) and a Support
(1) Discrete model: it is based on evolutionary features (Calvo & Vector Machines (SVM). This GA-SVM approach heuristically
D’Mello, 2010) that include basic emotions (happiness, sadness, searches for the best sets of features initially represented as
fear, disgust, anger, surprise), and mixed emotions such as Mo- chromosomes of features which evolves as the GA goes on,
tivational (thirst, hanger, pain, mood), Self-awareness (shame, so that these can then be provided as an input to an SVM
disgrace, guilt), etc. classifier (Wang et al., 2011). A major drawback with this
(2) Dimensional model: it is expressed in terms of two emotions method is the time spent to converge toward good results
provoking people: Valence (disgust, pleasure) and Arousal (calm, and the redundancy of the selected features assessed in each
excitement) Yisi et al. (2010). iteration of the GA.
In order to deal with this issue, other EEG feature selec-
Emotion recognition enables systems to get non-verbal informa- tion technique known as minimum-Redundancy-Maximum-
tion from human subjects so as to put events in context based on Relevance (mRMR) selects the features that correlate the
underlying captured emotions. Humans are capable of recognizing strongest with a classification variable, reducing information
emotions either from speech (voice tone and discourse) with an ac- redundancy. This method selects features that are mutually
curacy around 60% or from facial expressions and body movements different from each other while still having a high correlation
with an accuracy of 78–90%. However, the recognition task is strongly make up the selection task of mRMR (Polat & Cataltepe, 2012),
dependent on the context and requires facial expressions to be delib- by reducing redundancy between bad and good features using
erately performed or even in a very exaggerated manner, which is far Mutual Information (MI) methods, so that a subset of features
away from the natural way a user interact with intelligent interfaces. that represents best the dataset can be obtained.
Other kinds of techniques use audio signals, obtaining classifica- (4) Emotions classification: once the FVs are extracted from the
tion accuracy close to 60–90% (Calvo & D’Mello, 2010), whereas some previous task, emotions must be classified according to pre-
other methods use non-linguistic vocalizations (i.e., laughs, tears, viously identified classes of emotions. Despite the large num-
screams, etc.) to recognize complex emotional states such as anxi- ber of features used by these methods, no feature selection is
ety, sexual interest, boredom. Bi-modal methods also combine audio usually carried out. There are plenty of state-of-the-art classi-
inputs and facial expressions based on the assumption that a human fiers for automatic emotion identification. For example, Near-
emotion can trigger multiple behavior and physiological responses est Neighbor classifiers used features such as FFT and Wavelets
whenever he/she experiences this emotion. to recognize 4 types of emotions (i.e., joy, sad, angry, relaxed)
Nevertheless, most of these methods require humans to express achieving accuracies ranging from 54% to 67%. On the other
their emotional (mind) states in a deliberated and exaggerated man- hand, statistical methods such as Quadratic Discriminant Analy-
ner, so that emotions cannot spontaneously be expressed. On the sis (QDA) used several statistical features for negative and pos-
other hand, extracting information from facial expressions requires itive arousal levels with an average accuracy of 63% (Koelstra
monitoring a subject by using one of several cameras, whereas for et al., 2012; Petrantonakis & Hadjileontiadis, 2010; Wu et al.,
audio-based approaches, emotions are very hard to recognize when- 2010; Yisi et al., 2010).
ever a subject does not speak or produce any sounds (Giakoumis,
Tzovaras, Moustakas, & Hassapis, 2011; Sourina et al., 2011).
A popular and effective non-invasive technique to measure 3. An adaptive BCI-based emotions recognition model
changes on brain activity is called (EEG), which transforms brain ac-
In this work, a novel approach that combines minimum-
tivity into images of variations of electrical potential by using small
Redundancy-Maximum-Relevance (mRMR) based feature selection
low-cost devices (AlMejrad, 2010). There are several approaches for
tasks and kernel classifiers for emotions recognition is proposed. The
EEG-based emotion recognition which are usually based on four main
method takes EEG signals received from BCI devices and incorporates
tasks (Calvo & D’Mello, 2010):
relevant features in order to detect several kinds of emotional states
(1) Signal preprocessing: an EEG device can directly get signals by using state-of-the-art classifiers. The main contribution of this re-
from the brain. However, there are some noise sources that search is that unlike other automatic emotion recognition methods
are not neurologically produced known as artifacts (i.e., blink- our approach
ing, muscular effects, vascular effects, etc.), so digital signal
processing techniques must be applied to represent signals (1) Incorporates a feature selection task into the classification
using frequencies and harmonic functions (Petrantonakis & task.
Hadjileontiadis, 2010; Yisi et al., 2010). (2) Uses multi-label classifiers to simultaneously recognize a
(2) Feature extraction: EEG signals are highly dimensional so com- wider range of emotion types based on a dimensional model.
putational processing becomes very complex. Hence different The overall model is composed of three tasks: signal preprocess-
features must be extracted in order to simplify the further ing, feature extraction and selection, and emotions classification (see
emotion classification task so to create input Feature Vectors Fig. 1).
(FV). Typical methods include statistical metrics of the signal’s
first difference (i.e., median, standard deviation, kurtosis sym- 3.1. EEG signal preprocessing
metry, etc.), spectral density (i.e., EEG signals with specific
frequency bands) Zhang, Yang, and Huang (2008), Logarith- In order to train the emotions classifier, a set of previously
mic Band Power (Log BP) (i.e., power of a band within the emotion-labeled EEG data extracted from subjects self-assessing
J. Atkinson, D. Campos / Expert Systems With Applications 47 (2016) 35–41 37
as:
1
max VI , VI = I (C, fi ) (2)
|S| fi ∈S
Table 1
Parameters setting for GA-SVM-based feature selection.
Table 2 the selected features. Note also mRMR generates features sets that are
Best results for setting parameters for different feature selection methods.
smaller than for GA-SVM, which makes it more suitable for real-time
Method Dimension Accuracy (%) Std. dev (%) No. of applications as it requires less work to extract features and achieve
features good classification accuracy.
mRMR Arousal 60.72 9.08 35
4.3. Overall evaluation
mRMR Valence 62.39 9.90 20
GA-SVM Arousal 56.69 9.34 95
GA-SVM Valence 53.46 9.05 94 A final overall experiment compared our approach against some
state-of-the-art methods. To this end, the best previously tunned con-
figurations were used: mRMR-based feature selection, 35 features for
Finally, Table 2 shows the best setting results for each dimension, dimension Arousal and 20 features for dimension Valence, and RBF
feature selection method, and the number of selected features. The kernel with γ = 0.05 for the SVM classifier. The model was then
SVM classifier using RBF Kernel with γ = 0.05, produces the highest trained using 40 stimuli tests for each of the 31 subjects of the dataset.
accuracy, and the performance of mRMR is better than GA-SVM for Experimental results are shown in Fig. 6, indicating a median of
60.7% and 62.33%, for dimension Arousal and Valence, respectively
(i.e., std. dev. of 9 is close to the median of both dimensions).
Graphic of Fig. 7 shows the classification accuracy for each dimen-
sion and subject. In addition, the lower row for each figure shows the
number of subjects for whom certain features were selected, where
darker points represent a larger number of subjects. This suggests
there is no relationship between the selected features for one or other
subject. Nevertheless, best selected features for both dimensions, cor-
respond to the statistical measures extracted from each channel (left-
hand side).
Classification accuracy of our model was also compared against
other approaches, indicating very promising results when dealing
Fig. 6. Classification accuracy per dimension for each subject. with combination of methods and different classes of emotions as
Polat, D., & Cataltepe, Z. (2012). Feature selection and classification on brain computer Wu, D., Courtney, C., Lance, B., Narayanan, S., Dawson, M., Kelvin, S., & Parsons, T.
interface (BCI) data. In Proceedings of signal processing and communications applica- (2010). Optimal arousal identification and classification for affective computing us-
tions conference (SIU) (pp. 1–4). IEEE. ing physiological signals: virtual reality stroop task. IEEE Transactions on Affective
Sourina, O., Wang, Q., Liu, Y., & Nguyen, M. (2011). A real-time fractal-based brain state Computing, 1(2), 109–118.
recognition from eeg and its applications. In F. Babiloni, A. L. N. Fred, J. Filipe, & Yisi, L., Sourina, O., & Minh, N. (2010). Real-time EEG-based human emotion recog-
H. Gamboa (Eds.), Biosignals (pp. 82–90). SciTePress. nition and visualization. In Proceedings of international conference on cyberworlds
Wang, L., Xu, G., Wang, J., Yang, S., Guo, L., & Yan, W. (2011). GA-SVM based feature (CW) (pp. 262–269).
selection and parameters optimization for BCI research. In Proceedings of seventh Zhang, A., Yang, B., & Huang, L. (2008). Feature extraction of eeg signals using power
international conference on natural computation (ICNC): vol. 1 (pp. 580–583). spectral entropy. In Proceedings of international conference on biomedical engineer-
Wright, F. (2010). Emotional instant messaging with the epoc headset, (Master’s thesis). ing and informatics, 2008. BMEI: vol. 2 (pp. 435–439). IEEE.
Baltimore County: University of Maryland.