A Review On Machine Learning For EEG Signal Processing in Bioengineering
A Review On Machine Learning For EEG Signal Processing in Bioengineering
14, 2021
(Methodological Review)
Abstract—Electroencephalography (EEG) has been a EEG signals are taken on the surface of the scalp, but there also
staple method for identifying certain health conditions in exists iEEG signals, which are taken inside the brain. In this
patients since its discovery. Due to the many different types paper, we will be focusing primarily on conventional scalp EEG
of classifiers available to use, the analysis methods are
also equally numerous. In this review, we will be examin- signals.
ing specifically machine learning methods that have been Conventionally, EEG recordings may be obtained by con-
developed for EEG analysis with bioengineering applica- necting electrodes to the scalp with the use of a conductive
tions. We reviewed literature from 1988 to 2018 to capture gel. A differential amplifier is then used to amplify each active
previous and current classification methods for EEG in
electrode compared to the reference before it is sent through an
multiple applications. From this information, we are able to
determine the overall effectiveness of each machine learn- anti-aliasing filter. Finally, this filtered signal is converted with
ing method as well as the key characteristics. We have an analog-to-digital converter.
found that all the primary methods used in machine learn- Clinically, EEG signals are used primarily to diagnose and
ing have been applied in some form in EEG classification. treat various brain disorders such as epilepsy, tremor, concus-
This ranges from Naive-Bayes to Decision Tree/Random
sions, strokes, and sleep disorders. More recent applications of
Forest, to Support Vector Machine (SVM). Supervised learn-
ing methods are on average of higher accuracy than their EEG include using machine learning as a method of analysis. In
unsupervised counterparts. This includes SVM and KNN. particular, there is much research on epileptic seizure detection
While each of the methods individually is limited in their and sleep disorder research in combination with machine learn-
accuracy in their respective applications, there is hope that ing. Additionally, there is also a growing interest in studying
the combination of methods when implemented properly
EEG signals for gaming to control and manipulate objects using
has a higher overall classification accuracy. This paper
provides a comprehensive overview of Machine Learning brainwaves due to EEG monitoring for brain activity during
applications used in EEG analysis. It also gives an overview tasks [36].
of each of the methods and general applications that each EEG waveforms vary based on the band, which denotes the
is best suited to. frequency range. The delta band is the slowest wave with the
Index Terms—Machine learning, eeg, survey, medical highest amplitude, having a frequency range below 4 Hz. For
applications, signal processing, signal analysis. adults, it is located frontally, while for children it is located
I. INTRODUCTION posteriorly. The theta band is between 4 to 7 Hz and is most com-
mon in young children while signifying drowsiness or arousal
LECTROENCEPHALOGRAPHY (EEG) is a method of
E testing electrical signals in the brain. It is often applied as
a technique for data analysis such as time and frequency series
in adults. This band tends to spike due to an active inhibition of
a movement or response. The alpha band is between 8 to 14 Hz,
and it is correlated to eye muscle movements. It is located on
analysis. The brain’s neurons contain ionic current, which cre- both sides of the head’s posterior regions. The beta band is above
ates voltage fluctuations that EEG can measure. This electrical 14 Hz and is correlated with general motor behavior. It is located
activity is spontaneous and recorded over a period of time from on both sides of the head’s frontal regions [44].
many scalp electrodes to form an EEG signal. [22] Traditionally, Some of the advantages of using EEG compared to other
Manuscript received April 10, 2019; revised August 9, 2019; accepted
techniques to study brain function are low costs, tolerance to
September 29, 2019. Date of publication January 28, 2020; date of motion from subjects, and no radiation exposure risks. Some of
current version January 22, 2021. (Corresponding author: Mohammad- the disadvantages of using EEG include low spatial resolution
Parsa Hosseini.)
Mohammad-Parsa Hosseini is with the Bioengineering Department,
and poor signal-to-noise ratio.
Santa Clara University, Santa Clara, CA 95053 USA and also with the
AI Research, Silicon Valley, CA USA (e-mail: [email protected]).
Amin Hosseini is with the Electrical and Computer Engineering De- II. MACHINE LEARNING METHODS FOR EEG
partment, Azad University, Central Tehran Branch, Tehran, Iran (e-mail:
[email protected]). A. Overview
Kiarash Ahi is with the University of Connecticut, Storrs, CT 06269
USA (e-mail: [email protected]). Machine learning is the use of a set of mathematical mod-
Digital Object Identifier 10.1109/RBME.2020.2969915 els and algorithms to gradually improve the performance of a
1937-3333 © 2020 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 205
Fig. 1. Machine learning applications on EEG have been developed based on supervised and unsupervised learning in the literature. Supervised
learning is categorized to classification and regression which produce discrete and continuous accordingly. Unsupervised learning is categorized
to clustering and dimensionality reduction which produce discrete and continuous accordingly.
Fig. 2. The overall steps for EEG analysis by machine learning include preprocessing, feature extraction, feature selection, model training, model
testing.
singular task. It takes training data sets as input to use as a guide data, unedited. Pre-processing involves the removal of noise and
for making estimates without being specifically programmed other outliers in the data set. Feature extraction determines the
to. The tasks vary widely in this space and can be categorized spectrum of the data point groupings and what features they
into two main groups: supervised and unsupervised learning. correspond to. Feature selection is the isolation of the desired
Unsupervised learning is the case when the algorithm builds a classifiers that the machine learning method will be testing for
pattern of recognition from a data set containing only inputs the following training. Machine learning training involves the
with no set outputs. Supervised learning has a subsection being use of training data sets, whether with or without known outputs
semi-supervised learning. They are identical in the sense that to refine the classification method. Lastly, the testing phase is
they both learn from data sets with given inputs and known the processing of true test data sets and comparing the overall
outputs with the exception that semi-supervised has parts of accuracy of the desired feature.
the data set missing. Supervised learning is primarily used in
applications of classification and regression while unsupervised
B. Regression
learning lends itself to feature learning and the inverse, dimen-
sionality reduction. This paper will discuss some of the most Regression modeling is a popular tool in statistics because
popular machine learning methods and categorize them based it is a simple way to create a functional relationship between
on the type of learning with some practical applications in EEG. variables. Various types of regression include: univariate and
EEG signals can be used as indicators of harder to detect multivariate for quantitative response variables; simple and mul-
medical conditions with the assistance of machine learning tiple for predictor variables; linear for linearly transformable
methods. In Fig. 1 the applications of machine learning on data; nonlinear for nonlinearly transformable data; analysis of
EEG signals are shown based on supervised and unsupervised variance for qualitative variable predictors; analysis of covari-
learning. Supervised learning develops a predictive model using ance for the combination of qualitative and quantitative variable
both input and desired output data is categorized to classification predictors; and logistic for qualitative response variable [84].
and regression which produce discrete and continuous accord- Legendre and Gauss first applied regression using the Method
ingly. Unsupervised learning develops a predictive model using of Least Squares. This method makes approximations by sum-
just input data is categorized to clustering and dimensionality ming the squares of each equation residual to best fit the data,
reduction which produce discrete and continuous accordingly. and it is applied in Linear Regression. as shown in the equation
Fig. 2 describes the general flow of how machine learning is below.
implemented to get the desired classification of the data sets.
The first step is signal acquisition. This is essentially the raw yi = B0 + B1 xi + ei,i=1,...,n (1)
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
206 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
TABLE I
REGRESSION MODELS APPLIED FOR EEG ANALYSIS
Linear Regression is one of the most common regression tech- Linear SVM classifier with hard margin
niques. In this model, the parameters are specified in the form
of a linear combination, while each independent variable is not
l
1
l
l
necessarily linear. Multiple linear regression is similar, except W (α) = − αi + 2 yi yj αi αi x i x j (3)
i=1 i=1 j=1
that there are several independent variables rather than just one.
When the parameters are not linear, nonlinear regression must
be used. This also uses a sum of squares technique, though it Kernel trick equation minimizing W subject to:
uses an iterative procedure to minimize the function.
l
yi αi = 0
C. SVM i=1
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 207
Fig. 3. Higher dimension kernel separation. The kernel trick involves the transformation of the existing algorithm from a lower dimensional data
set to a higher one.
TABLE II
SUPPORT VECTOR MACHINE APPLICATIONS WITH EEG
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
208 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
TABLE III
ARTIFICIAL NEURAL NETWORKS APPLICATION FOR EEG ANALYSIS
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 209
TABLE IV
NAIVE BAYES APPLICATIONS WITH EEG
TABLE V
REVIEW ON DECISION TREE AND RANDOM FOREST
in Fig. 6 which shows the backpropagation of information By applying the tan hyperbolic function the dot product of
from one layer to another and to a specifically intended associated weights from previous states and the dot product of
node. associated weights and input state, we shall have the value of
To understand the working of RNN it is important to define the new state. We can have the final output function as:
the transitions from one previous state to a new state. Let Xt be
the input vector, Ht be the new state, and Ht-1 be the previous yt = Why .ht (11)
state. RNN is observed to be a function of the input vector and
the previous state, which will land us to the new state Ht. We F. Naive Bayes
can represent a simple Vanilla version of the RNN by obtaining
the weight function Fw and implementing that to find the output Naive Bayes classifier is a popular text categorization method
function Yt. This can be represented as follows: that applies Bayes’ theorem to separate data based on simple
trained features. Essentially, the model assigns labels as feature
ht = fw (ht−1 , xt ) (9) vectors within a finite set. While simple in nature, with adequate
pre-processing it can match more advanced methods such as
ht = tanh(Whh .ht−1 + Wxh .xt ) (10) SVM discussed above. The one disadvantage of the naive Bayes
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
210 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
TABLE VI
A REVIEW ON ENSEMBLE LEARNING STATE OF ARTS
TABLE VII
FUZZY LOGIC FOR EEG ANALYSIS
[45] method is that it considers all of the feature vectors as vectors and it assigns probabilities to a given outcome or case.
independent from one another regardless of any real correlation. Event models can be separated into 2 main classes, Gaussian
The main advantage of it is that it only needs a small number of Naive Bayes and Multinomial Naive Bayes. In a data set with
training data sets to begin correctly estimating the parameters continuous values, a good assumption would be that it follows
necessary for classification. Several models can be implemented a Gaussian distribution. Using this method the Bayes method
for the Bayes method. The most common of which is the prob- assigns probabilities based on the curve. A multinomial event
abilistic model. In this model, the features are represented by model represents the frequencies of specific events spawned
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 211
TABLE VIII
LINEAR DISCRIMINANT ANALYSIS
TABLE IX
K MEANS FOR EEG ANALYSIS
Fig. 6. Recurrent Neural Network where connections between nodes Fig. 7. Example for decision tree technique to determine a health
form a directed graph along a temporal sequence. It makes previous condition.
outputs to be used as inputs.
from multinomials, often as a histogram. A potential concern The probabilistic Naive Bayes Model
is when a feature does not occur in the data set at all. This
causes the multiple of all the estimates to be zero. It can be (v−μk )2
−
2σ 2
corrected with a pseudocount to smooth out any outliers in the P (x = v | Ck ) = √ 1 2
e k (13)
2πσk
data set [91].
P (x)|(c)
P (c|x) = P (x) (12) The Gaussian Naive Bayes Model
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
212 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
H. Ensemble Learning
Ensemble learning is a supervised learning algorithm. As the
name suggests, ensemble learning ensemble’s many different
algorithms to make a model that gives a better predictive perfor-
mance. The general idea is to improve the overall performance
by combining decisions received from different multiple models.
It is based on the concept of diversity, more diverse models
are considered for obtaining the results for the same problem
in comparison to single models. This gives a set of hypotheses
which can be combined to gain better performance. All the single
models are called as base learners when combined are called as
an ensemble. The ensemble is mostly better than the base learn-
Fig. 8. Random Forest is an ensemble learning method which is used
mostly for classification and regression. It operates by creating a multi- ers from which the ensemble is made. Ensemble learning can be
tude of decision trees on various sub-samples of the dataset and uses used in the fields of medicine, fraud detection, banking, malware
majority voting or averaging for finding output. This model improves the and intrusion detection, face and emotion recognition, etc.
accuracy of prediction and can control over-fitting.
I. Fuzzy Logic
Almost every household machine or equipment (like the air
G. Decision Tree and Random Forest
conditioner, washing machine, etc.) operates on the concept of
Decision trees use questions about the features of an item to Fuzzy Logic. This logic is fed to a control system usually called
classify data. Each question can be represented as a node, in the Fuzzy system control, where each component is designed to
which there is a child node for each answer to that question. function and alter another physical operating system, to achieve
This creates a hierarchy, in other words, a tree. The most basic the desired functionality. To understand how a fuzzy system
tree would be a binary one in which each question results in a yes works, it is necessary to analyze the system requirements and
or no answer. Therefore there is a yes and no child node for each the intent for using a fuzzy system [20]. To make a system
parent node question. Data is sorted through the tree by starting a knowledge-based functioning element with the capacity to
at the top-most node, also known as the root, and maneuvering apply the human cognitive processes, such as reasoning and
its way down to the leaf, or the node that has no children. The thinking, has to have a stable component that can provide output
path taken is dependent on the data’s features. Once the data on the perspective of the degree of truth for a given set of
reaches the leaf, it can be classified under the class associated input variables. Fig. 9 shows the breakdown of a typical fuzzy
with that particular leaf [64]. system. For a fuzzy system to work effectively, the following
The advantages of decision trees are that they are simplistic components need to be assured of performance:
and can be easily combined with other techniques for decision 1) Fuzzy sets: A fuzzy set is considered to be correspon-
making. The disadvantages of decision trees are that they are dent with the member function, which is defined in a
somewhat unstable as well as inaccurate, especially with varying fuzzy space where the variables are set. The feature of
level sizes which cause biases towards larger levels. a member function is to provide a degree of membership
In the study of machine learning, and different classifying to any element within the well defined fuzzy sets. Then
and distribution methods, we come across the Random Forest the member function assigns these elements a numerical
technique, which can be used for both data classification and value between 0 to 1, where 0 implies the corresponding
regression operations. As the name suggests, Random Forest element is not an element in the fuzzy set or 1 means the
operates by producing a multitude of decision trees and trained corresponding element is an element of the fuzzy set.
by performing bagging operation to combine multiple decision 2) Fuzzy Rules: The way a fuzzy logic is intended to function
trees or models to arrive at a more stable and accurate data is defined by a set of applied fuzzy rules, which deter-
prediction. Random Forest creates additional randomness to the mines the output which will be specified by the IF-THEN
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 213
Fig. 9. Example for Fuzzy System. For a Fuzzy system to work effectively, the following features and components needs to be assured of
performance: 1. Fuzzy sets, 2. Fuzzy Rules, 3. Fuzzy Logic Inference, 4. Fuzzy Score.
Fig. 10. General K-means classification. K-means works based on using an algorithm to locate a partition in order to minimize the error between
a cluster’s empirical mean and points within. Using these K clusters, K-means tries to minimize the summation of the squared errors.
rules. The IF-THEN rules are observed to create a condi- linguistic terms. In the second stage, the system processes
tional statement that will consist of fuzzy logic. For exam- the rules according to the strengths of each input variable.
ple, the IF-THEN assumes where X and Y are intended in the third stage, the resulting fuzzy values are converted
terms and are evaluated by the terms of fuzzy sets with the back to numerical values, by the process of Defuzzifica-
range being U and V. This divides the statement into two tion. This process thereby maps the fuzzy domain output
parts namely antecedent and consequent. If the antecedent back to the crisp domain, which makes the output clear.
is a preceding statement which specifies the terms X and 4) Fuzzy Score: The output from the FIS system is in the
U, then the consequent statement should conclude with form of a fuzzy score, for all the individual input scores
Y and V. These combined makes a rule which states: if X that are known to be generated by the system. The FIS
is U, then Y is V. However, these rules are based on the system calculates the fuzzy score by taking into consider-
natural language and model representation, based on the ations all the defined fuzzy constraints and membership
given fuzzy sets and logic. functions. The score is dependent on the type of rules ap-
3) Fuzzy Logic Inference or Fuzzy Inference System (FIS): plied and the type of input variables. Every input variable
Once the set of fuzzy rules and membership functions is assigned a score by the FIS based on the fuzzy rules
have been defined, the FIS is implemented for process criteria.
simulation, and control, and is done by the type of data or As the main application of Machine Learning is found to be in
knowledge provided. The FIS system usually operates on pattern recognition of EEG signals, Fuzzy Logic can be used to
3 stages: In the first stage, the numerical input variables determine the correct recognition rate of EEG classifications
which are provided to the system, are mapped for a degree at different stages. However, a combination of Fuzzy logic
of compatibility for the respective fuzzy sets. This is with Neural networks often called the Neuro-Fuzzy system, is
called the Fuzzification process. This process allows the adopted, where the system can apply the fuzzy parameter (like
system to express the input and output in fuzzy-readable fuzzy sets, fuzzy rules) and combine that with the neural network
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
214 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 215
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
216 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
HOSSEINI et al.: REVIEW ON MACHINE LEARNING FOR EEG SIGNAL PROCESSING IN BIOENGINEERING 217
[41] M.-P. Hosseini, M.-R. Nazem-Zadeh, D. Pompili, K. Jafari-Khouzani, [65] J. S. Kirar and R. Agrawal, “Relevant feature selection from a combina-
K. Elisevich, and H. Soltanian-Zadeh, “Automatic and manual segmen- tion of spectral-temporal and spatial features for classification of motor
tation of hippocampus in epileptic patients MRI,” arXiv:1610.07557, imagery EEG,” J. Med. Syst., vol. 42, no. 5, p. 78, 2018.
2016. [66] J. Laton et al., “Single-subject classification of schizophrenia patients
[42] M.-P. Hosseini, M. R. Nazem-Zadeh, D. Pompili, and H. Soltanian- based on a combination of oddball and mismatch evoked potential
Zadeh, “Statistical validation of automatic methods for hippocampus paradigms,” J. Neurological Sci., vol. 347, no. 1/2, pp. 262–267,
segmentation in MR images of epileptic patients,” in Proc. IEEE 36th 2014.
Annu. Int. Conf. Eng. Medicine Biol. Soc., 2014, pp. 4707–4710. [67] J. Le Douget, A. Fouad, M. M. Filali, J. Pyrzowski, and M. Le Van
[43] M.-P. Hosseini, D. Pompili, K. Elisevich, and H. Soltanian-Zadeh, “Op- Quyen, “Surface and intracranial EEG spike detection based on discrete
timized deep learning for EEG big data and seizure prediction BCI via wavelet decomposition and random forest classification,” in Proc. IEEE
Internet of Things,” IEEE Trans. Big Data, vol. 3, no. 4, pp. 392–404, 39th Annu. Int. Conf. IEEE Eng. Medicine Biol. Soc., 2017, pp. 475–478.
Dec. 2017. [68] Y.-H. Lee et al., “A cross-sectional evaluation of meditation experience
[44] M.-P. Hosseini, D. Pompili, K. Elisevich, and H. Soltanian-Zadeh, “Ran- on electroencephalography data by artificial neural network and support
dom ensemble learning for EEG classification,” Artif. Intell. Medicine, vector machine classifiers,” Medicine, vol. 96, no. 16, 2017.
vol. 84, pp. 146–158, 2018. [69] P. Li, C. Karmakar, J. Yearwood, S. Venkatesh, M. Palaniswami, and
[45] M. P. Hosseini, H. Soltanian-Zadeh, and S. Akhlaghpoor, “Computer- C. Liu, “Detection of epileptic seizure based on entropy analysis of short-
aided diagnosis system for the evaluation of chronic obstructive pul- term EEG,” PLOS ONE, vol. 13, no. 3, 2018, Art. no. e0193691.
monary disease on CT images,” Tehran University Med. J., vol. 68, [70] X. Li et al., “An ocular artefacts correction method for discriminative
no. 12, 2011. EEG analysis based on logistic regression,” in Proc. IEEE 23rd Eur.
[46] M. P. Hosseini, H. Soltanian-Zadeh, and S. Akhlaghpoor, “Three cuts Signal Process. Conf., 2015, pp. 2731–2735.
method for identification of COPD,” Acta Medica Iranica, vol. 51, no. 11, [71] Y.-H. Liu, S. Huang, and Y.-D. Huang, “Motor imagery EEG classifi-
pp. 771–778, 2013. cation for patients with amyotrophic lateral sclerosis using fractal di-
[47] M.-P. Hosseini, H. Soltanian-Zadeh, K. Elisevich, and D. Pompili, mension and fisher’s criterion-based channel selection,” Sensors, vol. 17,
“Cloud-based deep learning of big EEG data for epileptic seizure pre- no. 7, 2017, Art. no. 1557.
diction,” in Proc. IEEE Global Conf. Signal Inf. Process. (GlobalSIP), [72] M. Manjusha and R. Harikumar, “Performance analysis of KNN classifier
2016, pp. 1151–1155. and k-means clustering for robust classification of epilepsy from EEG
[48] M.-P. Hosseini, T. X. Tran, D. Pompili, K. Elisevich, and H. Soltanian- signals,” in Proc. IEEE Int. Conf. Wireless Commun., Signal Process.
Zadeh, “Deep learning with edge computing for localization of epilepto- Netw., 2016, pp. 2412–2416.
genicity using multimodal RS-FMRI and EEG big data,” in Proc. IEEE [73] T. Meyer, J. Peters, T. O. Zander, B. Schölkopf, and M. Grosse-Wentrup,
Int. Conf. Autonomic Comput., 2017, pp. 83–92. “Predicting motor learning performance from electroencephalographic
[49] M. P. Hosseini, T. X. Tran, D. Pompili, K. Elisevich, and H. Soltanian- data,” J. Neuroeng. Rehabil., vol. 11, no. 1, p. 24, 2014.
Zadeh, “ Multimodal data analysis of epileptic EEG and RS-fMRI via [74] M. Mirsadeghi, H. Behnam, R. Shalbaf, and H. J. Moghadam, “Charac-
deep learning and edge computin,” Artif. Intell. Medicine, vol. 104, 2020, terizing awake and anesthetized states using a dimensionality reduction
Art no. 101813. method,” J. Med. Syst., vol. 40, no. 1, p. 13, 2016.
[50] A. E. Hramov et al., “Classifying the perceptual interpretations of a [75] W. Mumtaz, S. S. A. Ali, M. A. M. Yasin, and A. S. Malik, “A ma-
bistable image using EEG and artificial neural networks,” Frontiers chine learning framework involving EEG-based functional connectivity
Neurosci., vol. 11, p. 674, 2017. to diagnose major depressive disorder (MDD),” Med. Biological Eng.
[51] W.-Y. Hsu, “Assembling a multi-feature EEG classifier for left–right Comput., vol. 56, no. 2, pp. 233–246, 2018.
motor imagery data using wavelet-based fuzzy approximate entropy [76] W. Mumtaz et al., “An EEG-based functional connectivity measure for
for improved accuracy,” Int. J. Neural Syst., vol. 25, no. 08, 2015, automatic detection of alcohol use disorder,” Artif. Intell. Medicine,
Art. no. 1550037. vol. 84, pp. 79–89, 2018.
[52] J. Hu and J. Min, “Automated detection of driver fatigue based on EEG [77] M. Murakami, S. Nakatani, N. Araki, Y. Konishi, and K. Mabuchi,
signals using gradient boosting decision tree model,” Cogn. Neurodyn., “Motion discrimination from EEG using logistic regression and schmitt-
vol. 12, no. 4, pp. 431–440, Aug. 2018. trigger-type threshold,” in Proc. IEEE Int. Conf. Syst., Man, Cybern.,
[53] J. Hu and Z. Mu, “EEG authentication system based on auto-regression 2015, pp. 2338–2342.
coefficients,” in Proc. IEEE 10th Int. Conf. Intell. Syst. Control, 2016, [78] M.-R. Nazem-Zadeh et al., “Lateralization of temporal lobe epilepsy
pp. 1–5. by imaging-based response-driven multinomial multivariate models,”
[54] A. Ishfaque, A. J. Awan, N. Rashid, and J. Iqbal, “Evaluation of ANN, in Proc. IEEE 36th Annu. Int. Conf. Eng. Medicine Biol. Soc., 2014,
LDA and decision trees for EEG based brain computer interface,” in Proc. pp. 5595–5598.
IEEE 9th Int. Conf. Emerging Technol., 2013, pp. 1–6. [79] E. Neto, F. Biessmann, H. Aurlien, H. Nordby, and T. Eichele, “Regular-
[55] I. Iturrate, L. Montesano, and J. Minguez, “Robot reinforcement learning ized linear discriminant analysis of EEG features in dementia patients,”
using EEG-based reward signals,” in Proc. IEEE Int. Conf. Robot. Autom., Frontiers Aging Neurosci., vol. 8, p. 273, 2016.
number EPFL-CONF-205134, pp. 4822–4829, 2010. [80] A. Onishi and K. Natsume, “Multi-class ERP-based BCI data analysis
[56] A. Jain, B. Abbas, O. Farooq, and S. K. Garg, “Fatigue detection and using a discriminant space self-organizing map,” in Proc. IEEE 36th
estimation using auto-regression analysis in EEG,” in Proc. IEEE Int. Annu. Int. Conf. Eng. Medicine Biol. Soc., 2014, pp. 26–29.
Conf. Adv. Comput., Commun. Informat., 2016, pp. 1092–1095. [81] M. S. Özerdem and H. Polat, “Emotion recognition based on EEG features
[57] A. K. Jain, “Data clustering: 50 years beyond k-means,” Pattern Recognit. in movie clips with channel selection,” Brain Informat., vol. 4, no. 4,
Lett., vol. 31, no. 8, pp. 651–666, 2010. p. 241, 2017.
[58] A. K. Jaiswal and H. Banka, “Epileptic seizure detection in EEG signal [82] A. Page, S. P. T. Oates, and T. Mohsenin, “An ultra low power feature
with GModPCA and support vector machine,” Bio-Med. Mater. Eng., extraction and classification system for wearable seizure detection,”
vol. 28, no. 2, pp. 141–157, 2017. in Proc. IEEE 37th Annu. Int. Conf. Eng. Medicine Biol. Soc., 2015,
[59] L. Jakaite, V. Schetinin, C. Maple, and J. Schult, “Bayesian decision pp. 7111–7114.
trees for EEG assessment of newborn brain maturity,” in Proc. IEEE UK [83] S. K. Prabhakar and H. Rajaguru, “PCA and k-means clustering for
Workshop Comput. Intell., 2010, pp. 1–6. classification of epilepsy risk levels from EEG signals? A comparitive
[60] A. Jalilifard, E. B. Pizzolato, and M. K. Islam, “Emotion classification study between them,” in Proc. IEEE Int. Conf. Intell. Informat. Biomed.
using single-channel scalp-EEG recording,” in Proc. IEEE 38th Annu. Sci., 2015, pp. 83–86.
Int. Conf. Eng. Medicine Biol. Soc., 2016, pp. 845–849. [84] S. Puntanen, “Regression analysis by example, by samprit chatterjee, Ali
[61] S. Jukić and J. Kevrić, “Majority vote of ensemble machine learning S. Hadi,” Int. Statistical Rev., vol. 81, no. 2, pp. 308–308, 2013.
methods for real-time epilepsy prediction applied on EEG pediatric data,” [85] A. F. Rabbi, L. Azinfar, and R. Fazel-Rezai, “Seizure prediction using
TEM J., vol. 7, no. 2, p. 313, 2018. adaptive neuro-fuzzy inference system,” in Proc. IEEE 35th Annu. Int.
[62] J. Kevric and A. Subasi, “The effect of multiscale PCA de-noising in Conf. Eng. Medicine Biol. Soc., 2013, pp. 2100–2103.
epileptic seizure detection,” J. Med. Syst., vol. 38, no. 10, p. 131, 2014. [86] A. F. Rabbi and R. Fazel-Rezai, “A fuzzy logic system for seizure onset
[63] J.-H. Kim, F. Bießmann, and S.-W. Lee, “Reconstruction of hand move- detection in intracranial EEG,” Comput. Intell. Neurosci., vol. 2012, p. 1,
ments from EEG signals based on non-linear regression,” in Proc. IEEE 2012.
Int. Winter Workshop Brain-Comput. Interface, 2014, pp. 1–3. [87] K. Rai, V. Bajaj, and A. Kumar, “Novel feature for identification of focal
[64] C. Kingsford and S. L. Salzberg, “What are decision trees? Nature EEG signals with k-means and fuzzy c-means algorithms,” in Proc. IEEE
Biotechnol., vol. 26, no. 9, p. 1011, 2008. Int. Conf. Digit. Signal Process., 2015, pp. 412–416.
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.
218 IEEE REVIEWS IN BIOMEDICAL ENGINEERING, VOL. 14, 2021
[88] H. Rajaguru and S. K. Prabhakar, “Logistic regression Gaussian mixture [111] S. Yuan, W. Zhou, and L. Chen, “Epileptic seizure prediction using dif-
model and softmax discriminant classifier for epilepsy classification from fusion distance and bayesian linear discriminate analysis on intracranial
EEG signals,” in Proc. IEEE Int. Conf. Comput. Methodologies Commun., EEG,” Int. J. Neural Syst., vol. 28, no. 1, 2018, Art. no. 1750043.
2017, pp. 985–988. [112] T. Zhang and W. Chen, “LMD based features for the automatic seizure
[89] H. Rajaguru and S. K. Prabhakar, “Non linear ICA and logistic regression detection of EEG signals using SVM,” IEEE Trans. Neural Syst. Rehabil.
for classification of epilepsy from EEG signals,” in Proc. IEEE Int. Conf. Eng., vol. 25, no. 8, pp. 1100–1108, Aug. 2017.
Electron., Commun. Aerosp. Technol., 2017, vol. 1, pp. 577–580. [113] N. Zhuang, Y. Zeng, K. Yang, C. Zhang, L. Tong, and B. Yan, “Investi-
[90] H. Rajaguru and S. K. Prabhakar, “Sparse PCA and soft decision gating patterns for self-induced emotion recognition from EEG signals,”
tree classifiers for epilepsy classification from EEG signals,” in Proc. Sensors, vol. 18, no. 3, p. 841, 2018.
IEEE Int. Conf. Electron., Commun. Aerosp. Technol., 2017, vol. 1,
pp. 581–584.
[91] I. Rish et al., “An empirical study of the naive bayes classifier,” in IJCAI
2001 Workshop on Empirical Methods in Artificial Intelligence, vol. 3.
IBM: New York, 2001, pp. 41–46. Mohammad-Parsa Hosseini (Senior Member,
[92] S. Roy, I. Kiral-Kornek, and S. Harrer, “Deep learning enabled automatic IEEE) received the B.Sc. degree in electrical
abnormal EEG identification,” in Proc. IEEE 40th Annu. Int. Conf. Eng. and electronic engineering in 2006, the M.Sc.
Medicine Biol. Soc., 2018, pp. 2756–2759. degree in biomedical engineering in 2008, and
[93] K.-M. Rytkönen, J. Zitting, and T. Porkka-Heiskanen, “Automated sleep the M.Sc. degree in electrical and communica-
scoring in rats and mice using the naive bayes classifier,” J. Neurosci. tion engineering in 2010. He received the Ph.D.
Methods, vol. 202, no. 1, pp. 60–64, 2011. degree in electrical and computer engineering
[94] C. Y. Sai, N. Mokhtar, H. Arof, P. Cumming, and M. Iwahashi, “Auto- with research in computer science from Rutgers
mated classification and removal of EEG artifacts with SVM and wavelet- University, New Brunswick, NJ, USA, in 2018.
ICA,” IEEE J. Biomed. Health Informat., vol. 22, no. 3, pp. 664–670, He did a graduate study in electrical engineering
2018. at Wayne State University, Detroit, MI, USA, in
[95] B. Sharif and A. H. Jafari, “Prediction of epileptic seizures from EEG us- 2013. He is collaborating with Medical Image Analysis Laboratory, Henry
ing analysis of ictal rules on poincaré plane,” Comput. Methods Programs Ford Health System, and with the Clinical Neuroscience Department,
Biomedicine, vol. 145, pp. 11–22, 2017. Spectrum Health, Grand Rapids, MI, USA. He has been a Senior Data
[96] A. Sharma, J. Rai, and R. Tewari, “Epileptic seizure anticipation and Scientist and Machine Learning Researcher in Silicon Valley, CA, USA,
localisation of epileptogenic region using EEG signals,” J. Med. Eng. since 2017. He has also been an Adjunct Lecturer and Faculty Member
Technol., vol. 42, no. 3, pp. 203–216, Apr. 2018. with several universities since 2009 and is currently with Santa Clara
[97] A. Sharmila and P. Geethanjali, “Effect of filtering with time domain University. His current research interests include machine learning, deep
features for the detection of epileptic seizure from EEG signals,” J. Med. learning, and signal and image processing. He was on the scientific
Eng. Technol., vol. 42, no. 3, pp. 217–227, Apr. 2018. committees and review boards of several national and international
[98] A. Sharmila and P. Mahalakshmi, “Wavelet-based feature extraction for conferences and journals.
classification of epileptic seizure EEG signal,” J. Med. Eng. Technol.,
vol. 41, no. 8, pp. 670–680, 2017.
[99] V. Srinivasan, C. Eswaran, and N. Sriraam, “Approximate entropy-based
epileptic EEG detection using artificial neural networks,” IEEE Trans.
Amin Hosseini (Member, IEEE) is with the De-
Inf. Technol. Biomedicine, vol. 11, no. 3, pp. 288–295, 2007.
partment of Electrical and Computer Engineer-
[100] A. F. Struck et al., “Association of an electroencephalography-based risk
ing with minor in computer science at Azad
score with seizure probability in hospitalized patients,” JAMA Neurol.,
University, Central Branch, Tehran, Iran. His re-
vol. 74, no. 12, pp. 1419–1424, 2017.
search interests include digital signal and im-
[101] A. Subasi, “Application of adaptive neuro-fuzzy inference system for
age processing, machine learning, artificial in-
epileptic seizure detection using wavelet feature extraction,” Comput.
telligence and biomedical engineering. He is a
Biol. Medicine, vol. 37, no. 2, pp. 227–244, 2007.
member of the IEEE Signal Processing Society
[102] T. Teramae, D. Kushida, F. Takemori, and A. Kitamura, “Estimation of
and the IEEE Machine Learning Society.
feeling based on EEG by using NN and k-means algorithm for massage
system,” in Proc. IEEE SICE Annu. Conf., 2010, pp. 1542–1547.
[103] A. Torabi, M. R. Daliri, and S. H. Sabzposhan, “Diagnosis of multiple
sclerosis from EEG signals using nonlinear methods,” Australasian Phys.
Eng. Sci. Medicine, vol. 40, no. 4, pp. 785–797, 2017.
[104] M. S. Treder, A. K. Porbadnigk, F. S. Avarvand, K.-R. Müller, and B.
Blankertz, “The LDA beamformer: Optimal estimation of ERP source Kiarash Ahi (Member, IEEE) received the
time series using linear discriminant analysis,” Neuroimage, vol. 129, M.Sc. degree in electrical and information engi-
pp. 279–291, 2016. neering from the Leibniz University of Hannover,
[105] V. Tuyisenge et al., “Automatic bad channel detection in intracranial Germany in 2012, and the Ph.D. degree in elec-
electroencephalographic recordings using ensemble machine learning,” trical and computer engineering from the Uni-
Clin. Neurophysiol., vol. 129, no. 3, pp. 548–554, 2018. versity of Connecticut, USA, in 2017. The focus
[106] V. Vijayakumar, M. Case, S. Shirinpour, and B. He, “Quantifying and of his M.Sc. degree has been smart grid, re-
characterizing tonic thermal pain across subjects from EEG data using newable energy systems, and power electron-
random forest models,” IEEE Trans. Biomed. Eng., vol. 64, no. 12, ics. His Ph.D. studies have been in the areas
pp. 2988–2996, 2017. of semiconductor technology, optics, machine
[107] Y. Wang, W. Chen, K. Huang, and Q. Gu, “Classification of neona- learning and natural computation, compressive
tal amplitude-integrated EEG using random forest model with com- sensing, and terahertz signal and image processing. He is currently
bined feature,” in Proc. IEEE Int. Conf. Bioinform. Biomed., 2013, a Senior Researcher and Lead Product Development Engineer in the
pp. 285–290. advanced semiconductor and software industry, where he researches
[108] S. Weichwald, T. Meyer, B. Scholkopf, T. Ball, and M. Grosse-Wentrup, areas where artificial intelligence can enhance the accuracy and effi-
“Decoding index finger position from EEG using random forests,” in ciency of semiconductor device manufacturing toward driving Moore’s
Proc. IEEE 4th Int. Workshop Cognitive Inf. Process., 2014, pp. 1–6. law beyond 7-nm technology node. He architects automated systems,
[109] X. Ying, H. Lin, and G. Hui, “Study on non-linear bistable dynamics empowered by machine learning and image processing, and leads
model based EEG signal discrimination analysis method,” Bioengi- multinational R&D teams. His scientific and research interests include
neered, vol. 6, no. 5, pp. 297–298, 2015. digital image and signal processing, optics and photolithography, MEMS
[110] W. Yu, T. Liu, R. Valdez, M. Gwinn, and M. J. Khoury, “Application of and semiconductor devices, machine learning and artificial intelligence,
support vector machine modeling for prediction of common diseases: The hardware security, bioengineering, wearable technologies, embedded
case of diabetes and pre-diabetes,” BMC Med. Informat. Decis. Making, systems, human-computer interaction, terahertz technology, and intel-
vol. 10, no. 1, p. 16, Mar. 2010. ligent software development.
Authorized licensed use limited to: CZECH TECHNICAL UNIVERSITY. Downloaded on July 15,2022 at 16:45:08 UTC from IEEE Xplore. Restrictions apply.