Human Activity Recognition For Elderly People Using Machine and Deep Learning Approaches
Human Activity Recognition For Elderly People Using Machine and Deep Learning Approaches
net/publication/361025278
Human Activity Recognition for Elderly People Using Machine and Deep
Learning Approaches
CITATIONS READS
30 880
4 authors:
All content following this page was uploaded by Ahatsham Hayat on 02 June 2022.
Article
Human Activity Recognition for Elderly People Using Machine
and Deep Learning Approaches
Ahatsham Hayat 1,2, *,† , Morgado Dias 1,2, *,† , Bikram Pratim Bhuyan 3,† and Ravi Tomar 3, *,†
Abstract: There are more than 962 million people aged 60 and up globally. Physical activity declines
as people get older, as does their capacity to undertake everyday tasks, effecting both physical and
mental health. Many researchers use machine learning and deep learning methods to recognize
human activities, but very few studies have been focused on human activity recognition of elderly
people. This paper focuses on providing assistance to elderly people by monitoring their activities in
different indoor and outdoor environments using gyroscope and accelerometer data collected from
a smart phone. Smart phones have been routinely used to monitor the activities of persons with
impairments; routine activities such as sitting, walking, going upstairs, going downstairs, standing,
and lying are included in the dataset. Conventional Machine Learning and Deep Learning algorithms
such as k-Nearest Neighbors, Random Forest, Support Vector Machine, Artificial Neural Network,
and Long Short-Term Memory Network are used for human activity recognition. Long Short-Term
Memory is a recurrent neural network variation that is best suited to handling temporal sequences.
Citation: Hayat, A.; Dias, M.; Two-fold and ten-fold cross-validation methods were performed to show the effect of changing the
Bhuyan, B.P.; Tomar, R. Human data in the training and testing dataset. Among all the classification techniques, the proposed Long
Activity Recognition for Elderly Short-Term Memory Network gave the best accuracy of 95.04%. However, Support Vector Machine
People Using Machine and Deep gave 89.07% accuracy with a very low computational time of 0.42 min using 10-fold cross-validation.
Learning Approaches. Information
2022, 13, 275. https://doi.org/ Keywords: human activity recognition; elderly people; machine learning; deep learning
10.3390/info13060275
of people with disabilities [4,5]. One of them is the requirement for a companion to monitor
their activities. To safeguard people with disabilities from injury, danger, or accidents,
they must be protected and supervised 24 h per day [6]. Mobile phones, on the other
hand, have been frequently used to monitor people’s activities [7,8]. In 2017, the global
population of people aged 60 and above totaled 962 million, more than doubling the
382 million people aged 60 and up in 1980. By 2050, the number of elderly people is
estimated to more than double, reaching over 2.1 billion. Two-thirds of the world’s elderly
reside in developing countries, where their numbers are increasing at a quicker rate than in
industrialized countries. Nearly 8 out of 10 of the world’s elderly people are anticipated to
live in underdeveloped countries by 2050. The proportion of people aged 60 and over who
live “independently”—alone or with a spouse exclusively—varied substantially among the
143 nations or territories with accessible data, ranging from 2.3 percent in Afghanistan to
93.4 percent in The Netherlands.
Artificial intelligence becomes more popular for HAR because of its self-learning
nature and robust classification models [9]. In recent years, several studies have been con-
ducted for HAR using machine learning and deep learning [10–12], but only a few focus on
developing a framework for the HAR system for elderly people. Kaixuan Chen et al. [13] in
their study gave an overview of challenges and opportunities in the area of human activity
recognition using deep learning techniques. Due to the diversity of sensor data, it is really
important to have multimodal features, which could help in maximizing the performance
of the system [14]. The attention-based mechanism could be used in HAR to highlight
the most important and differentiable modalities (Zeng et al. [15] and Chen et al. [16]).
Chen et al. [16] used multiple agents to focus on modalities that are related to sub-motions.
Although they have outperformed all the state-of-the-art methods, still, they have not been
validated on a dataset of the elderly. This study mainly focuses on finding out the most
effective approach for developing the HAR framework for elderly people. Both machine
learning and deep learning approaches such as k-Nearest Neighbors (k-NN), Random
Forest (RF), Support Vector Machine (SVM), Artificial Neural Network (ANN), and Long
Short-Term Memory (LSTM) are used in this study to see which approach is more suitable.
HAR algorithms that give a high accuracy, precision, and recall of activity classification
are required for this human-activity-monitoring application for elderly people. Machine
learning approaches for classifying human activities have previously been mentioned in
a few prior studies [17]. Classification or supervised learning is a data mining (machine
learning) system used for the prediction of the group to which a particular set of information
belongs [18]. Data analysis can be performed using classification and prediction; it helps in
the division of data into classes making the prediction of future trends easier. Grouping
helps put labels on the data, and prediction helps in determining the data value [19].
There are different algorithms of characterization that have been utilized for HAR: Naive
Bayes, RF, k-NN, SVM, ANNs, and Recurrent Neural Networks (RNN) are among the
most well-known approaches investigated. Among all, the RF algorithm is designed to
deliver high accuracy and consistent speed for data mining, particularly classification with
advanced features [20–23]. Unfortunately, all of these state-of-the-art algorithms are not
compared very effectively in terms of performance, such as precision, accuracy, F1-score,
and computation speed. In this work, all of the algorithms are evaluated with the above
criteria for better understanding.
Figure 1 shows the basic building blocks for human activity recognition. Raw data
will be collected from the smartphone, and various preprocessing steps will be performed
on the dataset such as Principal Component Analysis (PCA), etc. After that, features will
be extracted from the dataset, and the model will be trained on those features. In the final
steps, user activity will be detected.
Information 2022, 13, 275 3 of 13
2. Literature Review
Several studies on human activity recognition have been conducted in the past few
years. Some of the significant literature works in this area that concentrate on real-time
processing, as well as offline processing are listed below.
Activity recognition using smartphones has its own advantages because smartphones
are very easy to establish and are robust in nature. Usharani J et al. [24] came up with
an idea for a human activity recognition system based on the Android platform. They
created an application using the accelerometer data for classification, which supported
on-line training and classification. They used the clustered k-NN approach to enhance the
performance, accuracy, and execution time of the k-NN classifier with limited resources
on the Android platform. They also concluded that the classification times were also
dependent on the device models and capabilities. Similarly, Meysam, Vakili et al. [25]
proposed a real-time HAR model for online prediction of human physical movements
based on the smartphone inertial sensors. A total of 20 different activities were selected,
and six incremental learning algorithms were used to check the performance of the system,
then all of them were also compared with the state-of-the-art HAR algorithms such as
Decision Trees (DTs), AdaBoost, etc. Incremental k-NN and Incremental Naive Bayesian
have given the best accuracy of 95%.
In Jirapond Muangprathub et al.’s [26] paper, they introduced a novel elderly person
tracking system using a machine learning algorithm. In this work, they used the k-NN
model with a k value of 5, which was able to achieve the best accuracy of 96.40% in detecting
the real-time activity of elderly people. Furthermore, they created a system that displays
information in a spatial format for an elderly person, and in case of an emergency, they can
use a messaging device to request any help.
In Davide Anguita et al.’s paper [27], they introduced the improvised Support Vector
Machine algorithm, which works with fixed point arithmetic to produce an energy-efficient
model for the classification of human activities using a smartphone. They aimed to use
the presented novel technology for various intelligence applications and smart environ-
ments for faster processing with the least possible use of system resources to save the
consumption of energy along with maintaining comparable results with other generally
used classification techniques.
To understand people’s behavior in different places such as homes, clinics, etc., Md
Zia Uddin et al. [28] proposed a body-sensor-based activity recognition system using
deep Neural Stretchered Learning based on Long Short-Term Memory (LSTM). For better
clustering of features from all the activities, Kernel-based Discriminant Analysis (KDA)
was applied, which will maximize inter-class scattering and minimize intra-class scattering
of the samples. The proposed model successfully achieved a recall of 99%, which was
further compared to the existing deep learning models such as the RNN, Convolutional
Neural Network (CNN), and Deep Belief Network (DBN). None of the existing deep
learning models would be able to achieve the recall rate of the proposed method. Similarly,
Abdulmajid Murad et al. [29] proposed a deep LSTM network for recognizing six different
activities based on smartphone data. The network was able to achieve an accuracy of
96.70% on the UCI-HAD dataset.
Baoding Zhou et al. [30] proposed a CNN for indoor human activity recognition. A
total of nine different activities were recognized based on accelerometers, magnetometers,
Information 2022, 13, 275 4 of 13
gyroscopes, and barometers collected by smartphones. The proposed method was able to
achieve an excellent accuracy of 98%. Yashi Nan et al. [31] used combinations of CNN with
LSTM to recognize the activity of older people. Out of all the combinations, a multichannel
CNN-LSTM model was able to achieve the best accuracy of 81.1%.
In conclusion, most of the research focused on recognizing overall human activities is
not specific to elderly people, and they are not evaluated on extensive evaluation param-
eters. This study focused on the overall activity recognition of elderly people, and more
complex evaluation parameters were used for checking the performance of the models.
3. Methodology
3.1. Dataset Description
A dataset from the University of California Irvine (UCI) was used for this study.
Data were collected from a variety of participants aged between 19 and 48 who had a
smartphone [32]. The phone’s accelerometer and gyroscope may be utilized to capture
the acceleration and angular velocity of the data at a sample rate of 50 Hz. To process the
signals, a median filter and a notch filter with a drop frequency of 20 Hz were utilized.
There were 561 features obtained from the dataset, and 10,299 sample were present in
the dataset. The distribution of the dataset is shown in Figure 2. The data were split into
two sections, with 80 percent being training data and 20 percent being testing data. In
other words, there were 8240 training data and 2059 testing data. This dataset has six
different categories of user activity: sitting, walking, going upstairs, going downstairs,
standing, lying.
Furthermore, for the validation of the existing system, it was cross-validated on one
more dataset “Activity recognition with healthy older people using a batteryless wearable
sensor dataset” [33]. This dataset was collected from 14 participant aged between 66
and 86 in two clinical rooms. This dataset 75,128 samples and categories into 4 different
user activities: sitting on the bed (16,406 instances), sitting on the chair (4911 instances),
lying on the bed (51,520 instances), and ambulating (walking or standing within the room)
(2291 instances).
Figure 2. Representation of the distribution of activities such as standing, sitting, laying, walking,
going upstairs, and going downstairs with respect to the total data of 10,299 samples. From the data,
it can be seen that the most-performed activity is sitting, and the least is going downstairs.
3.3. Classification
3.3.1. Random Forest
RF classification consists of a multitude of decision trees. All these trees work together
to give the result. A simple concept of voting is utilized where the class with the highest
number of votes is the class prediction [35]. The bagging method was used in this study to
improve the classification performance of a single tree. Different numbers of decision trees
of RF were used in the training process {5, 10, 15, 20, 25, 30, 35, 40, 45, 50} to see how many
decision trees were required to achieve the best results. After comparing all the results from
all the activities, an RF with a 30-decision-tree classifier gave the best accuracy among all.
However, in order to achieve the best result, the RF requires a large number of labeled data.
For impurity reduction, the Gini impurity index was calculated. When all the elements in
the node are unique, it is called a pure node and its value will be 0. That means this node
cannot be split after. The Gini index value will be maximum when the probability of two
classes is the same [36].
2
Gini (n) = 1 − ∑ p2j (1)
j =1
3.3.2. k-NN
A simple, easy-to-implement method for supervised learning, k-NN, can be used for
HAR [37]. It finds the most similar data points in training data to classify and makes an
educated guess about the classification. The Euclidean distance was used to calculate the
distance between two points. “k” is a user-defined constant, which is generally an odd
number, which helps with the classification by querying the nearest point [38].
s
n
d( x, y) = ( ∑ ( x i − x j )2 ) (2)
i =1
In this study, the value of “k” for training was taken from 2 to 50, and it was observed that
when value of the “k” increases, the accuracy of the model also increases till a certain point,
and then, it starts decreasing because the model becomes too generalized. k-NN with 20 as
the “k” value gave the best accuracy among all.
|| X1 − X2 ||2
K ( X1 , X2 ) = exp(− ) (3)
σ2
where X1 and X2 are data points. || X1 − X2 || represents the Euclidean distance between X1
and X2 . σ is the variance, which is inversely proportional to γ.
Information 2022, 13, 275 6 of 13
Figure 3. Architecture of the proposed artificial feed-forward neural network with 3 hidden layers.
The output layer contains 6 nodes and uses the softmax activation function described
in Equation (8) [8]. The Rectified Linear Unit (ReLu) activation function was used in the
hidden layers [42,43]. The Adam optimizer was used for training the neural network with
categorical cross-entropy as the loss function [44]. The proposed MLP’s mathematical
computation can be described as follows:
Input layer: The first layer has “N” linear combinations of the d-dimensional inputs.
∑ wij
(1) (1)
ai = x j + bi (4)
j
Here, ai is called the activation, which is further used by the activation function. i = 1, 2,
(1) (1)
. . ., N. wij are the weights; x j represents the input; bi is a biased value. Superscript (1)
indicates that this is the first layer of the neural network.
Hidden layer: In each hidden layer, the weights and output vector of the previous
layer are the input of the new layer.
For some kth layer,
0
hik = φ (∑ wij x j + bi )
(k) (k)
(5)
j
0
Here, φ is the ReLu activation function, which is described as follows.
0
φ = 0∀ z < 0
(6)
= z∀z>=0
Output layer:
00 ( n −1) ( n −1)
yi = φ (∑ wij x j + bi ) (7)
j
00
Here, φ represents the activation function. The output layer’s activation function
specifies how the weighted sum of the input is converted. The softmax function was used
as the activation function for the output layer.
Information 2022, 13, 275 7 of 13
00 exp(zi )
φ = (8)
∑ j exp(z j )
Here, ht and h(t−1) represent the activation at time t and t − 1. φ is the activation function;
wih and whh are the input–hidden and hidden–hidden weights; respectively; bt is the
bias value.
Figure 4. Architecture of the proposed LSTM model for human activity recognition.
The output of LSTM was flattened and fed to the fully connected layer with the ReLu
activation function, and at the output layer, the softmax activation function was used for
the activity recognition. The Adam optimizer was used for network optimization, and the
categorical cross-entropy function was used to calculate the loss.
the right activity, and FN represents the number of incorrect predictions of the wrong
activity. The precision, recall, accuracy, and F1-score values were calculated using the above
parameters defined in Equations (10)–(13).
TP
Precision = (10)
( TP + FP)
TP
Recall = (11)
( TP + FN )
( TP + TN )
Accuracy = (12)
( TP + TN + FP + FN )
(2TP)
F1-score = (13)
(2TP + FP + FN )
Processing time was also considered as an evaluation criterion in this study because of
the complex nature of the dataset, so it is important to measure the processing time of the
algorithms. Two-fold cross-validation and 10-fold cross-validation were performed as the
train–test processes in order to observe the effect of the change in the train–test dataset
balance on the output. Furthermore, the “activity recognition with healthy older people
dataset” was used as a validation dataset to check the generalization capabilities of the
proposed methods.
Figure 5. Confusion matrix of the (a) RF, (b) k-NN (c) SVM (d) ANN, and (e) LSTM.
Information 2022, 13, 275 9 of 13
From Tables 1 and 2, it is clearly visible that the LSTM approach outperformed all the
other methods in most of the cases; the ANN gave the best accuracy in classifying “sitting”
for both 2-fold and 10-fold cross-validation. If we compare deep learning methods (ANN
and LSTM) with all the other methods, it is clearly depicted that deep learning methods
outperformed other conventional machine learning methods in terms of accuracy. The
overall accuracy of LSTM (95.05%) was the highest.
Table 1. Accuracy table of all the algorithms for each activity using 2-fold cross-validation.
Activity RF (%) k-NN (%) SVM (%) ANN (%) LSTM (%)
Walking 85.60 86.31 89.47 93.78 95.08
Going Upstairs 76.02 78.88 82.84 86.50 92.78
Going Downstairs 76.54 75.42 80.10 86.55 94.75
Sitting 87.62 93.81 93.92 95.57 96.85
Standing 87.22 90.54 91.54 94.88 95.94
Laying 83.10 87.34 89.50 90.10 91.77
Overall Performance 82.68 85.38 87.90 91.23 94.53
Table 2. Accuracy table of all the algorithms for each activity using 10-fold cross-validation.
Activity RF (%) k-NN (%) SVM (%) ANN (%) LSTM (%)
Walking 86.60 88.50 91.37 94.67 96.12
Going Upstairs 80.12 81.62 83.45 88.92 93.18
Going Downstairs 78.45 79.53 81.88 87.15 95.57
Sitting 92.76 94.16 95.89 97.22 97.11
Standing 90.14 91.12 92.14 95.14 96.12
Laying 85.70 88.16 89.73 90.38 92.18
Overall Performance 85.63 87.18 89.08 92.25 95.05
Additionally, from Figure 5 and Tables 1 and 2, it is observed that most of the misclas-
sified records came from “going upstairs” and “going downstairs”, as both activities are
somewhat similar, and the model had problems separating these two activities. Now, it is
possible to verify the performance of the system in terms of the other parameters as well.
Tables 3 and 4 show that the best precision, recall, accuracy, and F1-score values were
92.87%, 85.32%, 95.05%, and 88.94%, respectively, and were achieved by the LSTM classifier
using 10-fold cross-validation. However, SVM showed the best processing time of 0.08 and
0.42 min in both 2-fold and 10-fold cross-validation. Considering the overall performance
of the methods, it is clearly visible that the proposed LSTM’s overall performance was
better than other machine learning or deep learning models.
Model Avg. Precision (%) Avg. Recall (%) Avg. Accuracy (%) F1-Score (%) Time (min)
RF 81.40 78.41 82.68 79.88 0.53
k-NN 83.57 81.71 85.38 82.63 0.92
SVM 87.14 85.62 87.90 86.37 0.08
ANN 88.81 89.11 91.23 88.96 1.22
LSTM 90.78 92.62 94.53 91.69 0.72
Model Avg. Precision (%) Avg. Recall (%) Avg. Accuracy (%) F1-Score (%) Time (min)
RF 82.22 80.10 85.63 81.15 2.24
k-NN 85.41 82.30 87.18 83.83 4.20
SVM 88.98 87.80 89.08 88.39 0.42
ANN 89.20 91.78 92.25 90.47 3.78
LSTM 92.87 94.32 95.05 93.59 2.92
Information 2022, 13, 275 10 of 13
Figure 6 shows the training and loss curve versus epochs using 10-fold cross-validation
for both deep learning models. It was observed that the ANN gave a training loss as low as
0.24 and a validation loss of 0.23, and LSTM gave a training loss of 0.16 and a validation
loss of 0.15. In both, the model training and validation losses were higher initially, but at
the end of the epoch cycle, the loss drastically decreased, while training and validation
curves became almost equal. Training was stopped in both models because of the stopping
criteria (complete continuous 10-epoch cycle without change in the loss). As it was already
discussed above that at the end of the epoch cycle, both losses became almost equal for
both deep leaning models, this implies that both models were working perfectly. It was
also observed that in LSTM, the model training loss and validation loss difference was
higher in the beginning, but this drastically decreased in between 20 and 40 epoch cycles.
Furthermore, to check the robustness and validity of the proposed methods, they were
also validated on the “activity recognition with healthy older people using a batteryless
wearable sensor dataset” [33]. Table 5 shows the performance assessment on this dataset
using 10-fold cross-validation.
Figure 6. Training and validation curve for loss using 10-fold cross-validation: (a) ANN model,
(b) LSTM model.
Table 5. Performance assessment using 10-fold cross-validation on activity recognition with the
healthy older people dataset [33].
Model Avg. Precision (%) Avg. Recall (%) Avg. Accuracy (%) F1-Score (%)
RF 81.47 79.27 85.61 80.35
k-NN 86.77 83.14 86.35 84.92
SVM 87.12 86.81 88.13 86.96
ANN 90.57 90.29 91.77 90.43
LSTM 91.16 93.24 94.84 92.19
Information 2022, 13, 275 11 of 13
From Table 5, it is possible to say that the proposed LSTM method gave satisfying
results of 94.84%, 85.32%, 95.05%, and 88.94% for the precision, recall, accuracy, and F1-
score, respectively. Although the dataset was highly imbalanced, still, the proposed system
was able to achieve good performance. Therefore, it is possible to say that the proposed
method can be used for detecting the activity of elderly people.
From this study, it was observed that deep learning methods are more suitable for the
HAR of elderly people as compared to conventional machine learning methods. However,
if the processing time is a constraint, machine learning methods are much faster than deep
learning methods, especially SVM.
5. Conclusions
This research aimed to monitor the everyday activities of elderly people which will
safeguard the elderly people’s health as elderly people usually follow the same routine
day. Therefore, if a sudden change is seen in everyday activities, the right action can
be taken. Automatic activity identification might assist older individuals in operating
independently in their own homes in an aging society. In the proposed study, various
conventional machine learning and deep learning methods were used to recognize the
activity of elderly people.
The proposed work aimed to develop an automatic activity monitoring system that
will automatically detect the activity of elderly people using different machine and deep
learning methods such as RF, k-NN, SVM, ANN, and LSTM. Out of all the methods, the
LSTM method gave the best accuracy of 95.05%, which is significant, as building an accurate
HAR system is very challenging due to the large diversity of activities, and sometimes,
these are very similar in nature. However, there is still a need for improvement of the
proposed models by validating with larger datasets and with more activities. Our proposed
models can also be used in various other fields where the constant supervision of activities
is required.
Author Contributions: Conceptualization, A.H. and B.P.B.; methodology, M.D.; software, A.H.;
validation, R.T., B.P.B. and A.H.; formal analysis, A.H.; investigation, M.D.; resources, R.T.; data
curation, M.D.; writing—original draft preparation, A.H.; writing—review and editing, A.H., M.D.
and B.P.B.; visualization, A.H.; supervision, R.T.; project administration, R.T.; funding acquisition,
M.D. All authors have read and agreed to the published version of the manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Paul, P.; George, T. An Effective Approach for Human Activity Recognition on Smartphone. In Proceedings of the 2015 IEEE
International Conference on Engineering and Technology (ICETECH), Coimbatore, India, 20–20 March 2015; pp. 1–3.
2. World Health Organization; The World Bank. World Report on Disability 2011; World Health Organization: Geneva, Switzerland,
2011; Volume 91, p. 549.
3. Kementrian Kesehatan Republic Indonesia. Situasi Penyandang Disabilitas. 2014. Available online: https://pusdatin.kemkes.go.
id/article/view/15033100002/situasi-penyandang-disabilitas.html (accessed on 3 March 2022)
4. Du, Y.; Lim, Y.; Tan, Y. A Novel Human Activity Recognition and Prediction in Smart Home Based on Interaction. Sensors 2019,
19, 4474. [CrossRef] [PubMed]
5. Bustoni, I.A.; Hidayatulloh, I.; Azhari, S.N.; Augoestin, N.G. Multidimensional Earcon Interaction Design for The Blind: A
Proposal and Evaluation. In Proceedings of the 2018 International Seminar on Research of Information Technology and Intelligent
Systems (ISRITI), Yogyakarta, Indonesia, 21–22 November 2018; pp. 384–388. [CrossRef]
6. Chernbumroong, S.; Cang, S.; Atkins, A.; Yu, H. Elderly activities recognition and classification for applications in assisted living,
Expert Systems with Applications. Expert Syst. Appl. 2013, 40, 1662–1674. [CrossRef]
Information 2022, 13, 275 12 of 13
7. Bulbul, E.; Cetin, A.; Dogru, I.A. Human Activity Recognition Using Smartphones. In Proceedings of the 2018 2nd International
Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT), Ankara, Turkey, 19–21 October 2018; pp. 1–6.
[CrossRef]
8. Bayat, A.; Pomplun, M.; Tran, D.A. A Study on Human Activity Recognition Using Accelerometer Data from Smartphones.
Procedia Comput. Sci. 2014, 34, 450–457. [CrossRef]
9. Nguyen, T.H.C.; Nebel, J.C.; Florez-Revuelta, F. Recognition of activities of daily living with egocentric vision: A review. Sensors
2016, 16, 72. [CrossRef]
10. Dhiman, C.; Vishwakarma, D.K. Review of state-of-the-art techniques for abnormal human activity recognition. Eng. Appl. Artif.
Intell. 2019, 77, 21–45. [CrossRef]
11. Cheng, G.; Wan, Y.; Saudagar, A.N.; Namuduri, K.; Buckles, B.P. Advances in human action recognition: A survey. arXiv 2015,
arXiv:1501.05964.
12. Saraswathi, D.; Srinivasan, E. Performance analysis of mammogram CAD system using SVM and KNN classifier. In Proceedings
of the 2017 International Conference on Inventive Systems and Control (ICISC), Coimbatore, India, 19–20 January 2017; pp. 1–5.
13. Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep Learning for Sensor-based Human Activity Recognition: Overview,
Challenges, and Opportunities. ACM Comput. Surv. 2021, 54, 77. [CrossRef]
14. Chen, K.; Yao, L.; Zhang, D.; Wang, X.; Chang, X.; Nie, F. A semisupervised recurrent convolutional attention model for human
activity recognition. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1747–1756. [CrossRef]
15. Zeng, M.; Gao, H.; Yu, T.; Mengshoel, O.J.; Langseth, H.; Lane, I.; Liu, X. Understanding and improving recurrent networks for
human activity recognition by continuous attention. In Proceedings of the 2018 ACM International Symposium on Wearable
Computers, Singapore, 8–12 October 2018; ACM: New York, NY, USA, 2018; pp. 56–63.
16. Chen, K.; Yao, L.; Zhang, D.; Guo, B.; Yu, Z. Multi-agent Attentional Activity Recognition. In Proceedings of the Twenty-Eighth
International Joint Conference on Artificial Intelligence, IJCAI, Macao, China, 10–16 August 2019; pp. 1344–1350.
17. Bustoni, I.A.; Hidayatulloh, I.; Ningtyas, A.M.; Purwaningsih, A.; Azhari, S.N. Classification methods performance on human
activity recognition. J. Phys. Conf. Ser. ICTVT 2019, 1456, 012027. [CrossRef]
18. Xia, K.; Huang, J.; Wang, H. LSTM-CNN Architecture for Human Activity Recognition. IEEE Access 2020, 8, 56855–56866.
[CrossRef]
19. Zhang, S.; Wei, Z.; Nie, J.; Huang, L.; Wang, S.; Li, Z. A review on human activity recognition using vision-based method. J.
Healthc. Eng. 2017. Available online: https://www.hindawi.com/journals/jhe/2017/3090343/ (accessed on 10 July 2019).
20. Ougiaroglou, S.; Nanopoulos, A.; Papadopoulos, A.N.; Manolopoulos, Y.; Welzer-Druzovec, T. Adaptive k-Nearest-Neighbor
Classification Using a Dynamic Number of Nearest Neighbors. In Proceedings of the 11th East European Conference, ADBIS
2007, Varna, Bulgaria, 29 September–3 October 2007; pp. 66–82.
21. Reyes-Ortiz, J.; Ghio, A.; Anguita, D.; Parra-Llanas, X.; Cabestany, J.; Catalá, A. Human activity and motion disorder recognition:
Towards smarter interactive cognitive environments. In Proceedings of the European Symposium on Artificial Neural Networks,
Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013.
22. Madarshahian, R.; Caicedo, J.M. Human Activity Recognition Using Multinomial Logistic Regression. Model Valid. Uncertain.
Quantif. 2015, 3, 363–372.
23. Uddin, T.; Billah, M.; Hossain, F. Random forests based recognition of human activities and postural transitions on smartphone.
In Proceedings of the 2016 5th International Conference on Informatics, Electronics and Vision (ICIEV), Dhaka, Bangladesh, 13–14
May 2016; pp. 250–255. [CrossRef]
24. Usharani, J.; Saktivel, U. Human Activity Recognition using Android Smartphone. In Proceedings of the International Conference
on Innovations in Computing & Networking ICICN16, Bengaluru, Karnataka, 12–13 May 2016.
25. Vakili, M.; Rezaei, M. Incremental Learning Techniques for Online Human Activity Recognition. arXiv 2021, arXiv:2109.09435.
26. Muangprathub, J.; Sriwichian, A.; Wanichsombat, A.; Kajornkasirat, S.; Nillaor, P.; Boonjing, V. A Novel Elderly Tracking System
Using Machine Learning to Classify Signals from Mobile and Wearable Sensors. Int. J. Environ. Res. Public Health 2021, 18, 12652.
[CrossRef] [PubMed]
27. Anguita, D.; Ghio, A.; Oneto, L.; Parra-Llanas, X.; Reyes-Ortiz, J. Energy Efficient Smartphone-Based Activity Recognition using
Fixed-Point Arithmetic. J. Univers. Comput. Sci. 2013, 19, 1295–1314.
28. Uddin, Z.; Soylu, A. Human activity recognition using wearable sensors, discriminant analysis, and long short-term memory-
based neural structured learning. Sci. Rep. 2021, 11, 16455. [CrossRef]
29. Murad, A.; Pyun, J.-Y. Deep Recurrent Neural Networks for Human Activity Recognition. Sensors 2017, 17, 2556. [CrossRef]
30. Zhou, B.; Yang, J.; Li, Q. Smartphone-Based Activity Recognition for Indoor Localization Using a Convolutional Neural Network.
Sensors 2019, 19, 621. [CrossRef]
31. Nan, Y.; Lovell, N.H.; Redmond, S.J.; Wang, K.; Delbaere, K.; van Schooten, K.S. Deep Learning for Activity Recognition in Older
People Using a Pocket-Worn Smartphone. Sensors 2020, 20, 7195. [CrossRef]
32. Anguita, D.; Ghio, A.; Oneto, L.; Parra Perez, X.; Reyes Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition
Using Smartphones. In Proceedings of the 21th European Symposium on Artificial Neural Networks, Computational Intelligence
and Machine Learning, ESANN 2013, Bruges, Belgium, 24–26 April 2013.
Information 2022, 13, 275 13 of 13
33. Torres, R.L.S.; Ranasinghe, D.C.; Shi, Q.; Sample, A.P. Sensor enabled wearable RFID technology for mitigating the risk of falls
near beds. In Proceedings of the 2013 IEEE International Conference on RFID, Orlando, FL, USA, 30 April–2 May 2013; IEEE:
Piscataway, NJ, USA, 2013; pp. 191–198.
34. Alrazzak, U.; Alhalabi, B. A survey on human activity recognition using accelerometer sensor. In Proceedings of the Joint 8th
International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision
& Pattern Recognition (icIVPR), Spokane, WA, USA, 30 May–2 June 2019; pp. 152–159. [CrossRef]
35. Mobark, M.; Chuprat, S. Recognition of complex human activity using mobile phones: A systematic literature review. J. Theor.
Appl. Inf. Technol. 2018, 96, 3756–3779.
36. Strobl, C.; Boulesteix, A.-L.; Augustin, T. Unbiased split selection for classification trees based on the gini index. Comput. Stat.
Data Anal. 2006, 52, 483–501. [CrossRef]
37. Karthick, G.; Harikumar, R. Comparative Performance Analysis of Naive Bayes and SVM classifier for Oral X-ray images. In
Proceedings of the 4th International Conference on Electronics and Communcation Systems (ICECS), Coimbatore, India, 24–25
February 2017; Volume 17, pp. 88–92.
38. Taunk, K.; De, S.; Verma, S.; Swetapadma, A. A Brief Review of Nearest Neighbor Algorithm for Learning and Classification. In
Proceedings of the 2019 International Conference on Intelligent Computing and Control Systems (ICCS), Madurai, India, 15–17
May 2019; pp. 1255–1260. [CrossRef]
39. Patle, A.; Chouhan, D.S. SVM kernel functions for classification. In Proceedings of the International Conference on Advances in
Technology and Engineering (ICATE), Mumbai, India, 23–25 January 2013. [CrossRef]
40. Ahatsham; Singh, A.; Shahare, V.; Arora, N. An Efficient System for Early Diagnosis of Breast Cancer using Support Vector
Machine. Int. J. Eng. Adv. Technol. (IJEAT) 2019, 9, 7029–7035. [CrossRef]
41. Chen, Y.-P.; Yang, J.-Y.; Liou, S.-N.; Lee, G.-Y.; Wang, J.-S. Online classifier construction algorithm for human activity detection
using a tri-axial accelerometer. Appl. Math. Comput. 2008, 205, 849–860. [CrossRef]
42. Nguyen, D.T.; Li, W.; Ogunbona, P.O. Human detection from images and videos: A survey. Pattern Recognit. 2016, 51, 148–175.
[CrossRef]
43. Khan, A.M.; Lee, Y.-K.; Lee, S.Y.; Kim, T.-S. Human Activity Recognition via an Accelerometer-Enabled-Smartphone Using Kernel
Discriminant Analysis. In Proceedings of the 2010 5th International Conference on Future Information Technology, Busan, Korea,
21–23 May 2010; pp. 1–6.
44. Ahatsham; Jain, S.; Shahare, V.; Arora, N.; Rab, S. Real Time Human Activity Recognition using Smart Phone. Adv. Sci. Eng. Med.
2020, 12, 1200–1203.
45. Maurer, U.; Smailagic, A.; Siewiorek, D.P.; Deisher, M. Activity recognition and monitoring using multiple sensors on different
body positions. In Proceedings of the International Workshop on Wearable and Implantable Body Sensor Networks (BSN’06),
Cambridge, MA, USA, 3–5 April 2006; pp. 113–116.
46. Chen, Y.; Zhong, K.; Zhang, J.; Sun, Q.; Zhao, X. LSTM Networks for Mobile Human Activity Recognition. In Proceedings of the
2016 International Conference on Artificial Intelligence: Technologies and Applications, Bangkok, Thailand, 24–25 January 2016;
pp. 50–53.
47. Rahmani, H.; Mian, A.; Shah, M. Learning a deep model for human action recognition from novel viewpoints. IEEE Trans Pattern
Anal Mach. Intell. 2018, 40, 667–681. [CrossRef] [PubMed]
48. Goutte, C.; Gaussier, E. A Probabilistic Interpretation of Precision, Recall and F-Score, with Implication for Evaluation. Lect. Notes
Comput. Sci. 2005, 3408, 345–359.