PermDroid - A Framework For Android Malware Detection
PermDroid - A Framework For Android Malware Detection
com/scientificreports
The challenge of developing an Android malware detection framework that can identify malware in
real-world apps is difficult for academicians and researchers. The vulnerability lies in the permission
model of Android. Therefore, it has attracted the attention of various researchers to develop an
Android malware detection model using permission or a set of permissions. Academicians and
researchers have used all extracted features in previous studies, resulting in overburdening while
creating malware detection models. But, the effectiveness of the machine learning model depends on
the relevant features, which help in reducing the value of misclassification errors and have excellent
discriminative power. A feature selection framework is proposed in this research paper that helps in
selecting the relevant features. In the first stage of the proposed framework, t-test, and univariate
logistic regression are implemented on our collected feature data set to classify their capacity for
detecting malware. Multivariate linear regression stepwise forward selection and correlation analysis
are implemented in the second stage to evaluate the correctness of the features selected in the first
stage. Furthermore, the resulting features are used as input in the development of malware detection
models using three ensemble methods and a neural network with six different machine-learning
algorithms. The developed models’ performance is compared using two performance parameters:
F-measure and Accuracy. The experiment is performed by using half a million different Android apps.
The empirical findings reveal that malware detection model developed using features selected by
implementing proposed feature selection framework achieved higher detection rate as compared
to the model developed using all extracted features data set. Further, when compared to previously
developed frameworks or methodologies, the experimental results indicates that model developed in
this study achieved an accuracy of 98.8%.
Keywords Android apps, API calls, Neural network, Deep learning, Feature selection, Intrusion detection,
Permissions model
1
Department of Computer Science and applications, D.A.V. University, Sarmastpur, Jalandhar 144012,
India. 2Department of Mathematics, Guru Nanak Dev University, Amritsar, India. 3Department of Nuclear and
Renewable Energy, Ural Federal University Named after the First President of Russia Boris Yeltsin, Ekaterinburg,
Russia 620002. 4Department of Electronics and Communication Engineering, Central University of Jammu,
Jammu 181143, UT of J&K, India. 5School of Electronics and Communication Engineering, Shri Mata Vaishno
Devi University, Katra 182320, UT of J&K, India. 6Department of Applied Data Science, Noroff University College,
Kristiansand, Norway. 7Artificial Intelligence Research Center (AIRC), Ajman University, Ajman, 346, United Arab
Emirates. 8MEU Research Unit, Middle East University, Amman 11831, Jordan. 9Applied Science Research Center,
Applied Science Private University, Amman, Jordan. 10Department of Software, Department of Computer Science
and Engineering, Kongju National University, Cheonan 31080, Korea. *email: [email protected];
[email protected]; [email protected]; [email protected]
Vol.:(0123456789)
www.nature.com/scientificreports/
Now-a-days, smartphones can do the same work as the computer has been doing. By the end of 2023, there will
be around 6.64 billion smartphone users worldwide (https://www.bankmycell.com/blog/how-many-phones-
are-in-the-world). According to the report (https://www.statista.com/statistics/272307/market-share-forecast-
for-smartphone-operating-systems/) at the end of 2023, Android operating systems captured 86.2% of the total
segment. The main reason for its popularity is that its code is written in open source which attracts developers
to develop Android apps on a daily basis. In addition to that it provides many valuable services such as process
management, security configuration, and many more. The free apps that are provided in its official store are the
second factor in its popularity. By the end of March 2023 data (https://w ww.a ppbra in.c om/s tats/n umber-o
f-a ndro
id-apps), Android will have 2.6 billion apps in Google play store.
Nonetheless, the fame of the Android operating system has led to enormous security challenges. On the daily
basis, cyber-criminals invent new malware apps and inject them into the Google Play store (https://play.google.
com/store?hl=en) and third-party app stores. By using these malware-infected apps cyber-criminals steal sensi-
tive information from the user’s phone and use that information for their own benefits. Google has developed
the Google Bouncer (https://krebsonsecurity.com/tag/google-bouncer/) and Google Play Protect (https://www.
android.com/play-protect/) for Android to deal with this unwanted malware, but both have failed to find out
malware-infected apps1–3. According to the report published by Kaspersky Security Network, 6,463,414 mobile
malware had been detected at the end of 2022 (https://s ecure list.c om/i t-t hreat-e volut ion-i n-q
1-2 022-m
obile-s tati
stics/1 06589/). Malware acts as a serious problem for the Android platform because it spreads through these apps.
The challenging issue from the defender’s perspective is how to detect malware and enhance its performance.
A traditional signature-based detection approach detects only the known malware whose definition is already
known to it. Signature-based detection approaches are unable to detect unknown malware due to the limited
amount of signatures present in its database. Hence, the solution is to develop a machine learning-based approach
that dynamically learns the behavior of malware and helps humans in defending against malware attacks and
enhancing mobile security.
Researchers and academicians have proposed different methods for analyzing and detecting malware from
Android. Some of them have been proposed by using static analysis, for example, A NASTASIA4, DREBIN5,
6 7
Droiddetector and D roidDet . On the other side, some researchers have proposed with the help of dynamic
analysis, for example, IntelliDroid8, DroidScribe9, StormDroid10 and M amaDroid11. But, the main constraints of
these approaches are present in its implementation and time consumption because these models are developed
with a number of features. On the other side, academicians and r esearchers3,12–19 also proposed malware detec-
tion frameworks that are developed by using relevant features. But, they have restrictions too. They implemented
only already proposed feature selection techniques in their work.
So, in this research paper, to overcome the hindrances a feature selection framework is proposed. This helps
in the evaluation of appropriate feature sets with the goal of removing redundant features and enhances the
effectiveness of the machine-learning trained model. Further, by selecting a significant features a framework
named PermDroid is developed. The proposed framework is based on the principle of artificial neural network
with six different machine learning techniques, i.e., Gradient descent with momentum (GDM), Gradient descent
method with adaptive learning rate (GDA), Levenberg Marquardt (LM), Quasi-Newton (NM), Gradient descent
(GD), and Deep Neural Network (DNN). These machine learning algorithms are considered on the basis of their
performance in the literature20. In addition to this, three different ensemble techniques with three dissimilar
combination rules are proposed in this research work to develop an effective malware detection framework.
F-measure and Accuracy have been considered as performance parameters to evaluate the performance. From
the literature r eview21–23, it is noticed that a number of authors have concentrated on bettering the functioning
of the malware detection models. However, their study had a key flaw, they only used a small amount of data to
develop and test the model. In order to address this issue, this study report takes into account 500,000 unique
Android apps from various categories.
The method for developing a reliable malware detection model is represented in Fig. 1. The initial collection
of Android application packages (.apk) comes from a variety of promised repositories (mentioned in “Creation
of experimental data set and extraction of features” section). Anti-virus software is used to identify the class of
.apk files at the next level (mentioned in “Creation of experimental data set and extraction of features” section).
Then, features (such as API calls and permissions) are retrieved from the .apk file using various techniques
described in the literature (mentioned in subsection 3.4). Additionally, a feature selection framework is applied
to evaluate the extracted features (discussed in “Proposed feature selection validation method” section). Then,
a model is developed using an artificial neural network using six different machine-learning techniques and
three different ensemble models, employing the selected feature sets as input. Finally, F-measure and Accuracy
are taken into consideration while evaluating the developed models. The following are the novel and distinctive
contributions of this paper:
Vol:.(1234567890)
www.nature.com/scientificreports/
• In this study, to develop efficient malware detection model half a million unique apps have been collected
from different resources. Further, unique features are extracted by performing dynamic analysis in this study.
• The methodology presented in this paper, is based on feature selection methodologies, which contributes in
determining the significant features that are utilized to develop malware detection models.
• In this study, we proposed three different ensemble techniques that are based on the principle of a heteroge-
neous approach.
• Six different machine learning algorithms that are based on the principle of Artificial Neural Network (ANN)
are trained by using relevant features.
• When compared to previously developed frameworks and different anti-virus software in the market, the
proposed Android malware detection framework can detect malware-infected apps in less time.
• A cost-benefit analysis shows that the proposed Android malware detection framework is more effective in
identifying malware-infected apps from the real world.
The remaining sections of this research paper are arranged as follows: “Related work” section presents the litera-
ture survey on Android malware detection as well as the creation of research questions. “Research methodology”
section gives an overview of the research methodology used to create the Android malware detection framework.
Different machine learning and ensemble techniques are addressed in “Machine learning technique” section. The
proposed feature selection validation technique is discussed in “Proposed feature selection validation method”
section. The experimental results are presented in “Experimental setup and results” section. Threats to validity
are presented in “Threats to validity” section. Conclusion and the future scope are discussed in “Conclusion
and future work” section.
Related work
The exploitation of the vulnerability is common these days to acquire higher privilege on Android platforms.
Since 2008, cybercriminals have started targeting Android devices. An exploit app, from the perspective of
Android security, can assist cyber-criminals in bypassing security mechanisms and gaining more access to users’
devices. Cybercriminals may exploit user data by selling their personal information for monetary gain if they
took advantage of these privileges. The detection process, which has been used by researchers in the past and is
based on Artificial Neural Networks (ANN) and feature selection techniques, is addressed in this subsection.
Androguard (https://c ode.g oogle.c om/a rchiv e/p/a ndrog uard/) is a static analysis tool that detects malware on
Android devices using the signature concept. Only malware that is already known to be present and whose defini-
tion is in the Androguard database is identified. It cannot, however, identify unidentified malware. Andromaly23,
is developed on a dynamic analysis tool that uses a machine learning technique. It monitored CPU utilization,
data transfer, the number of effective processes, and battery usage in real-time. The test was carried out on a
few different types of simulated malware samples, but not on the applications that are present in the real-world.
By using the semantics of the code in the form of code graphs collected from Android apps, Badhani et al.24
developed malware detection methodology. Faruki et al.21 introduced AndroSimilar, which is based on the
principles of generated signatures that are developed from the extracted features, which are used to develop
malware detection model.
Aurasium25 takes control of an app’s execution by examining arbitrary security rules in real-time. It repackages
Android apps with security policy codes and informs users of any privacy breaches. Aurasium has the problem
of not being able to detect malicious behavior if an app’s signature changes. They performed dynamic analysis
of Android apps and considered call-centric as a feature. The authors tested their method on over 2900 Android
malware samples and found that it is effective at detecting malware activity. A web-based malware evaluation
method has been proposed by A ndrubis26, it operates on the premise that users can submit apps via a web service,
and after examining their activity, it returns information on whether the app is benign or malicious. Ikram et al.27
suggested an approach named as DaDiDroid based on weighted directed graphs of API calls to detect malware-
infected apps. The experiment was carried out with 43,262 benign and 20,431 malware-infected apps, achieving
a 91% accuracy rate. Shen et al.28 developed an Android malware detection technique based on the information
flow analysis principle. They implement N-gram analysis to determine common and unique behavioral patterns
present in the complex flow. The experiment was carried out on 8,598 different Android apps with an accuracy
of 82.0 percent. Yang et al.29 proposed an approach named EnMobile that is based on the principle of entity
characterization of the behavior of the Android app. The experiment was carried out on 6,614 different Android
apps, and the empirical results show that their proposed approach outperformed four state-of-the-art approaches,
namely Drebin, Apposcopy, AppContext, and MUDFLOW, in terms of recall and precision.
CrowDroid34, which is built using a behavior-based malware detection method, comprises of two components:
a remote server and a crowdsourcing app that must both be installed on users’ mobile devices. CrowDroid uses a
crowdsourcing app to send behavioral data to a remote server in the form of a log file. Further, they implemented
2-mean clustering approach to identify that the app belongs to malicious or benign class. But, the crowDroid app
constantly depletes the device’s resources. Yuan et al.52 proposed a machine learning approach named Droid-Sec
that used 200 extracted static and dynamic features for developing the Android malware detection model. The
empirical result suggests that the model built by using the deep learning technique achieved a 96% accuracy rate.
TaintDroid30 tracks privacy-sensitive data leakage in Android apps from third-party developers. Every time any
sensitive data leaves the smartphone, TaintDroid records the label of the data, the app that linked with the data,
as well as the data’s destination address.
Zhang et al.53 proposed a malware detection technique based on the weighted contextual API dependency
graph principle. An experiment was performed on 13500 benign samples and 2200 malware samples and achieved
an acceptable false-positive rate of 5.15% for a vetting purpose.
Vol.:(0123456789)
www.nature.com/scientificreports/
AndroTaint54 works on the principle of dynamic analysis. The features extracted were used to classify the
Android app as dangerous, harmful, benign, or aggressive using a novel unsupervised and supervised anomaly
detection method. Researchers have used numerous classification methods in the past, like Random forest55,
J4855, Simple logistic55, Naïve Bayes55, Support Vector Machine56,57, K-star55, Decision tree23, Logistic regression23
and k-means23 to identify Android malware with a better percentage of accuracy. D roidDetector6, Droid-Sec52,
and Deep4MalDroid58 work on the convention of deep learning for identifying Android malware. Table 1 sum-
marizes some of the existing malware detection frameworks for Android.
The artificial neural network (ANN) technique is used to identify malware on Android devices
Nix and Z hang59 developed a deep learning algorithm by using a convolution neural network (CNN) and used
API calls as a feature. They utilized the principle of Long Short-Term Memory (LSTM) and joined knowledge
from its sequences. McLaughlin et al.60, implemented deep learning by using CNN and considered raw opcode
as a feature to identify malware from real-world Android apps. Recently, researchers6,58 used network param-
eters to identify malware-infected apps. Nauman et al.61, implemented connected, recurrent, and convolutional
neural networks, and they also implemented DBN (Deep Belief Networks) to identify malware-infected apps
from Android. Xiao et al.62, presented a technique that was based on the back-propagation of the neural net-
works on Markov chains and considered the system calls as a feature. They consider the system call sequence as
a homogenous stationary Markov chain and employed a neural network to detect malware-infected apps. Mar-
tinelli et al.63, implemented a deep learning algorithm using CNN and consider the system call as a feature. They
performed an experiment on a collection of 7100 real-world Android apps and identify that 3000 apps belong
to distinct malware families. Xiao et al.64, suggested an approach that depends on the principle of LSTM (Long
Short-Term Memory) and considers the system call sequence as a feature. They trained two LSTM models by the
system call sequences for both the benign and malware apps and then compute the similarity score. Dimjas̈evic
et al.65, evaluate several techniques for detecting malware apps at the repository level. The techniques worked
on the tracking of system calls at the time the app is running in a sandbox environment. They performed an
experiment on 12,000 apps and able to identify 96% malware-infected apps.
Vol:.(1234567890)
www.nature.com/scientificreports/
Research questions
To identify malware-infected apps and considering the gaps that are present in the literature following research
questions are addressed in this research work:
• RQ1 Does the filtering approach helps to identify that whether an app is a benign or malware-infected (first
phase of the proposed feature selection framework)? To determine the statistical significance among mali-
cious and benign apps, the t-test is used. After, determining significant features, a binary ULR investigation is
applied to select more appropriate features. For analysis, all the thirty different feature data sets are assigned
(shown in Table 5) as null hypotheses.
• RQ2 Do already existing and presented work’s sets of features show an immense correlation with each other?
To answer this question, both positive and negative correlations are examined to analyze the sets of features,
which help in improving the detection rate.
• RQ3 Can the identified features assist in determining whether the app is malware-infected or not? The pri-
mary objective of this question is to use the feature selection framework validation approach to determine
the appropriate features. In this paper, four stages (i.e., ULR, t-test, Correlation analysis, and multivariate
linear regression stepwise forward selection) are implemented to identify the appropriate features, that helps
in identifying whether an app contains malicious behavior or not.
• RQ4 Which classification algorithm among the implemented machine learning algorithms is most appropriate
for identifying malware-infected apps? To answer to this question the efficiency of various machine learn-
ing approaches are evaluated. In this study, three different ensemble approaches and six different machine
learning algorithms based on neural networks are considered.
• RQ5 Is the feature collected (such as an app’s rating, API calls, permissions, and the number of people who
have downloaded the app) sufficient for identifying a malicious app or not? This question helps in determin-
ing whether or not considering features can detect malware-infected apps in the real world. To answer this
question, the performance of our suggested model is compared with previously published frameworks as
well as several anti-virus scanners in the market.
Research methodology
Based on the research questions mentioned above, the methodology that is used in this research paper is men-
tioned in the following subsections. In order to improve the detection rate for malware, the obtained data set
has been normalized, and dependent and independent variables have been selected.
Independent variables
In this study, the model is developed by applying the proposed feature selection approach, which helps in the
detection of malware-infected apps. Additionally, as shown in Fig. 2, five different strategies to select the best
Vol.:(0123456789)
www.nature.com/scientificreports/
features are used. The best features are selected from other accessible features created on intermediate explore
models at each level.
Dependent variables
The focus of this research is to find a link between Android apps and the features (such as app rating, API calls,
permission, and the number of users who have downloaded an app) retrieved from the collected data set. The
malware app characteristics are separated from the benign app features in the dependent variable of Android
apps.
Vol:.(1234567890)
www.nature.com/scientificreports/
Author/approach The technique for selecting features was used Machine learning algorithm used
ANASTASIA4 Randomized tree group Decision tree (J48), Support vector machine (SVM),
(i.e., Extra trees-classifier) Naïve Bayes (NB), Logistic regression,
K-Nearest neighbours, random forest(RF),
Deep learning, and Adaboost
Andromaly23 Chi-square, Fisher score and Information gain k-Means, Naïve Bayes (NB),
Bayesian network, decision tree (J48)
Histogram or logistic regression
Mas’ ud et al.66 Information gain and Chi-square test Naïve Bayes (NB), K-nearest Neighbour (KNN),
Decision Tree (J48), Multi-layer perceptron (MLP),
and random forest (RF)
Allix et al.74 Information gain Support vector machine (SVM), C4.5,
RIPPER, and Random forest
Yerima et al.67 Mutual information Bayesian classification
MKLDroid75 Chi-squared Kernel methods
Azmoodeh et al.76 Information gain Deep Eigenspace learning approach
Chen et al.77 Using manual pruning while gaining information Random forest (RF), support vector machine (SVM), )
and K-nearest neighbor (KNN)
Narudin et al.68 ClassifierSubsetEval Random forest, Multi-layer perceptron,
J48, K-Nearest neighbours, and Bayes network
Yerima et al.78 Information gain Bayesian classifier
Table 2. In the literature, there are feature selection methods and machine learning algorithms that have been
implemented.
Vol.:(0123456789)
www.nature.com/scientificreports/
n
1 ′i
Mean square error = (O − Oi )2 . (2)
n
i=1
′
Here, O is the actual output value, and O is the desired output value. Various methods were proposed by
researchers20,84 to train the neural network. In this research work, six different kinds of machine learning algo-
rithms (namely, Gradient Descent approach, Quasi-Newton approach, Gradient Descent with Momentum
approach, Levenber-Marquardt approach, Gradient Descent with Adaptive learning rate approach, and Deep
neural network) are considered to develop malware detection model. These models are effective in the field of
software fault prediction20, intrusion detection and desktop malware predictions85 too.
Vol:.(1234567890)
www.nature.com/scientificreports/
Quasi‑Newton approach
In order to compute the total error function, this approach requires the evaluation of the second order deriva-
tives for each component of the gradient v ector20,84. The iterative scheme for the Weight vector X is given as:
∂
Xk+1 = Xk − Hk−1 (Ek ), (7)
∂X
where Xk and Xk+1 are the current and updated weight vectors, accordingly. H is the Hessian matrix given by
2
∂2E 2
∂ E
∂X 2 ∂X1 X2
· · · ∂X∂1 XEN
21
∂2E ∂2E
∂ E
∂X1 X2 ∂X22 · · · ∂X2 XN
H =
. .. .. ..
.. . . .
2
∂ E ∂2E ∂2E
∂X1 XN ∂X2 XN · · · ∂X 2
N
• Homogenous ensemble approach: In this approach, all classification models, are of the same kinds, but the
difference is in generating the training set.
• Heterogenous ensemble approach: Here, all base classification approaches are of distinct types.
On the basis of combination rules, ensemble approaches are divided into two distinct categories:
• Linear ensemble approach: While developing the model, with a linear ensemble approach an arbitrator com-
bines the results that come from the base learners, i.e., selection of classification approach, average weighted,
etc.
Vol.:(0123456789)
www.nature.com/scientificreports/
• Nonlinear ensemble approach: While developing the model, with the nonlinear ensemble approach, it fed the
result of the base classifier, which is a nonlinear malware detection model for example Decision tree (DT),
Neural network (NN), etc.
In this work, a heterogenous ensemble approach having three distinct combination rules is adapted. The ensemble
techniques are detailed in Table 6.
Vol:.(1234567890)
www.nature.com/scientificreports/
• False positive (FP) A false positive occurs when the developed model identifies the positive class incorrectly.
• False negative (FN) When the developed model successfully identifies the negative class, a false negative
occurs.
• True negative (TN) An accurate identification of the negative class by the developed model represents a true
negative conclusion.
• True positives (TP) An accurate identification of the positive class by the developed model represents a real
positive conclusion.
• Recall The data set’s positive classes that are made up of all other positive classes are identified.
x
Recall = , (9)
x+z
where x = NMalware→Malware , z = NMalware→Benign
• Precision The accuracy measures the proportion of forecasts in the positive class that are indeed in the posi-
tive class.
x
Precision = . (10)
x+y
where y = NBenign→Malware
Accuracy Accuracy is measured a s3:
x+w
Accuracy = , (11)
Nclasses
where Nclasses = x + y + z + w,
w = NBenign→Benign
F-measure F-measure is measured a s3:
Vol.:(0123456789)
www.nature.com/scientificreports/
2 ∗ Recall ∗ Precision
F − measure =
Recall + Precision
2∗x (12)
=
2∗x+y+z
Vol:.(1234567890)
www.nature.com/scientificreports/
Figure 2 demonstrates the phases of the proposed feature selection validation framework. Without using
machine learning algorithms, this framework aims to determine whether the selected features are useful in
detecting malicious apps. The wrapper strategy is used to pick the sets of features that are useful in identifying
malware apps after all crucial components have been examined. It keeps track of the progress of the learning
algorithm that was used to identify each feature subset. In this work, the selected features are investigated using
linear discriminant analysis (LDA).
i. Data set Table 3 summarized the data set used in this research work. The considered data set belongs to
141 different malware families.
ii. Normalization of data By using the Min-max normalizing approach, all features are normalized between
the ranges of 0 and 1.
iii. Partition of data We examined at the data set that wasn’t used for training in order to evaluate the pro-
posed feature selection approach. Further, the data set is divided into two different parts one part is used
for training, and the remaining is used for testing. The group ratios in the training and testing of the data
sets are nearly identical.
iv. Filter approach Pre-processing is the term that describes this technique because it eliminates extraneous
features. In this step, the t-test and ULR analysis are implemented.
a. t-test analysis It examine the statistical significance of benign and malware apps using the t-test method.
In a 2-class problem (malware apps and benign apps), analysis of the null hypothesis (H0) significant
that the two populations are not equal, or it is seen that there is a noticeable variance among their mean
values and features used by both of them are d ifferent95. Furthermore, it shows that the features affect
the malware detection result. Hence, those features are considered, which have significant differences
in their mean values, and others are excluded. Hence, it is essential to approve the null hypothesis (i.e.,
H0) and discard the alternative ones95. t-test is implemented on each of the attributes and then P value
for each feature is calculated, which indicates how well it distinguishes the group of apps. According to
research by95, features with an P value of < 0.05 show significant biases.
b. Univariate logistic regression (ULR) analysis After identifying features that make a significant difference
between malware and benign apps, binary ULR analysis is implemented to test the correlation among
features that helps in malware detection95. ULR analysis is implemented on each selected feature set,
which helps in discovering whether the above-selected features were essential to detect the malware-
infected apps or not. Only those features are considered, which are having P value < 0.05. From the
results of the ULR analysis and t-test, the hypothesis are rejected and accepted mentioned in Table 5.
v. Wrapper approach To determine optimum sets of the feature, cross-correlation analysis and multivariate
linear regression stepwise forward selection is implemented in this stage.
a. Cross correlation analysis After finding the important features, the correlation analysis is implemented
and then examination for both negative and positive correlation coefficients (i.e., r-value) between
features is performed. If a feature has a value of r > = 0.7 or r-value < =0.7 with other features, i.e., have
a higher correlation then the performance of these features is studied separately. Further, those features
are selected, which perform better.
b. Multivariate linear regression stepwise forward selection It is not imply that, features that are achieved
are relevant to develop malware detection framework. In this stage, ten-fold cross-validation technique
is applied to determine the significant features.
vi. Performance evaluation Further, to validate that proposed framework is able to identify malware-infected
apps that were developed by implementing the steps mentioned above by using independent test data.
Additionally, the efficiency of the essential feature sets used for malware detection is validated. On thirty
different categories of Android apps, nine different machine learning classifiers were used to develop the
investigation model. To evaluate the framework two separate performance parameters, are considered i.e.,
F-measure and Accuracy. The effectiveness of our detection model is then evaluated using the proposed
malware detection methodology.
a. Comparison with previously used classifiers Parameters like Accuracy and F-measure are compared with
existing classifiers proposed by researchers in the literature to see if our suggested model is feasible or not.
b. Comparison with AV scanners To compare the effectiveness of our suggested work, ten different anti-virus
scanners are considered and their performance is evaluated on the collected data set.
c. Detection of unknown and known malware families The proposed framework is also examined to see whether
it can identify known and unknown malware families.
Vol.:(0123456789)
www.nature.com/scientificreports/
Vol:.(1234567890)
www.nature.com/scientificreports/
Figure 5. Deep learning neural network (DNN) method constructed with DBN.
Following are the phases that are pursued in this study, to develop an effective and efficient malware detec-
tion framework. The proposed feature selection framework is applied to all the extracted feature data sets, to
select significant features. After that, six different machine learning algorithms based on the principle of neural
network and three different ensemble algorithms are considered to develop a malware detection model. So, in
this study, a total of 540 (30 different Android apps data sets * 9 different machine learning techniques * (one
Vol.:(0123456789)
www.nature.com/scientificreports/
Benign Malware
Benign Benign-> Benign (TP) Benign-> Malware (FP)
Malware Malware-> Benign (FN) Malware-> Malware (TN)
Table 7. An Android app’s maliciousness can be determined using a confusion matrix.
takes into account all extracted features, and another takes into account features identified using the suggested
feature selection framework. )) different detection models are developed. The following are a detailed description
of the model followed in this study:
1. Thirty different extracted feature data sets are used to implement the proposed feature selection framework.
2. The first stage, which involved identifying significant features, was employed as an input to train the model
using various classification and ensemble machine learning approaches. In this research paper, ten-fold
cross-validation technique is implemented to verify the develop m odel16. Further, outliers are eliminated,
which effect the performance of the proposed framework. The performance of outliers is measured using
the equation below:
if |zji − zˆj | > 3 ∗ σ for Effective outliers,
ei =
if |zji − zˆj | ≤ 3 ∗ σ for Non Effective outliers (13)
3. The developed model using the aforementioned two processes is evaluated using the collected data set in
order to determine whether or not the proposed framework is successful in identifying malicious apps.
t‑Test analysis
t-test analysis is used to determine the statistical significance of detecting the malware from Android apps. In
this work, t-test is applied on extracted feature sets and calculated its P value. Further, in this study, the cut-off P
value considered is 0.05, i.e., it denotes that feature sets that have P value < 0.05 has a strong prediction capabil-
ity. Figure 7 illustrates the findings of a t-test performed on the thirty various categories of Android apps that
comprise up our obtained data set. The P value is provided using two forms for simplicity of use (box with black
Vol:.(1234567890)
www.nature.com/scientificreports/
circle (·) means P value < 0.05 and blank box means P value > than 0.05). The sets of features with emphasis
P values of < 0.05 have a significant impact on identifying malicious or benign apps. Figure 7 shows how the
S29, S27, S25, S23, S22, S21, S19, S18, S13, S10, S8, S5, S3, and S1 feature sets might help to detect malicious and
benign apps in the Arcade and Action categories. As a result, in this study, we rule out the hypotheses H1, H3,
H5, H8, H10, H13, H18, H19, H21, H22, H23, H25, H27, and H29, coming to the conclusion that these sets of
features are capable of identifying apps in the Arcade and Action category that are malicious or benign.
To understand the relationship between malware and benign apps, we have drawn an error box-plot dia-
gram. These box-plot diagrams verify the outcomes of the t-test analysis. If there is no overlapping in means and
their confidence intervals (CI), then it means there will be a statistical difference between malware and benign
apps else. There is no significant difference between them. An error box-plot of the 95% confidence intervals
throughout the sets of features and the mean for Arcade and Action category apps is demonstrated in Fig. 8. The
outcomes of other categories of Android apps are of similar types. Based on Fig. 8, we can observe that the boxes
of S29, S27, S25, S23, S22, S21, S19, S18, S13, S10, S8, S5, S3, and S1 sets of feature do not overlap which means
they are significantly different from each other. The mean value of the malware group is higher than the benign
group apps. Based on error box-plots, we consider the hypotheses H1, H3, H5, H8, H10, H13, H18, H19, H21,
H22, H23, H25, H27 and H29 concluding that these feature sets can able to identify the malware-infected apps
for Arcade and Action category Android apps.
ULR analysis
To examine whether the selected sets of feature after implementing t-test analysis are significant to identify
malware apps or not, in this study, ULR analysis is performed on selected sets of features. A set of features is
considerably associated with malware detection if its P value is < 0.05. In every task, some sets of features are
essential for the evolution of the malware detection model, while different sets of features do not seem to be
appropriate for malware detection. The outcomes of the ULR approach are demonstrated in Fig. 9. Equivalent
to t-test analysis, the same representation is used as such in P values, i.e., blank box means P value > 0.05 and
box having black square has P value ≤ to 0.05.
From Fig. 9, it is clear that among thirty different categories of features, only S5, S3, S1, S13, S10, S23, S19,
S29, and S25 sets of features are significant detectors of malware apps. As a result, we reject null hypotheses H1,
H3, H5, H10, H13, H19, H23, H25, and H29 and conclude that these sets of features are directly related to the
functioning of the apps. After implementing t-test and ULR analysis on our collected sets of features, rejection
and acceptance of the hypotheses is done that is presented in the Table 5. Figure 10 demonstrates the rejection
and acceptance of the hypotheses for all of the thirty different categories of Android apps. The horizontal and
vertical axes indicate the name of the hypothesis and the equivalent category of the Android app, accordingly.
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 8. Error box-plots for all the set of permissions in Arcade and Action category apps.
To represent the rejection and acceptance of the hypotheses, the cross symbol (×) and black circle (·), are used
respectively. Based on Fig. 10, it is observed that only sixteen hypotheses out of thirty are accepted. Others are
rejected for Arcade and Action category Android apps.
Vol:.(1234567890)
www.nature.com/scientificreports/
Figure 10. Hypothesis.
Vol.:(0123456789)
www.nature.com/scientificreports/
Figure 11. Correlation between set of features (here LT stands for lower triangle and UT stands for Upper
triangle.
Vol:.(1234567890)
www.nature.com/scientificreports/
Figure 13. Features selected after implementing multivariate linear regression stepwise forward selection.
Vol.:(0123456789)
www.nature.com/scientificreports/
significant set of features. A set of features that were taken into account in this paper while building a malware
detection model is represented by a black circle with the symbol (·).
• Empty circle symbol: Features are relevant after implementing t-test analysis.
• Triangle symbol: Features are relevant after implementing ULR analysis and t-test.
• Diamond symbol: Features are relevant after applied cross-correlation analysis, ULR, and t-test.
• Filled circle symbol: Features are relevant after implementing multivariate linear regression stepwise forward
selection method, cross-correlation analysis, ULR, and t-test.
Vol:.(1234567890)
www.nature.com/scientificreports/
performance metrics for various categories of Android apps are shown in Tables 8 and 9. From Tables 8 and 9,
the following conclusions can be drawn:
• The model developed by features selected using proposed framework (Model also developed by using distinct
feature selection approaches are shown in Tables S1 to S14 in “Online Appendix A”) as an input produces
Vol.:(0123456789)
www.nature.com/scientificreports/
better results when compared to a model constructed by taking into account all sets of features, presenting
a significant value of F-measure and Accuracy for identifying malware.
• In compared to the others, the neural network with Deep Neural Network (DNN) training method yields
higher outcomes.
Figures 16 and 17 show the Accuracy and F-measure box-plot diagrams for each model built using classification
methods. Each figure has two box plots, one containing all of the extracted features (EF) and the other contain-
ing only selected feature sets (SF).
The Box-plot diagram assists us in analyzing the performance of all the implemented approaches based on a
single diagram. The line drawn in the middle of each box-plot diagram, i.e. the median, is used to determine its
value. If a model’s median value is high, it’s regarded as the best model for detecting malware. It can be inferred
from Figs. 16 and 17 that:
• The models developed utilizing a significant set of features have high median values. The box-plot diagrams
in Figs. 16 and 17 show that SF outperformed all extracted features in terms of detecting Android malware.
• The DNN-based model yields the best results out of all the machine learning techniques for classification
that have been used.
Vol:.(1234567890)
www.nature.com/scientificreports/
Further, it is also noticed that ensemble approaches detect more malware as compared to other implemented
machine learning algorithms except DNN.
Vol.:(0123456789)
www.nature.com/scientificreports/
exists between the two procedures. Table 10a shows that the P value is < 0.0013, indicating that there is a signifi-
cant difference between the applied processes; out of 36 pairs of training techniques, 22 are offered as a significant
outcome. By examining the mean difference value in Table 10a, it can be seen that the DNN method outper-
formed the performance of other machine learning techniques. In addition, the value of the mean difference of
ensemble techniques is better when compared to other models, with the exception of the model built using DNN.
On the basis of all selected sets of feature using proposed framework and extracted features
By taking into consideration each set of features, a total of 270 different data points ((3 ensemble techniques +
neural network with six machine learning techniques) * 30 types of Android apps) are developed in this study
(one for each performance measure). Wilcoxon signed-rank test performance was described in Table 10b. It is
seen from Table 10b that there is a significant difference between the models developed because the P value is
less than 0.05. Additionally, it is evident that the features taken into account employing the feature selection
framework outperformed the model developed by using all extracted feature sets when comparing the mean
difference values from Table 10b to it.
Vol:.(1234567890)
Scientific Reports |
(2024) 14:10724 |
Accuracy F-measure
www.nature.com/scientificreports/
GD GDM GDX NM LM DNN BTE MVE NDTF GD GDM GDX NM LM DNN BTE MVE NDTF
(a) Training methods
P value
GD 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.001 0.000 0.000 0.000 0.000 0.000 0.000 0.000
GDM 0.425 0.000 0.000 0.000 0.000 0.000 0.000 0.300 0.000 0.000 0.000 0.000 0.000 0.000
GDX 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
NM 0.000 0.410 0.0220 0.0710 0.050 0.081 0.050 0.0701 0.110 0.107
LM 0.000 0.0120 0.0510 0.020 0.040 0.0601 0.100 0.170
DNN 0.0020 0.0010 0.030 0.0001 0.000 0.450
BTE 0.0040 0.010 0.000 0.020
MVE 0.010 0.010
NDTF
Mean
GD 0.000 1.508 2.456 − 2.040 − 3.010 − 3.500 − 2.200 − 2.090 − 2.908 0.000 0.001 0.018 − 0.050 − 0.020 − 0.060 − 0.027 − 0.0280 − 0.030
GDM − 1.006 0.000 1.130 − 4.425 − 4.080 − 5.080 − 3.800 − 4.480 − 4.600 − 0.020 0.000 0.005 − 0.030 − 0.060 − 0.080 − 0.040 − 0.045 − 0.050
https://doi.org/10.1038/s41598-024-60982-y
GDX − 2.788 − 1.33 0.000 − 4.000 − 4.213 − 6.170 − 3.120 − 3.780 − 4.230 − 0.011 − 0.002 0.000 − 0.021 − 0.060 − 0.080 − 0.050 − 0.070 − 0.078
NM 2.890 3.899 4.898 0.000 0.890 − 2.410 − 0.620 − 0.0710 − 0.810 0.021 0.034 0.039 0.000 − 0.004 − 0.006 − 0.007 − 0.008 − 0.0015
LM 3.477 5.025 6.671 1.333 0.000 − 1.322 − 0.233 − 0.851 − 0.880 0.020 0.310 0.040 0.300 0.000 − 0.006 − 0.007 − 0.0078 − 0.012
DNN 4.311 4.220 3.780 2.0981 5.678 0000 1.200 2.180 1.910 0.060 0.058 0.052 0.038 0.041 0.000 − 0.001 − 0.003 − 0.005
BTE 2.997 2.633 2.431 2.100 1.890 − 1.988 0.000 − 0.0540 − 0.679 0.028 0.041 0.045 0.007 0.009 0.001 0.000 − 0.002 − 0.020
MVE 2.882 4.488 5.672 0.889 0.998 − 0.533 0.560 0.000 − 0.054 0.028 0.042 0.046 0.002 0.008 0.001 0.002 0.000 − 0.008
NDTF 2.944 4.552 5.661 0.991 0.789 − 0.450 0.646 0.054 0.000 0.036 0.050 0.050 0.015 0.009 0.007 0.008 0.007 0.000
Accuracy F-measure
Mean P value Mean P value
(b) All EF and SF
EF SF EF SF EF SF EF SF
EF 0.00 − 5.76 0.00 0.00 − 0.048 0.00
SF 5.76 0.00 0.048 0.000
Table 10. To examine the rank test findings, the Wilcoxon signed technique was applied.
27
Vol.:(0123456789)
www.nature.com/scientificreports/
features)* 30 data sets). Figure 18 illustrates both the classifiers employed in this study and the most frequently
used classifiers in the literature.
On the basis of Fig. 18, it can be seen that the model produced using neural networks has a higher median
value and achieves better results than the model developed using the literature’s used classifiers. Further, to decide
that, which model produces better results, a pairwise Wilcoxon signed rank test is implemented. Table 11 summa-
rizes the results of the Wilcoxon test with Bonferroni correction examination for accuracy outcomes. Further, the
Table 11 is divided into two sections, the first of which indicates the P value and the second of which demonstrates
the mean difference between different pairs of classifiers. We implemented thirteen different machine learning
approaches in this research paper (4 previously applied classifier in the literature + 9 implemented classifier in
this study); thus, an aggregate of seventy eight (78) individual pairs are possible 13techniques C2 = 78, and all clas-
sifier outcomes are examined at the 0.05 significance level. Only those null hypotheses with an P value is less
than 0.05/78 = 0.000641 are rejected in this study. Table 11 shows that there is a significant difference between
different implemented classifier approaches in a number of cases when the P value is less than 0.000641, i.e.,
66 out of 78 pairs of classification approaches have significant outcomes. Table 11 demonstrates that the DNN
approach outperforms other machine learning classifiers in terms of mean difference value.
Vol:.(1234567890)
Scientific Reports |
(2024) 14:10724 |
www.nature.com/scientificreports/
Accuracy
P value Mean
LOGR NBC SVM DT GD GDM GDX NM LM DNN BTE MVE NDTF LOGR NBC SVM DT GD GDM GDX NM LM DNN BTE MVE NDTF
https://doi.org/10.1038/s41598-024-60982-y
LOGR 0.000 0.000 0.160 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 2.010 3.940 − 1.30 − 9.087 − 7.890 − 6.67 − 11.890 − 12.880 − 15.880 − 11.77 − 12.99 − 13.00
NBC 0.000 0.000 0.0100 0.670 0.642 0.000 0.000 0.000 0.000 0.000 0.000 8.77 0.000 12.22 7.77 − 1.08 − 1.66 0.08 1.77 − 3.89 − 6.00 − 3.88 − 4.88 − 5.89
SVM 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 − 3.89 − 12.89 0.000 − 5.88 − 13.88 − 11.55 − 10.88 − 15.77 − 13.88 − 17.88 − 15.77 − 16.99 − 17.00
DT 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 1.88 − 7.88 5.88 0.000 − 7.88 − 6.77 − 7.09 − 5.66 − 5.00 − 11.88 − 8.99 − 9.99 − 10.89
GD 0.000 0.000 0.000 0.000 0.000 0.000 0.000 9.88 8.99 12.88 11.20 0.000 1.678 2.77 − 2.06 − 3.88 − 6.05 − 4.55 − 5.08 − 5.89
GDM 0.660 0.000 0.000 0.000 0.000 0.000 0.000 7.88 5.00 6.77 1.99 2.89 0.000 1.110 − 0.89 − 2.88 − 7.88 − 6.77 − 5.77 − 6.990
GDX 0.000 0.000 0.000 0.000 0.000 0.000 6.77 7.08 6.88 7.88 1.99 2.001 0.000 − 1.88 − 2.88 − 8.99 − 6.88 − 7.88 − 8.09
NM 0.000 0.041 0.0020 0.000 0.000 11.99 10.99 15.22 16.77 2.33 1.550 1.7701 0.000 − 1.107 − 11.22 − 8.99 − 7.99 − 10.99
LM 0.000 0.000 0.0080 0.000 0.000 13.79 11.79 14.82 18.67 1.93 1.950 2.701 1.990 0.000 − 13.22 − 7.99 − 9.69 − 12.99
DNN 0.008 0.000 0.004 18.79 17.90 13.33 16.07 6.93 5.050 5.501 4.090 3.780 0.000 1.99 1.69 1.09
BTE 0.003 0.004 12.90 14.90 12.03 15.97 5.73 4.050 3.701 3.801 0.780 − 1.003 0.000 − 0.69 − 0.99
MVE 0.003 10.90 12.890 10.403 13.097 2.73 2.950 2.701 2.201 1.680 − 3.003 − 1.088 0.000 − 1.999
NDTF 17.90 16.890 12.403 15.097 3.631 3.850 2.491 2.881 1.980 − 1.003 0.078 0.070 0.000
Table 11. Wilcoxon signed rank test analysis is implemented to the previously used classifier.
29
Vol.:(0123456789)
www.nature.com/scientificreports/
Comparison of results based on the amount of time it takes to identify malware in real‑world apps
In this section of the article, the performance of PermDroid is compared in terms of the time needed to iden-
tify malware in real-world apps. For this experiment, we download the data set from two different repositories
Drebin (https://www.sec.cs.tu-bs.de/~danarp/drebin/download.html) and AMD (http://amd.arguslab.org/) and
experimented by implementing the individual frameworks. Table 14 shows that, when compared to the individual
frameworks available in the literature, our suggested technique can identify malware in less time.
Comparison of the results on the basis of detection rate with different approaches or frameworks available in the
literature
Furthermore, proposed malware detection model (i.e., PermDroid) is compared to previously developed tech-
niques or frameworks present in the literature. The names, methodology, deployment, purpose, data collection,
Vol:.(1234567890)
www.nature.com/scientificreports/
Frameworks TPR FPR Average time in identifying app is malicious or not (Sec)
MADAM (2012)38 0.88 0.6 1200
DroidScope (2012)39 0.89 0.6 1280
AppGuard (2012)40 0.87 0.7 1100
TstructDroid (2013)41 0.88 0.7 1200
AppsPlayground (2013)42 0.88 0.7 1100
AppProfiler (2013)43 0.89 0.8 1000
Andrubis (2014)26 0.88 0.8 980
Androguard (2015)44 0.88 0.7 1100
CopperDroid (2015)45 0.78 0.7 1300
DroidDetector (2016)6 0.80 0.7 1000
MAMADROID (2016)11 0.82 0.6 800
DroidSieve (2017)46 0.88 0.7 920
PIndroid (2017)47 0.89 0.8 810
MOCDroid (2017)48 0.88 0.5 500
DroidDet (2018)7 0.88 0.32 430
MalDozer (2018)49 0.90 0.3 320
Enmobile (2018)29 0.88 0.7 380
SeqDroid (2019)50 0.92 0.2 290
MaMaDroid (2019)97 0.93 0.2 300
DaDiDroid (2019)27 0.91 0.6 330
DeepDroid (2019)13 0.91 0.6 330
DL-Droid (2020)51 0.93 0.19 200
PerbDroid (2020)15 0.91 0.6 330
Proposed approach (i.e., PermDroid) 0.982 0.1 100
Table 14. Compare PermDroid’s performance to earlier frameworks that have been developed. Averaged time
is calculated by taking training and testing time-period and using Drebin data set.
and detection rate of proposed methodologies or frameworks are listed in Table 15. Empirical result revealed
that our proposed framework produced a 3 percent greater detection rate. Experiment was performed by using
Drebin data set (https://www.sec.cs.tu-bs.de/~danarp/drebin/download.html).
Vol.:(0123456789)
www.nature.com/scientificreports/
Framework/approach Goal Methodology Deployment Data set used while developing Detection rate Availability
Paranoid Android31 (2010) Detection Behavioural and Dynamic Off-device Limited – –
Crowdroid34 (2011) Detection Dynamic, Distributed Very-Limited High –
System call/API and Behavioural
Aurasium25(2012) Detection Dynamic and Behavioural Off-device Limited High Free
Andromaly23 (2012) Detection Dynamic and Profile-based Distributed Very-Limited High Free
AndroSimilar21(2013) Detection Static Off-device Limited Lesser –
TaintDroid30 (2014) Detection Dynamic Off-Device Very-Limited Lesser Free
System call/API and Behavioural
Andrubis26 (2014) Analysis and Detection Static, Dynamic, Off-device Higher Lesser Free
Profile-based and Behavioural
CopperDroid45(2015) Analysis and Detection Dynamic, System/API Off-Device Limited Lesser Free
and VMI
HinDroid98(2017) Detection Dynamic and API Off-device Limited Lesser –
HEMD99(2018) Detection Dynamic and Permissions Off-device Limited Lesser –
MalDozer49(2018) Detection Dynamic Off-Device Limited Lesser –
DroidDet7(2018) Detection Static Off-device Limited Lesser –
Wei Wang100(2019) Detection Dynamic Off-device Limited Lesser –
MalInsight101(2019) Detection Dynamic Off-device Limited High –
MLDroid3 (2020) Detection Dynamic On-device Unlimited High Free
GDroid2 (2021) Detection Static Off-device Limited Lesser Free
IntDroid102 (2021) Detection Static Off-device Limited Lesser –
DNNDroid103 (2022) Detection Dynamic Off-device Limited Moderate Free
PARUDroid104(2023) Detection Dynamic On-device Limited Moderate Free
YarowskyDroid105 (2023) Detection Dynamic Off-device Limited Lesser Free
PermDroid (our proposed
Detection Dynamic,Permissions, Off-device Unlimited Higher Free
framework)
API calls, user-rating
and Number of user download
app
Table 15. Comparison with different approaches/frameworks proposed in the literature. Experiment was
performed by using Drebin data set (https://www.sec.cs.tu-bs.de/~danarp/drebin/download.html).
CAT
Cyren Ikarus VIPRE McAfee AVG AVware ESET NOD32 QuickHeal AegisLab NANO Antivirus SF with DNN SF with DNN
Full data set 82% 82.68% 89% 89% 90% 92.8% 92.9% 96.9% 97.1% 96.2% 98.8% 98.8%
Speed in detect-
60 62 40 30 32 30 20 32 30 20 12 12
ing
malware in Sec
Table 16. PermDroid and antivirus scanner detection rates. For this experiment, we use .apk file that’s less
than 27 MB in size. The experiment was carried out using 1000 different Android apps from the real world.
Experimental outcomes
The conclusions reached after conducting experimental work are presented in this section of the paper. The
empirical work was done using a neural network and six different machine learning techniques, including GDA,
NM, GD, GDM, LM, and DNN, as well as three ensemble approaches. The developed models outperform previ-
ously used classifiers in the literature (Table 11) and can detect malware from both known and unknown families
(Table 18, Fig. 20). Additionally, they increase the rate of detection by different Antivirus scanners (stated in
Table 15). It is clear from Fig. 20 and Tables 14, 15, 16, and 18 that:
• PermDroid can detect 98.8% of Android malware, which is impossible for most AV scanners on the market.
• With a detection rate of 98.8% for both known and unknown malware types, PermDroid is capable of finding
malware.
The proposed framework is able to answer the research questions mentioned in “Research questions” section:
Vol:.(1234567890)
www.nature.com/scientificreports/
Table 17. Top malware families are taken into account in our data set.
1. To verify the importance of the correlation between the feature sets and the malware detection model, the
t-test and ULR analysis are used. It is discovered that there are several separate sets of features that are highly
connected with the creation of malware detection models as a result of this research.
2. From Fig. 11, it can be noticed that certain sets of features pass a high correlation with other sets of features
(i.e., the case with a black square is having high negative correlation, and the case with a black circle is hav-
ing a high positive correlation). It is essential to remove the collinearity among the features, for calculating
the ability of each feature. In this manner, the models developed by selecting sets of the feature are capable
to detect malware and do not suffer from the aspect of collinearity.
Vol.:(0123456789)
www.nature.com/scientificreports/
Table 18. Detecting unknown malware families with the help of the PermDroid framework proposed in this
study.
3. Forward stepwise selection process, ULR, correlation analysis, and t-test analysis are implemented to select
features that are able to identify whether the app is malicious or not. The model built by applying the speci-
fied sets of features produces better outcomes when compared to the rest, according to t-test analysis.
4. Six various types of machine learning techniques based on neural network principles, such as NM, GD, LM,
GDM, GDA, and DNN, as well as three ensemble approaches, are implemented in detecting whether an app
is benign or malicious. From the Tables 8 and 9, it is apparent that the model developed using an ANN and
the Deep Neural Network (DNN) approach produces the best results when compared to other techniques.
5. Tables 8 and 9 and Figs. 18, 19 and 20 show that our suggested model is effective in identifying malware from
real-world apps when API calls, permissions, app rating, and the number of people that have downloaded
the app are all considered features.
Threats to validity
In this section, threats to validity are discussed that are experienced while performing the experiment. Three
different threats are mentioned below:
i. Construct validity The Android malware detection methodology in this research study is capable of detect-
ing whether an app is benign or malicious, however it does not specify how many features are needed to
find vulnerabilities in Android apps.
ii. Internal validity The homogeneity of the data set employed in this research work is the second threat. Apps
are collected from a variety of promised repositories. Any errors made while gathering data from these
sources are not taken into account in this study. Although, it cannot promise that the data collected and
retrieved for our analysis is 100 percent accurate, it can be believed that it assembled consistently.
iii. External validity To train the Android malware detection algorithm, 141 different malware families are
considered. Furthermore, the research can be extended to include other malware families in order to train
the technique to identify malicious apps.
• Based on the feature selection method, it is discovered that there is a limited group of attributes that can
detect malware or benign apps with greater accuracy and lower values of incorrectly classified errors.
• Using our feature selection method sets S25, S28, S19, S14, S9, and S4 of features were discovered to be
important malware detectors.
• Based on the Wilcoxon signed-rank test, it is found that there is a significant difference between all extracted
features and the selected feature sets. It is found that, after calculating the mean difference that the model
developed with the input of the selected feature sets outperformed the model with the input of all extracted
feature sets.
Vol:.(1234567890)
www.nature.com/scientificreports/
• Different classification algorithms differ significantly, according to the Wilcoxon signed-rank test. By calcu-
lating the mean difference value, it is discovered that the model created by combining a neural network with
the Deep-Learning machine-learning algorithm produced superior results than the other machine learning
methods used in this study.
• It may be inferred from the results of the experiments that the NDTF approach performed better than other
ensemble methods.
• Our used classifier outperformed the performance of the classifiers used in the literature, as shown in Fig. 20
and Tables 11 and 14.
• According to the results of the experiments (Tables 8, 9), the malware detection model built was not signifi-
cantly harmed after deleting 60% of the possible number of sets of features; in fact, in almost all cases, the
results were better.
• As shown in Table 18 and Fig. 20, our proposed malware detection system can detect malware from both
known and undiscovered malware families.
This study established that a malware detection method merely identifies whether an app is malicious or benign.
Several avenues can be explored for future research. Firstly, a large amount of Android apps are required to
develop the model, memorize and disclose information related to the data set. Second, it is also difficult to make a
centralized system at the time of training and testing the model. Third, decentralized, privacy-preserving classifier
model will be proposed for detecting Android malwares. Further, it is also be discovered how many permissions
are necessary to evaluate whether an app is dangerous or not, more investigation may be done.
Data availibility
For materials should be addressed to corresponding authors.
References
1. Faruki, P. et al. Android security: A survey of issues, malware penetration, and defenses. IEEE Commun. Surv. Tutor. 17(2),
998–1022 (2014).
2. Gao, H., Cheng, S. & Zhang, W. Gdroid: Android malware detection and classification with graph convolutional network.
Comput. Secur. 106, 102264 (2021).
3. Mahindru, A. & Sangal, A. MLDroid—framework for android malware detection using machine learning techniques. Neural
Comput. Appl. 33, 1–58 (2020).
4. Fereidooni, H., Conti, M., Yao, D. & Sperduti, A. Anastasia: Android malware detection using static analysis of applications. In
2016 8th IFIP International Conference on New Technologies, Mobility and Security (NTMS), 1–5 (IEEE, 2016).
5. Arp, D. et al. Drebin: Effective and explainable detection of android malware in your pocket. Ndss 14, 23–26 (2014).
6. Yuan, Z., Lu, Y. & Xue, Y. Droiddetector: Android malware characterization and detection using deep learning. Tsinghua Sci.
Technol. 21(1), 114–123 (2016).
7. Zhu, H. J. et al. Droiddet: Effective and robust detection of android malware using static analysis along with rotation forest
model. Neurocomputing 272, 638–646 (2018).
8. Wong, M. Y. & Lie, D. Intellidroid: A targeted input generator for the dynamic analysis of android malware. NDSS 16, 21–24
(2016).
9. Dash, S. K., Suarez-Tangil, G., Khan, S., Tam, K., Ahmadi, M., Kinder, J. & Cavallaro, L. Droidscribe: Classifying android malware
based on runtime behavior. In: 2016 IEEE Security and Privacy Workshops (SPW), 252–261 (IEEE, 2016).
10. Chen, S., Xue, M., Tang, Z., Xu, L. & Zhu, H. Stormdroid: A streaminglized machine learning-based system for detecting android
malware. In Proceedings of the 11th ACM on Asia Conference on Computer and Communications Security, 377–388 (2016).
11. Mariconti, E., Onwuzurike, L., Andriotis, P., Cristofaro, E. D., Ross, G. & Stringhini, G. Mamadroid: Detecting Android Malware
by Building Markov Chains of Behavioral Models. arXiv:1612.04433 (2016)
12. Kabakus, A. T. DroidMalwareDetector: A novel android malware detection framework based on convolutional neural network.
Expert Syst. Appl. 206, 117833 (2022).
13. Mahindru, A. & Sangal, A. Deepdroid: Feature selection approach to detect android malware using deep learning. In: 2019 IEEE
10th International Conference on Software Engineering and Service Science (ICSESS), 16–19 (IEEE, 2019).
14. Mahindru, A. & Sangal, A. Feature-based semi-supervised learning to detect malware from android. In Automated Software
Engineering: A Deep Learning-Based Approach, 93–118 (Springer, 2020).
15. Mahindru, A. & Sangal, A. Perbdroid: Effective malware detection model developed using machine learning classification
techniques. In A Journey Towards Bio-inspired Techniques in Software Engineering 103–139 (Springer, 2020).
16. Mahindru, A. & Sangal, A. Hybridroid: An empirical analysis on effective malware detection model developed using ensemble
methods. J. Supercomput. 77(8), 8209–8251 (2021).
17. Mahindru, A. & Sangal, A. Semidroid: A behavioral malware detector based on unsupervised machine learning techniques
using feature selection approaches. Int. J. Mach. Learn. Cybern. 12(5), 1369–1411 (2021).
18. Zhao, Y. et al. On the impact of sample duplication in machine-learning-based android malware detection. ACM Trans. Softw.
Eng. Methodol. (TOSEM) 30(3), 1–38 (2021).
19. Yumlembam, R., Issac, B., Jacob, S. M. & Yang L. IoT-based android malware detection using graph neural network with adver-
sarial defense. IEEE Internet Things J. (2022).
20. Kumar, L., Misra, S. & Rath, S. K. An empirical analysis of the effectiveness of software metrics and fault prediction model for
identifying faulty classes. Comput. Stand. Interfaces 53, 1–32 (2017).
21. Faruki, P., Ganmoor, V., Laxmi, V., Gaur, M. S. & Bharmal, A. Androsimilar: Robust statistical feature signature for android
malware detection. In Proceedings of the 6th International Conference on Security of Information and Networks, 152–159 (2013).
22. Milosevic, J., Malek, M. & Ferrante, A. Time, accuracy and power consumption tradeoff in mobile malware detection systems.
Comput. Secur. 82, 314–328 (2019).
23. Shabtai, A., Kanonov, U., Elovici, Y., Glezer, C. & Weiss, Y. Andromaly: A behavioral malware detection framework for android
devices. J. Intell. Inf. Syst. 38(1), 161–190 (2012).
24. Badhani, S. & Muttoo, S. K. Android malware detection using code graphs. In System Performance and Management Analytics,
203–215 (Springer, 2019).
Vol.:(0123456789)
www.nature.com/scientificreports/
25. Xu, R., Saïdi, H. & Anderson, R. Aurasium: Practical policy enforcement for android applications. In Presented as part of the
21st{USENIX}Security Symposium ( {USENIX}Security 12), 539–552 (2012).
26. Lindorfer, M., Neugschwandtner, M., Weichselbaum, L., Fratantonio, Y., Veen, V. V. D. & Platzer, C. (2014) Andrubis–1,000,000
apps later: A view on current android malware behaviors. In 2014 Third International Workshop on Building Analysis Datasets
and Gathering Experience Returns for Security (BADGERS), 3–17 (IEEE).
27. Ikram, M., Beaume, P. & Kâafar, M. A. Dadidroid: An Obfuscation Resilient Tool for Detecting Android Malware via Weighted
Directed Call Graph Modelling. arXiv:1905.09136 (2019).
28. Shen, F., Vecchio, J. D., Mohaisen, A., Ko, S. Y. & Ziarek, L. Android malware detection using complex-flows. IEEE Trans. Mob.
Comput. 18(6), 1231–1245 (2018).
29. Yang, W., Prasad, M. R. & Xie, T. Enmobile: Entity-based characterization and analysis of mobile malware. In Proceedings of the
40th International Conference on Software Engineering, 384–394 (2018).
30. Enck, W. et al. Taintdroid: an information-flow tracking system for realtime privacy monitoring on smartphones. ACM Trans.
Comput. Syst. (TOCS) 32(2), 1–29 (2014).
31. Portokalidis, G., Homburg, P., Anagnostakis, K. & Bos, H. (2010) Paranoid android: Versatile protection for smartphones. In
Proceedings of the 26th Annual Computer Security Applications Conference, 347–356.
32. Bläsing, T., Batyuk, L., Schmidt, A. D., Camtepe, S. A. & Albayrak, S. An android application sandbox system for suspicious
software detection. In 2010 5th International Conference on Malicious and Unwanted Software, 55–62 (IEEE, 2010).
33. Aubery-Derrick, S. Detection of Smart Phone Malware. Unpublished Ph.D. Thesis, 1–211 (Electronic and Information Technol-
ogy University, Berlin, 2011).
34. Burguera, I., Zurutuza, U. & Nadjm-Tehrani, S. Crowdroid: Behavior-based malware detection system for android. In Proceed-
ings of the 1st ACM Workshop on Security and Privacy in Smartphones and Mobile Devices, 15–26 (2011).
35. Grace, M. C., Zhou, Y., Wang, Z. & Jiang, X. Systematic detection of capability leaks in stock android smartphones. In NDSS,
vol 14, 19 (2012).
36. Grace, M., Zhou, Y., Zhang, Q., Zou, S. & Jiang, X. Riskranker: Scalable and accurate zero-day android malware detection. In
Proceedings of the 10th International Conference on Mobile Systems, Applications, and Services, 281–294 (2012).
37. Zheng, C., Zhu, S., Dai, S., Gu, G., Gong, X., Han, X. & Zou, W. Smartdroid: An automatic system for revealing UI-based trigger
conditions in android applications. In Proceedings of the Second ACM Workshop on Security and Privacy in Smartphones and
Mobile Devices, 93–104 (2012).
38. Dini, G., Martinelli, F., Saracino, A. & Sgandurra, D. Madam: A multi-level anomaly detector for android malware. In Interna-
tional Conference on Mathematical Methods, Models, and Architectures for Computer Network Security, 240–253 (Springer, 2012).
39. Yan, L. K. & Yin, H. Droidscope: Seamlessly reconstructing the {OS} and Dalvik semantic views for dynamic android malware
analysis. In Presented as part of the 21st{USENIX}Security Symposium ( {USENIX}Security 12), 569–584 (2012).
40. Backes, M., Gerling, S., Hammer, C., Maffei, M. & von Styp-Rekowsky, P. Appguard–enforcing user requirements on android
apps. In International Conference on TOOLS and Algorithms for the Construction and Analysis of Systems, 543–548 (Springer,
2013).
41. Shahzad, F., Akbar, M., Khan, S. & Farooq, M. Tstructdroid: Realtime malware detection using in-execution dynamic analysis of
kernel process control blocks on android. Tech Rep (National University of Computer and Emerging Sciences, Islamabad, 2013).
42. Rastogi, V., Chen, Y. & Enck, W. Appsplayground: Automatic security analysis of smartphone applications. In Proceedings of the
third ACM Conference on Data and Application Security and Privacy, 209–220 (2013).
43. Rosen, S., Qian, Z. & Mao, Z. M. Appprofiler: A flexible method of exposing privacy-related behavior in android applications
to end users. In Proceedings of the Third ACM Conference on Data and Application Security and Privacy, 221–232 (2013).
44. Desnos, A. et al. Androguard-reverse engineering, malware and goodware analysis of android applications. URL code google
com/p/androguard 153 (2013).
45. Tam, K., Khan, S. J., Fattori, A. & Cavallaro, L. Copperdroid: Automatic reconstruction of android malware behaviors. In Ndss
(2015).
46. Suarez-Tangil, G., Dash, S. K., Ahmadi, M., Kinder, J., Giacinto, G. & Cavallaro, L. Droidsieve: Fast and accurate classification
of obfuscated android malware. In Proceedings of the Seventh ACM on Conference on Data and Application Security and Privacy,
309–320 (2017).
47. Idrees, F., Rajarajan, M., Conti, M., Chen, T. M. & Rahulamathavan, Y. Pindroid: A novel android malware detection system
using ensemble learning methods. Comput. Secur. 68, 36–46 (2017).
48. Martín, A., Menéndez, H. D. & Camacho, D. Mocdroid: Multi-objective evolutionary classifier for android malware detection.
Soft. Comput. 21(24), 7405–7415 (2017).
49. Karbab, E. B., Debbabi, M., Derhab, A. & Mouheb, D. Maldozer: Automatic framework for android malware detection using
deep learning. Digit. Investig. 24, S48–S59 (2018).
50. Lee, W. Y., Saxe, J. & Harang, R. Seqdroid: Obfuscated android malware detection using stacked convolutional and recurrent
neural networks. In Deep Learning Applications for Cyber Security, 197–210 (Springer, 2019).
51. Alzaylaee, M. K., Yerima, S. Y. & Sezer, S. DL-Droid: Deep learning based android malware detection using real devices. Comput.
Secur. 89, 101663 (2020).
52. Yuan, Z., Lu, Y., Wang, Z. & Xue, Y. Droid-sec: Deep learning in android malware detection. In Proceedings of the 2014 ACM
Conference on SIGCOMM, 371–372 (2014).
53. Zhang, M., Duan, Y., Yin, H. & Zhao, Z. Semantics-aware android malware classification using weighted contextual API depend-
ency graphs. In Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communications Security, 1105–1116 (2014).
54. Shankar, V. G., Somani, G., Gaur, M. S., Laxmi, V. & Conti, M. Androtaint: An efficient android malware detection framework
using dynamic taint analysis. In 2017 ISEA Asia Security and Privacy (ISEASP), 1–13 (IEEE, 2017).
55. Mahindru, A. & Singh, P. Dynamic permissions based android malware detection using machine learning techniques. In Pro-
ceedings of the 10th Innovations in Software Engineering Conference, 202–210 (2017).
56. Shi, B. et al. Prediction of recurrent spontaneous abortion using evolutionary machine learning with joint self-adaptive sime
mould algorithm. Comput. Biol. Med. 148, 105885 (2022).
57. Zhang, Q., Wang, D. & Wang, Y. Convergence of decomposition methods for support vector machines. Neurocomputing 317,
179–187 (2018).
58. Hou, S., Saas, A., Chen, L. & Ye, Y. Deep4maldroid: A deep learning framework for android malware detection based on linux
kernel system call graphs. In 2016 IEEE/WIC/ACM International Conference on Web Intelligence Workshops (WIW), 104–111
(IEEE, 2016).
59. Nix, R. & Zhang, J. Classification of android apps and malware using deep neural networks. In 2017 International Joint Confer-
ence on Neural Networks (IJCNN), 1871–1878 (IEEE, 2017).
60. Zhang, X. A deep learning based framework for detecting and visualizing online malicious advertisement. Ph.D. Thesis, Uni-
versity of New Brunswick (2018)
61. Nauman, M., Tanveer, T. A., Khan, S. & Syed, T. A. Deep neural architectures for large scale android malware analysis. Clust.
Comput. 21(1), 569–588 (2018).
62. Xiao, X., Wang, Z., Li, Q., Xia, S. & Jiang, Y. Back-propagation neural network on Markov chains from system call sequences: a
new approach for detecting android malware with system call sequences. IET Inf. Secur. 11(1), 8–15 (2016).
Vol:.(1234567890)
www.nature.com/scientificreports/
63. Martinelli, F., Marulli, F. & Mercaldo, F. Evaluating convolutional neural network for effective mobile malware detection. Procedia
Comput. Sci. 112, 2372–2381 (2017).
64. Xiao, X., Zhang, S., Mercaldo, F., Hu, G. & Sangaiah, A. K. Android malware detection based on system call sequences and
LSTM. Multim. Tools Appl. 78(4), 3979–3999 (2019).
65. Dimjašević, M., Atzeni, S., Ugrina, I. & Rakamaric, Z. Evaluation of android malware detection based on system calls. In Pro-
ceedings of the 2016 ACM on International Workshop on Security and Privacy Analytics, 1–8 (2016).
66. Mas’ud, M. Z., Sahib, S., Abdollah, M. F., Selamat, S. R. & Yusof, R. Analysis of features selection and machine learning classifier
in android malware detection. In 2014 International Conference on Information Science and Applications (ICISA), 1–5 (IEEE,
2014).
67. Yerima, S. Y., Sezer, S., McWilliams, G. & Muttik, I. A new android malware detection approach using Bayesian classification. In
2013 IEEE 27th International Conference on Advanced Information Networking and Applications (AINA), 121–128 (IEEE, 2013).
68. Narudin, F. A., Feizollah, A., Anuar, N. B. & Gani, A. Evaluation of machine learning classifiers for mobile malware detection.
Soft. Comput. 20(1), 343–357 (2016).
69. Wang, W. et al. Exploring permission-induced risk in android applications for malicious application detection. IEEE Trans. Inf.
Forensics Secur. 9(11), 1869–1882 (2014).
70. Ayar, M., Isazadeh, A., Gharehchopogh, F. S. & Seyedi, M. NSICA: Multi-objective imperialist competitive algorithm for feature
selection in arrhythmia diagnosis. Comput. Biol. Med. 161, 107025 (2023).
71. Hu, H. et al. Dynamic individual selection and crossover boosted forensic-based investigation algorithm for global optimization
and feature selection. J. Bionic Eng. 20, 1–27 (2023).
72. Zhong, C., Li, G., Meng, Z., Li, H. & He, W. A self-adaptive quantum equilibrium optimizer with artificial bee colony for feature
selection. Comput. Biol. Med. 153, 106520 (2023).
73. Zhou, P. et al. Unsupervised feature selection for balanced clustering. Knowl.-Based Syst. 193, 105417 (2020).
74. Allix, K. et al. Empirical assessment of machine learning-based malware detectors for android. Empir. Softw. Eng. 21(1), 183–211
(2016).
75. Narayanan, A., Chandramohan, M., Chen, L. & Liu, Y. A multi-view context-aware approach to android malware detection and
malicious code localization. Empir. Softw. Eng. 23(3), 1222–1274 (2018).
76. Azmoodeh, A., Dehghantanha, A. & Choo, K. K. R. Robust malware detection for internet of (battlefield) things devices using
deep eigenspace learning. IEEE Trans. Sustain. Comput. 4(1), 88–95 (2018).
77. Chen, K. Z., Johnson, N. M., D’Silva, V., Dai, S., MacNamara, K., Magrino, T. R., Wu, E. X., Rinard, M. & Song, D. X. Contextual
policy enforcement in android applications with permission event graphs. In: NDSS, 234 (2013).
78. Yerima, S. Y., Sezer, S. & McWilliams, G. Analysis of Bayesian classification-based approaches for android malware detection.
IET Inf. Secur. 8(1), 25–36 (2014).
79. Gonzalez, H., Stakhanova, N. & Ghorbani, A. A. Droidkin: Lightweight detection of android apps similarity. In International
Conference on Security and Privacy in Communication Networks, 436–453 (Springer, 2014) .
80. Kadir, A. F. A., Stakhanova, N. & Ghorbani, A. A. Android botnets: What urls are telling us. In International Conference on
Network and System Security, 78–91 (Springer, 2015).
81. Zhou, Y. & Jiang, X. Android malware genome project. Disponibile a http://www.malgenomeproject.org (2012).
82. Garcia, J., Hammad, M. & Malek, S. Lightweight, obfuscation-resilient detection and family identification of android malware.
ACM Trans. Softw. Eng. Methodol. (TOSEM) 26(3), 1–29 (2018).
83. Mahindru, A. & Sangal, A. Parudroid: Validation of android malware detection dataset. J. Cybersecur. Inform. Manag. 3(2),
42–52 (2020).
84. McCulloch, W. S. & Pitts, W. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 5(4), 115–133
(1943).
85. Faruk, M. J. H., Shahriar, H., Valero, M., Barsha, F. L., Sobhan, S., Khan, M. A., Whitman, M., Cuzzocrea, A., Lo, D., Rahman,
A., et al. Malware detection and prevention using artificial intelligence techniques. In 2021 IEEE International Conference on
Big Data (Big Data), 5369–5377 (IEEE, 2021).
86. Battiti, R. First-and second-order methods for learning: Between steepest descent and newton’s method. Neural Comput. 4(2),
141–166 (1992).
87. Levenberg, K. A method for the solution of certain non-linear problems in least squares. Q. Appl. Math. 2(2), 164–168 (1944).
88. Bengio, Y. Learning deep architectures for AI. Found. Trends® Mach. Learn.2(1), 1–127 (2009).
89. Kaur, J., Singh, S., Kahlon, K. S. & Bassi, P. Neural network-a novel technique for software effort estimation. Int. J. Comput.
Theory Eng. 2(1), 17 (2010).
90. Doraisamy, S., Golzari, S., Mohd, N., Sulaiman, M. N. & Udzir, N. I. A study on feature selection and classification techniques
for automatic genre classification of traditional Malay music. In ISMIR, 331–336 (2008).
91. Forman, G. An extensive empirical study of feature selection metrics for text classification. J. Mach. Learn. Res. 3(Mar), 1289–
1305 (2003).
92. Furlanello, C., Serafini, M., Merler, S. & Jurman, G. Entropy-based gene ranking without selection bias for the predictive clas-
sification of microarray data. BMC Bioinform. 4(1), 54 (2003).
93. Coronado-De-Alba, L. D., Rodríguez-Mota, A. & Escamilla-Ambrosio, P. J. Feature selection and ensemble of classifiers for
android malware detection. In 2016 8th IEEE Latin-American Conference on Communications (LATINCOM), 1–6 (IEEE, 2016).
94. Deepa, K., Radhamani, G. & Vinod, P. Investigation of feature selection methods for android malware analysis. Procedia Comput.
Sci. 46, 841–848 (2015).
95. Kothari, C. R. Research methodology: Methods and techniques. New Age International (2004).
96. Chaikla, N. & Qi, Y. Genetic algorithms in feature selection. In IEEE SMC’99 Conference Proceedings. 1999 IEEE International
Conference on Systems, Man, and Cybernetics (Cat. No. 99CH37028), vol 5, 538–540 (IEEE, 1999).
97. Onwuzurike, L. et al. Mamadroid: Detecting android malware by building Markov chains of behavioral models (extended ver-
sion). ACM Trans. Privacy Secur. (TOPS) 22(2), 1–34 (2019).
98. Hou, S., Ye, Y., Song, Y. & Abdulhayoglu, M. Hindroid: An intelligent android malware detection system based on structured
heterogeneous information network. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery
and Data Mining, 1507–1515 (2017) .
99. Zhu, H. J. et al. HEMD: A highly efficient random forest-based malware detection framework for android. Neural Comput. Appl.
30(11), 3353–3361 (2018).
100. Wang, W., Zhao, M. & Wang, J. Effective android malware detection with a hybrid model based on deep autoencoder and con-
volutional neural network. J. Ambient. Intell. Humaniz. Comput. 10(8), 3035–3043 (2019).
101. Han, W., Xue, J., Wang, Y., Liu, Z. & Kong, Z. Malinsight: A systematic profiling based malware detection framework. J. Netw.
Comput. Appl. 125, 236–250 (2019).
102. Zou, D. et al. Intdroid: Android malware detection based on API intimacy analysis. ACM Trans. Softw. Eng. Methodol. (TOSEM)
30(3), 1–32 (2021).
103. Mahindru, A. & Arora, H. Dnndroid: Android malware detection framework based on federated learning and edge computing.
In International Conference on Advancements in Smart Computing and Information Security, 96–107 (Springer, 2022).
Vol.:(0123456789)
www.nature.com/scientificreports/
104. Mahindru, A. & Arora, H. Parudroid: Framework that enhances smartphone security using an ensemble learning approach. SN
Comput. Sci. 4(5), 630 (2023).
105. Mahindru, A., Sharma, S. K. & Mittal, M. Yarowskydroid: Semi-supervised based android malware detection using federation
learning. In 2023 International Conference on Advancement in Computation & Computer Technologies (InCACCT), 380–385
(IEEE, 2023).
Acknowlegment
This work was partly supported by the Technology Innovation Program funded by the Ministry of Trade, Indus-
try & Energy (MOTIE) (No.20022899) and by the Technology Development Program of MSS (No.S3033853).
Author contributions
All the authors have contributed equally.
Competing interests
The authors declare no competing interests.
Additional information
Supplementary Information The online version contains supplementary material available at https://doi.org/
10.1038/s41598-024-60982-y.
Correspondence and requests for materials should be addressed to A.M., S.K.G., S.M. or J.K.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or
format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the
Creative Commons licence, and indicate if changes were made. The images or other third party material in this
article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the
material. If material is not included in the article’s Creative Commons licence and your intended use is not
permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Vol:.(1234567890)