Enhanced Malware Detection Using Grey Wolf Optimization and Deep Belief Neural Networks
Enhanced Malware Detection Using Grey Wolf Optimization and Deep Belief Neural Networks
DOI: https://doi.org/10.54692/ijeci.2024.0803206
1
Faculty of Electronics and Information Engineering, Beijing University of
Technology, Beijing, China.
2
Department of Computer Science, University of Alabama at Birmingham AL 35205,
USA.
3
Department of Computer Science, Ripah international University, Sahiwal Campus,
Sahiwal, Pakistan
4
Department of Computer Science, Government Postgraduate College, Nowshera, KP,
Pakistan
ABSTRACT
Standard identification methods are flattering and less effective as attacks from malware
get increasingly sophisticated. Considering current malware outbreaks employ tactics
such as polymorphism, obfuscation and encryption, to avert identification, growing
complicated approaches must be developed. This paper deals with a mixed model
utilizing Deep Belief Neural Network (DBNN) for classifying and Grey Wolf
Optimization (GWO) for choosing features. Whereas DBNN encodes complicated
patterns by hierarchical learning, GWO optimizes the choosing of the more essential
features, lowering the cost of computing and dataset complexity. Investigations reveal
that the suggested GWO-DBNN model beats existing machine learning procedures in
terms of detection accuracy, recall, precision, and false positive rate (FPR). These mixed
tactics offer dependable and scalable solutions to the challenges faced by modern
malware threats.
training and classifications. The actual model that was suggested. Five-
malware datasets utilized in the present fold cross-validation is
research were derived from the implemented to verify stability
Microsoft Malware Classification
while avoiding overfitting.
Challenge [9] on Kaggle. This
collection includes more than 10,000
3.2. Feature Selection using Grey
samples from numerous malware
Wolf Optimization (GWO)
families, among them Ramnit, Simda,
Kelihos, and Vundo. The dataset covers The procedure called GWO [11] [12] is
the static and dynamic information, a metaheuristic approach influenced by
such as system call traces and opcode grey wolves' natural hunting techniques
frequency, making it suitable for both and leadership framework. GWO
static and dynamic training. organizes wolves into four categories:
Preprocessing encompasses numerous alpha, beta, delta, and omega. The alpha
stages: wolf reveals the optimum respond
Missing values can have a serious (optimal feature subset), whilst the beta
influence on the model's and delta wolves direct the search
efficiency. In this research, the process.
The GWO method estimates the
missing data has been solved via
distance that exists among the wolves
mean imputation for numerical and their prey and repeatedly updates
parameters and median imputation their locations to arrive at the optimal
for categorical characteristics. response. The location of updates are
Normalization applies Min-Max determined using the following formula
scaling to align every value of the as:
𝐷∝ = |𝐶1 . 𝑋∝ (𝑡) = 𝑋(𝑡)| (1)
feature across 0 and 1. This
Xα provides the alpha wolf's position,
guarantees that characteristics with while Dα provides the distance from the
wide ranges do not have an optimal feature subset. The technique
excessive effect on the learning repeatedly alters the wolves'
process. placements to reduce the space of
For categorical features, such as features and pick the most appropriate
subset for categorization.
malware families, one-hot
encoding is used to convert 3.3. Classification Using Deep Belief
categorical values into binary Neural Networks (DBNN)
vectors. This avoids any ordinal The architecture of DBNN is shown in
interpretation of categorical figure 1. After identifying the
variables, ensuring that the model appropriate feature subset with GWO,
does not infer unnecessary DBNN employs it for identifying
relationships between malware malware. DBNNs are made up of
numerous layers of Restricted
families. Boltzmann Machines (RBMs) that are
The dataset is divided into train unsupervised learning models [13][14].
(80%) and test (20%) batches for RBM learns to rebuild input by
evaluating the efficacy of the retaining the statistical connections
among hidden and visible units [15]. represents the learning rate. After
The weight adjustment for every RBM pretraining with RBMs, the DBNN is
follows a certain rule is fine-tuned using backpropagation
∆𝑊𝑖𝑗 = 𝜂(≺ 𝑉𝑖 ℎ𝑗 ≻𝑑𝑎𝑡𝑎 − (≺ method to classify malware as either
𝑉𝑖 ℎ𝑗 ≻𝑟𝑒𝑐𝑜𝑛 ) (2) malicious or benign based on the
Where ℎ𝑗 is the hidden unit, 𝑉𝑖 optimized feature subset. The flow
designates the visible unit, and η diagram of the whole methodology has
been shown in figure 2
.