0% found this document useful (0 votes)
37 views

Intrusion Detection System Based on Pattern Recognition

Uploaded by

electro-ub ub
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
37 views

Intrusion Detection System Based on Pattern Recognition

Uploaded by

electro-ub ub
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 9

Arabian Journal for Science and Engineering

https://doi.org/10.1007/s13369-022-07421-0

RESEARCH ARTICLE-COMPUTER ENGINEERING AND COMPUTER SCIENCE

Intrusion Detection System Based on Pattern Recognition


Mohamed M. Abdeldayem1,2

Received: 24 April 2022 / Accepted: 24 October 2022


© King Fahd University of Petroleum & Minerals 2022

Abstract
Artificial intelligence has been developed to be able to solve difficult problems that involve huge amounts of data and that
require rapid decision-making in most branches of science and business. Machine learning is one of the most prominent
areas of artificial intelligence, which has been used heavily in the last two decades in the field of network security, especially
in Intrusion Detection Systems (IDS). Pattern recognition is a machine learning method applied in medical applications,
image processing, and video processing. In this article, two layers’ IDS is proposed. The first layer classifies the network
connection according to the used service. Then, a minimum number of features that optimize the detection accuracy of
malicious activities on that service are identified. Using those features, the second layer classifies each network connection as
an attack or normal activity based on the pattern recognition method. In the training phase, two multivariate normal statistical
models are created: the normal behavior model and the attack behavior model. In the testing and running phases, a maximum
likelihood estimation function is used to classify a network connection into attack or normal activity using the two multivariate
normal statistical models. The experimental results prove that the proposed IDS has superiority over related IDSs for network
intrusion detection. Using only four features, it successfully achieves DR of 97.5%, 0.001 FAR, MCC 95.7%, and 99.8%
overall accuracy.

Keywords Intrusion detection system (IDS) · Pattern recognition · Machine learning techniques · Network security

1 Introduction passes through the network. It provides a thorough inspec-


tion of the packet and then determines whether or not it
In the era of the communication and information revolution, exhibits abnormal behavior according to network packet fea-
the use of the Internet has grown with a steady increase in tures. IDSs can be categorized based on the detection process
various areas of life. In addition, the Internet of Things (IoT) into misuse or anomalies. In misuse-based IDS, a database of
and the precautionary precautions as a result of the COVID- known malicious behaviors is created, and each new packet is
19 pandemic have led to tremendous use of the Internet in compared against this database to be identified as an attack or
various fields such as distance education, government trans- not. In the case of anomaly IDS, a model of normal behavior is
actions, banking transactions, entertainment, and electronic constructed, and then, each new packet is examined to deter-
commerce. With the growth of Internet use, the risks increase mine if it belongs to the model of normal behavior or not, and
and the need for security becomes urgent to build trust and as a result, it can be identified as a normal or harmful activity.
protect the network assets. Therefore, cybersecurity always As Machine Learning (ML) has evolved, researchers pro-
strives to protect these assets and to provide availability, con- vided training model datasets to detect anomalous malignant
fidentiality, authentication, and integrity of information. An behaviors. ML can be categorized into three types of learning:
IDS is a cybersecurity system that captures each packet that supervised, semi-supervised, and unsupervised. The train-
ing datasets are developed by experts; therefore, supervised
B Mohamed M. Abdeldayem learning is the most accurate. Pattern recognition is one of
[email protected] the ML methods used in many applications. In this paper, a
1 Computers Science and Engineering Department, College of pattern recognition service-based IDS is proposed. It can be
Applied Studies and Community Service, King Saud classified as supervised ML. Many related works are intro-
University, Riyadh, Saudi Arabia duced in the next section. Section 3 explains the proposed
2 Information Technology Department, Faculty of Computers IDS architecture, training phase, and testing phase. Section
and Artificial Intelligence, Cairo University, Cairo, Egypt

123
Arabian Journal for Science and Engineering

4 presents IDS performance metrics; however, Sect. 5 dis-


cusses the experimental results based on these performance
metrics. Section 6 concludes and summarizes the paper. In
addition, future works are illustrated.

2 Related IDSs Works

Some of the related IDSs will be introduced in this section. Fig. 1 Training dataset service classifier
Some MLs are designed based on the decision tree hierar-
chy [1–4], which is a simple supervised ML technique. Each
leaf on the decision tree includes some rules; according to ples were preserved. Moreover, other types of malicious
these rules, the data are classified down from the root to behaviors were preserved by OvA-HMC, which makes the
the final decision tree leaf until ranked. Decision trees have distribution learned from applying the techniques on the C-
been used with success in several classification problems NSLKDD standard dataset similar to the learned distribution
recently. For instance, a decision tree algorithm detects mali- from applying the techniques on a real dataset. Consequently,
cious domain in domain name systems [4]. Cyber-attacks are OvA-HMC exceeds OvR-HMC and similar IDS techniques.
predicted based on a deep neural network [5]; in addition, An IDS is proposed to detect malicious behavior on HTTP
many datasets are evaluated to conduct the benchmark. Self- network service [16]. First, it identifies the connection fea-
Taught Learning method (STL) [6] proposed a framework tures that best classify HTTP packets, and then, it classifies
for selecting features. Filter and wrapper features selection network traffic into normal or attack behavior based on the
methodologies are used by J48 and Naïve Bayes ML clas- Naive Bayes equation. It satisfies good IDS performance met-
sifiers to detect network anomalies [7]. In [8], an IDS using rics with a small number of features compared to other IDSs.
Bayesian network algorithms is proposed, and Cyber-attacks That is because it limits classification to a single network
are sorted into stages; therefore, earlier stages of the attack service rather than working on all network services; there-
can be detected. In addition, the behavior of the attack can be fore, it prevents any feature contradiction that degrades attack
predicted, which helps in cyber-attack defense. The proposed detection on each service.
IDS in [9] uses convolutional and recurrent neural network
to find and classify attacks. The NSL-KDD data set is used
to train and evaluate them. In [10], a proposed IDS based on
3 The Proposed IDS
convolutional neural network classifies network traffic into
normal or malicious packets. It was trained and evaluated
The proposed IDS consists of two classifiers: a service-based
using CICIDS2017. The back-propagation neural network
classifier and an intrusion detection classifier. In Subsect.
[11,12] is a supervised learning algorithm that propagates
(3.1), the IDS architecture is illustrated. The training phase
the error signal back to update the weight while reducing
is explained in Subsect. (3.2), while Subsect. (3.2) describes
the error. It is popular in decision-making as it is easy to
the testing phase.
implement, as well as being accurate in forecasting and clas-
sification problems. During the training phase, the actual
and target outputs are compared to update the weights. The 3.1 The IDS Architecture
Support Vector Machine (SVM) classifier sets the support
vectors for feature separation in the hyperplane [13,14], and The IDS includes two classifications levels. The first level
SVM is a supervised ML algorithm. In various applications, classifies the dataset based on the network services, such that
SVMs have produced successful results that include image each subset includes only data that are belonging to a spec-
analysis, biometrics and bioinformatics [13–15]. In [15], the ified network service (HTP, FTP, SMTP, etc.). Then, in the
SVM includes two phases: training and testing, in addition select features process, the target features of the service’s sub-
to radial basis function, which is the preferred kernel. The set are extracted. As illustrated in Fig. 1, the training data are
multi-class classification yields multiple binary classification classified based on network services, and the service training
cases based on two strategies: OvR-HMC and OvA-HMC. dataset is extracted based on a target features. Finally, in the
OvR-HMC in each layer, the class with the largest number, separation process, two datasets are produced: Input Train-
is identified as 1 and the remaining classes as 0. The pre- ing Normal Dataset to model normal behavior and the Input
viously classified samples were then removed. OvA-HMC Training Attack Dataset to model attack behavior.
specifies class and labels as 1 for each layer, with the other The Input Testing dataset for each network service is
labeled as 0. Previously classified as abnormal, these sam- obtained from the Testing dataset, and it includes only the

123
Arabian Journal for Science and Engineering

In contrast, the purpose of the attack behavior model is to


calculate the probability that it is an attack. In the testing
phase, the input testing dataset is normalized using equation
(1); then, the behavior of each connection is compared with
each of normal model and attack models; then, it is classi-
Fig. 2 Testing dataset service classifier fied as attack or normal behavior based on the model that
it belongs to. Finally, the accuracy and benchmark matrices
are determined by comparing the testing decisions with the
actual states of the testing connections.

3.2 Training Phase

In the training phase, the normal behavior of the service net-


work connection is modeled using the Input Training Normal
dataset. A multivariate normal distribution is identified and
its covariance and mean vectors are determined to model
the normal behavior. In addition, the Input Training Attack
dataset is used to identify a multivariate normal distribution
and to determine its covariance and mean vectors. This distri-
bution models the attack behavior against the target network
service. In the same manner as [17], let us define the normal
and attack datasets that input into the training phase after
Fig. 3 Machine learning decision model
normalization process as follows: Each of the normalized
datasets (normal and attack) is within the real metric space
features that are selected in the Input Training dataset figure RN. Each  ofthem can be defined as:  
2. f  k = f 1 , f 2 , f 3 , . . . , f j , . . . , f k−1 , f k ∈ R N
 T
In the second classification level, Fig. 3 is based on
f i =  , f , f ,..., f
f i,1 i,2 i,3 , f
i,N −1 i,N
 ; ∀i ∈
supervised ML. Equation (1) is used to normalize the Input
{1, 2, 3, . . . , k}
Training Normal Dataset and the Input Training Attack
where N is the number of connections and
Dataset. Please note that in Fig. 3 Machine Learning Deci-
k is the number of features.
sion Model, the normalization process is implemented twice
Each of multivariate normal and attack datasets is pre-
to scale each feature within the range of values [0,1]. The
sented as the following matrix (Eq. 2):
normalization process is used in the training phase to nor-
malize the training dataset and during the testing phase to
normalize the testing dataset. ⎡ ⎤
f 1,1 f 2,1 · · · f k,1
  ⎢ f 1,1 f 2,2 · · · f k,2 ⎥
⎢ ⎥
fi f =⎢ .. .. .. .. .. .. ⎥ (2)
∀ f j ∈ {F}k , f j,i =  , f j,i ∈ [0, 1]
k ⎣ . . . . . . ⎦
max f n j ,i , f a j ,i )
f 1,N f 2,N · · · f k,N
(1)

The multivariate normal probability density function is


where
illustrated in Eq. 3 as follows [18]:
{F}k : the set of k features
f j,i : value of ith feature of {F}k for j connection
f n j ,i : value of ith feature of normal dataset for j connec-
1 1 
tion f (x) = e− 2 (x−μ)  −1 (x − μ) (3)
f a j,i : value of ith feature of attack dataset for j connection (2π ) p/2 || 1/2

f j,i :normalized value of ith feature for j connection


Based on the pattern recognition method, normal behavior where -∞ < x j < ∞, j=1,2, …, p
and attack behavior models are created in the training phase. p is the number of variable, and the data vector is [x1,
The purpose of the normal behavior model is to determine x2,xp];
the probability that the connection is a normal connection.  is the covariance p× p matrix.

123
Arabian Journal for Science and Engineering

To apply Eq. 3 to the proposed


  IDS, x will be replaced by The covariance of the multivariate attack dataset are sa jk
the values of the features f  k and p will be replaced by k (Eq. 12),
as shown in Eq. 4:

  1    1 
N
  
e− 2 ( f −μ)  −1 f  − μ
1 
f f = (4) sa jk = f ai j − f a j f aik − f a k (12)
(2π )k/2 |  |1/2 N −1
i=1

where-∞ < x j < ∞, j = 1, 2, . . . , k


The mean vector of the sample, In addition, the sample 3.3 Testing Phase
covariance matrix can be used in Eq. (3) [18]. Therefore,
both of them are used to find the normal distribution model. During testing phase, each connection in the normalized test-
In the same way, the attack distribution model can be found ing dataset is classified as normal or attack behavior. For each
using the multivariate attack dataset sample mean vector and connection, the probability of it being a normal behavior or
covariance matrix. Let the mean vector and the covariance attack behavior is determined by calculating Maximum Like-
matrix for the multivariate normal dataset be f n (Eq. 5) and lihood Estimation (MLE) [19–23]. In order to find the most
Sn (Eq. 6) in order, and the mean vector and the covariance accurate unknown probability model parameters [23], MLE
matrix for the multivariate attack dataset be f a (Eq. 9), and methods have advantages over alternate methods: they are
Sa (Eq. 10) in order. fixed under the parameter conversion, they are simpler, they
are functions of sufficient statistics, and in addition, they
satisfy good convergence properties and are asymptotically
unbiased with a large training dataset [22,23]. Let D be a
1 
N
fn = f ni (5) training data set belonging to the probability density func-
N tion p(D | θ ). It is used to calculate the unknown vector θ .
i=1
Assume D includes independent n samples (x1, x2,...., xn),
1 
N
 
Sn = f ni − f n f ni − f n (6) the likelihood function can be estimated using Eq. (13) [23].
N −1
i=1

The variances of the multivariate normal dataset on the 


n

main diagonal of Sn are sn 2j (Eq.7), p(D | θ ) = p (X k | θ ) (13)


k=1

The MLE of θ that maximize the likelihood function


1 
N
2 p(D | θ ) is  θ . Because our model is based on multivari-
sn 2j = f ni j − f n j (7)
N −1 ate normal distribution, the unknown parameter θ includes
i=1
the distribution’s mean vector μ and the covariance matrix
The covariance of the multivariate normal dataset are sn jk . To determine  we find  , 
μ and  μ the MLE of μ and 
(eq.8), is MLE of  . Both of   will be estimated from the
μ and 
training dataset D (Eqs. 14 and 15) [6].
1 
N
 
sn jk = | f n i j − f n j f n ik − f n k (8)
N −1 1
n
i=1

μ= xk (14)
1  n
N
k=1
fa = f ai (9)
N 1 n
i=1 =
 (xk − 
μ) (xk − 
μ)t (15)

N n
1    k=1
Sa = f ai − f a f ai − f a (10)
N −1
i=1 3.4 MLE for the Proposed Model
The variances of the multivariate attack dataset on the  will be estimated from the proposed training
Both μ and 
main diagonal of Sn are sn 2j (Eq. 11),
dataset; however, the proposed model includes the normal
training dataset f  and
the attack training dataset f a ; thus,
θ̂1 includes μ̂1 and  1 that will be estimated for normal
1 
N
2 
sn 2j = f ni j − f n j (11) behavior, and θ̂2 includes μ̂2 and  2 for attack behavior. Let
N −1 the normalized test dataset be F , the Eqs. 13, 14 and 15 will
i=1

123
Arabian Journal for Science and Engineering

be rewritten to be applied in the proposed model for each Table 1 Connections classifications
connection Ft in F as follows: For normal behavior: Classified as normal Classified as attack

1  
N Normal TN FP
μ̂1 = fi = f n (16) Attack FN TP
N
i=1

 1 
N
 
= f n i j − f n j f n ik − f n k = sn jk
1 N −1 as attacks to the total number of true normal instances. Acc
i=1
is the total number of true attacks and true normal instances
(17)
detected by the system (TP and TN) across all data set sam-
  
k   ples. Mcc can be used as a measure of the quality of binary
p1 Ft | θ̂1 = p Fi | θ̂1 (18)
classification [30]. While the prediction is perfect if MCC =
i=1
+1, MCC =-1 identifies the worst possible prediction. Table
For attack behavior: 1 shows how the connection can be classified as: TN, FN, TP
or FP [10].

1 
N
μ̂2 = f ai = f a (19) TP
N AD R = (23)
i=1 T P + FN
1 
N
 
2 =
 f ai j − f a j f aik − f a k = sa jk (20) The ADR ratio is used to identify the IDS capability to
N −1
i=1 detect attacks based on the features vector [24–30].
  k  
p2 Ft | θ̂2 = p Fi | θ̂2 (21)
i=1 FP
    F AR = (24)
p1 Ft | θ̂1 > p2 Ft | θ̂2 = { yes, Normal FP + T N

Otherwise, Attack (22) FAR is the ratio of normal connections that are identified
as attacks to all normal connections. If FAR is large, this
(MLE)N is determined using the feature vector of the con- indicates that the IDS is unreliable because it will generate
nection and the multivariate normal distribution of normal too many false alarms [24–30]. The Acc represents ratio of
behavior (Eqs. 16 and 18). In addition, (MLE)A is deter- accurate detected connections to all connections [27].
mined using the feature vector of the connection and the
multivariate normal distribution of attack behavior (Eqs. 19 TP +TN
Acc =
and 21); then, the connection is classified as normal behavior T P + T N + FP + FN
or attack behavior based on Eq. 22. T P × T N − FP × FN
Mcc = √ (25)
(T P + F P)(T P + F N )(T N + F P)(T N + F N )

4 Performance Metrics
5 Results
After classifying all connections in the testing dataset, the
performance metrics for the proposed IDS are determined. The Dell Inspiration 15, 5570 Series, 64-based PC, Core i7,
These include [24–30]: True Positive (TP), False Positive 16 GB RAM, and 1 TB HD and 1 GB VGA are used to
(FP), True Negative (TN), False Negative (FN), Attack test the proposed IDS performance. Jupyter Notebook took
Detection Rate (ADR), Accuracy (Acc), Mathew’s correla- 0.0588 seconds to train the data, while MATLAB Software
tion coefficient (Mcc), and False Alarm Rate (FAR). TP is took 2.4507 seconds. To evaluate IDS, an IDS dataset must
the number of attacks that the IDS detected as attacks; FP is be used in the training and testing phases. MIT Lincoln Labs
the number of attacks that the IDS did not detected as attacks; published KDD-DARPA, which is the most famous standard
it classified them as normal instances; TN is the number of IDS data set. This includes a wide range of intrusion connec-
normal instances that the IDS detected as normal instances; tions. A KDD is used to test a large number of IDSs. Although
FN is the number of normal instances that the IDS didn’t it is old, it is still used to evaluate the capability of IDS. The
detected as normal; it classified them as attacks. ADR is the KDD includes forty-one ordered features for each connec-
ratio of the number of detected attacks to all true attacks. tion, and a table describing these ordered features is found
FAR is the ratio of the number of normal instances detected in [31]. NSL-KDD was published as an enhanced version of

123
Arabian Journal for Science and Engineering

Table 2 NSL-KDD selected features


Index Features

A src-bytes, dst_bytes, hot, count, and srv_count


B src-bytes, hot, count, and srv_count
C src-bytes, dst_bytes, count, and srv_count
D Duration, src-bytes, dst_bytes,hot,count, and srv _count
Bays_9f Duration, src-bytes, dst_bytes,hot, count, and srv_count
Bays_8f src-bytes, dst_bytes, count, and srv_count

Table 3 True/false alarms for Index TP FN TN FP


each IDS
A 307 4 6630 9
Fig. 4 ADR comparison
B 307 4 6626 13
C 306 5 6631 8
D 308 3 6630 9
Bays_9f 307 4 6631 8
Bays_8f 306 5 6630 9

Table 4 ADR, FAR, Acc, and Mcc


Index ADR FAR Acc Mcc

A 97.2% 0.1% 99.7% 95.0%


B 95.9% 0.2% 99.6% 93.1%
C 97.5% 0.1% 99.8% 95.3%
D 97.2% 0.1% 99.7% 95.1%
Bays_9f 97.5% 0.1% 99.8% 95.4% Fig. 5 FAR comparison
Bays_8f 97.1% 0.1% 99.7% 95.6%

the proposal C and Bays_9F have the same ADR (97.5 %),
which is the largest value, while the proposal C has a slightly
the KDD [32]. It eliminates some of the KDD’s inconsisten- lower ADR value (97.2%), as shown in Fig. 4. The FAR for
cies [32]. Therefore, I used NSL-KDD to test proposed IDS all IDSs has the same value (0.1%) except the proposal B has
performance. This includes two level classifiers. The first (0.2%) value (Fig. 5).
level classifies the connection based on the three attribute In Fig. 6, the largest value of Acc is 99.8% which is for the
features: Protocol type, service and flag (features numbers 2, proposal C and Bayes_9F. The Mcc value for Bayes_8F is
3 and 4 in NSL-KDD dataset); in addition, the second level the highest, but it is similar to the Mcc values for Bayes_9F
classifies the connection based on some other features. In and proposal C, as shown in Fig. 7. We can conclude that the
the second level, different combinations of features (A-D) performance of proposal C and Bayes_9F is similar; however,
are selected as shown in Table 2 In addition, it includes the the proposed IDS C uses lower number of features that can
group of features that are used in Bays_9f IDSs. MATLAB enhance the speed of implementations and detection rate.
software [33] is used to test the performance of the proposed In addition, it is based on machine learning technique that
IDS, and Tables 3 and 4 illustrate the performance results accelerate the IDS, while Bayes_9F based on conditional
for each combination of features compared to the results for probability which is more complex and may reduce the IDS
Bayes_9F and Bayes_8f. detection rate.
As shown in Table 3, there are a little difference in the The proposals A and C are coded using PyTorch [34] and
results among all techniques. However, proposal D has the other python library for machine learning, such as pandas
greatest TP alarms and the lowest FN alarms, and propos- [35] and tensorflow [36]. The IDS python code is imple-
als C and Bayes_9F have the largest TN and the lowest FP. mented using Jupyter under Anaconda Navigator [37]. The
Table4illustrates the comparison of ADR, FAR, Acc, and experimental results compared with MATLAB implemen-
Mcc for all proposed IDSs. Please keep in mind that all tation are presented in Tables 5, 6, and Fig. 8. As shown
measurements are in percentages. As the ADR increases, in Tables 5 and 6, and Fig. 8, there is no significant dif-

123
Arabian Journal for Science and Engineering

Table 6 Performance measurements


Index ADR FAR Acc Mcc

A 97.2% 0.1% 99.7% 95.0%


Python-A 93.6% 0.1% 99.5% 95.3%
C 97.5% 0.1% 99.8% 95.3%
Python-C 93.9% 0.2% 99.6% 95.7%

Fig. 6 Acc comparison

Fig. 8 Comparison MATLAB and Python implementation

Table 7 Performance measurements


IDSs ACC (%) FAR DR (%)

Proposed 99.8 0.001 97.5


BAT-MLP 87.7 0.188 92.5
BBO-MLP 87.9 0.146 89.8
CS-MPL 86.82 0.192 91.4
Fig. 7 Mcc comparison DA-MLP 88.3 0.214 95.6
DE-MLP 84.8 0.213 89.5
EBAT-MLP 97.5 .022 97.2
Table 5 True/false experimental results
GA-MLP 88.57 0.182 93.7
Index TP FN TN FP GSA-MLP 84.4 0.310 96.1

A 307 4 6630 9 PSO-MLP 83.9 0.290 93.7

Python-A 409 28 6663 10


C 306 5 6631 8
Python-C 410 27 6665 8

ference between the two implementations. However, the


MATLAB implementation has a little superiority over the
Python implementation in terms of accurate results, but the
Python implementation has a higher speed.
In [11], the performance of 10 IDS ML algorithms is mea-
sured using NSL-KDD data set. Table 7, and Figs. 9, 10 and
11 illustrate the performance of the proposed IDS compared
to the 10 IDS algorithms. The proposed IDS has superior- Fig. 9 Comparison of IDSs ACC
ity over the other IDSs in terms of ACC, FAR, and DR. It
achieves an ACC of 99.8%,a FAR of 0.001, a DR of 97.5%.

123
Arabian Journal for Science and Engineering

References
1. Yee Jian, C., Shih Yin, O., Kok-Seng, W., and Ying Han, P.:
Decision tree with sensitive pruning in network-based intrusion
detection system. In: Computational Science and Technology, pp.
1–10. Springer (2020)
2. Tony, T., Athira, V. P., Sabu, E.: Machine learning approaches in
cyber security analytics. Springer (2020)
3. Chandrashekhar, A., Ashok Kumar, M., Vijay Kumar, J.: Evolu-
tionary decision tree-based intrusion detection system. In: Pro-
ceedings of the third international conference on microelectronics,
computing and communication systems , pp. 271–282. Springer
(2019)
4. Ilham, R., Parman, S., and Muhammad, A. N.: Comparative analy-
Fig. 10 Comparison of IDSs FAR sis of k-nearest neighbor and decision tree in detecting distributed
denial of service. In: 2020 8th international conference on infor-
mation and communication technology (ICoICT)
5. Ravi, V.; Mamoun, A.; Soman, K.P.; Prabaharan, P.; Al-Nemrat, A.;
Sitalakshmi, V.: Deep learning approach for intelligent intrusion
detection system. IEEE Access 7, 41525–41550 (2019)
6. Javaid, Ahmad; Niyaz, Quamar; Sun, Weiqing; Alam, Mansoor:
A deep learning approach for network intrusion detection system.
Eai Endors. Trans. Secur. Saf. 3(9), e2 (2016)
7. Hebatallah, M. A., Mohamed, F., and Ayman, A.-H.: A framework
for efficient network anomaly intrusion detection with features
selection. In: 2018 9th international conference on information and
communication systems (ICICS), pp. 157–162. IEEE (2018)
8. Pivarníková, M.; Sokol, P.; Bajtoš, T.: Early-stage detection of
cyber attacks. Information 11(12), 560 (2020)
9. Sara, A.-E., Aisha, A.-M., Felwa, A.-S.: Using deep learning tech-
niques for network intrusion detection. In: 2020 IEEE international
conference on informatics, IoT, and enabling technologies (ICIoT),
Fig. 11 Comparison of IDSs DR pp. 171–176. IEEE (2020)
10. Samson, H.; Saleh, A.J.; Khalil, D.; Mohammad, M.: A novel
intrusion detection model for detecting known and innovative
cyberattacks using convolutional neural network. IEEE Open J.
Comput. Soc. 2, 14–25 (2021)
6 Conclusion 11. Waheed, A.H.M.G.; Aman, J.: A new approach for intrusion detec-
tion system based on training multilayer perceptron by using
enhanced bat algorithm. Neural Comput. Appl. 32(15), 11665–
Nowadays, machine learning is used in several applications. 11698 (2020)
Pattern recognition is a machine learning method that is used 12. AHMG, Waheed; Aman, J.: An enhanced bat algorithm with muta-
in applications that require rapid decision-making. In this tion operator for numerical optimization problems. Neural Comput.
Appl. 31(1), 617–651 (2019)
paper, a pattern recognition-based IDS is proposed, its struc- 13. Yunqian, M., Guodong, G.: Support vector machines applications,
ture and implementation are described, and its performance is vol. 649. Springer (2014)
tested on the NSL-KDD dataset. The proposed IDS consists 14. Ankita, S., Bhaswati, S., Siddharth, S. R., Manjusha, P.: Analysis
of two levels of classification. The first identifies the required of breast cancer dataset using big data algorithms for accuracy
of diseases prediction. In: International conference on computer
service for the network packet and accordingly selects the set networks and inventive communication technologies, pp. 271–277.
of features that best classifies the packet. The second classi- Springer (2019)
fies the packet as normal or attacking based on the maximum 15. Hsiu-Min, C., Hui-Ying, H., Fanpyn, L., Chung-Hsien, T.: Classi-
likelihood estimation function that is applied to the selected fication of intrusion detection system based on machine learning.
In: International cognitive cities conference, pp. 492–498. Springer
features. The experimental results illustrated that the pro- (2019)
posed intrusion detection system is superior to other intrusion 16. Abd-Eldayem, Mohamed M.: A proposed http service based ids.
detection systems in the matters of accuracy, DR, and FAR. Egypt. Inf. J. 15, 13–24 (2014)
Using only 4 features, it classifies the network packets with 17. Plamen A. P., Xiaowei, G.: Empirical approach to machine learn-
ing. Springer (2019)
99.8% accuracy, 97.5% DR, and 0.001 FAR. For future work, 18. Douglas, C.: Montgomery. Introduction to statistical quality con-
deep learning, the most recent machine learning techniques trol. Wiley (2020)
will be evaluated in network IDS. Other classifiers will be 19. Ulisses, B.-N.: Fundamentals of Pattern Recognition and Machine
applied to the proposed IDS to achieve better detection rate; Learning. Springer, Cham (2020)
20. Arcangelo, D.; Cosimo, D.; Wheeler, D.: Handbook of Image Pro-
in addition, the proposed IDS will be applied on other net- cessing and Computer Vision. Springer, Cham (2020)
work services and will be installed in a real-life network.

123
Arabian Journal for Science and Engineering

21. Tsai, Ming-Tien.: On the maximum likelihood estimation of a 30. Hanan, H.; David, B.; Ethan, B.; Amar Kumar, S.; Christos, T.;
covariance matrix. Math. Methods Stat. 27(1), 71–82 (2018) Robert, A.; Xavier, B.: A taxonomy of network threats and the effect
22. Ameet, J.V.: Machine Learning and Artificial Intelligence. of current datasets on intrusion detection systems. IEEE Access 8,
Springer, Cham (2020) 104650–104675 (2020)
23. Richard, R.J.: Mathematical Statistics: An Introduction to Likeli- 31. KDD Cup. Data (1999)http://kdd.ics.uci.edu/databases/kddcup99/
hood Based Inference. Wiley, London (2018) kddcup99.html (2018)
24. Zeeshan, A., Adnan, S. K., Cheah, W. S., Johari, A., and Farhan, A.: 32. The nsl-kdd data set, https://www.unb.ca/cic/datasets/nsl.html.
Network intrusion detection system: a systematic study of machine accessed 22 sep. (2022)
learning and deep learning approaches. Trans. Emerging Telecom- 33. Matlab,https://www.mathworks.com/products/matlab.
mun. Technol., 32(1):e4150 (2021) html.accessed 22 sep.(2022)
25. Jecheva, V.; Nikolova, Evgeniya: Classification trees as a technique 34. Pytorch, https://pytorch.org/. accessed 22 sep. (2022)
for creating anomaly-based intrusion detection systems. Serdica J. 35. pandas, https://pandas.pydata.org/. accessed 22 sep. (2022)
Comput. 3(4), 335–358 (2009) 36. Tensorflow, https://www.tensorflow.org/. accessed 22 sep. (2022)
26. Natesan, P.; Balasubramanie, P.; Gowrison, G.: Improving attack 37. Anaconda navigator, https://www.anaconda.com/products/
detection rate in network intrusion detection using adaboost algo- individual. accessed 22 sep. (2022)
rithm with multiple weak classifiers. J. Inf. Comput. Sci. 9(8),
2239–2251 (2012)
27. Wang, M.; Zheng, K.; Yang, Y.; Wang, Xiujuan: An explainable
Springer Nature or its licensor (e.g. a society or other partner) holds
machine learning framework for intrusion detection systems. IEEE
exclusive rights to this article under a publishing agreement with the
Access 8, 73127–73141 (2020)
author(s) or other rightsholder(s); author self-archiving of the accepted
28. Adriana-Cristina, E., Valentin, S.: An improved bat algorithm
manuscript version of this article is solely governed by the terms of such
driven by support vector machines for intrusion detection. In:
publishing agreement and applicable law.
Computational intelligence in security for information systems
conference, pp. 41–51. Springer (2015)
29. Mehdi, H.; Amir Masoud, R.; Bay, V.; Moazam, B.; Mohammad,
M.; Mehran, Z.: Improving security using svm-based anomaly
detection: issues and challenges. Soft Comput. 25(4), 3195–3223
(2021)

123

You might also like