0% found this document useful (0 votes)
46 views8 pages

Enhanced Intrusion Detection System Usin

This document summarizes a research paper that proposes an enhanced intrusion detection system using feature selection and ensemble learning algorithms. The paper aims to select the most relevant features from a dataset to improve the accuracy of the detection model while reducing complexity. It applies feature selection techniques to the NSL-KDD dataset and evaluates different classification algorithms. Experimental results show that the proposed feature selection method improves system accuracy with fewer features. When combining feature selection with ensemble learning algorithms like Random Forest and PART, the best results are achieved using a voting learning algorithm with the product probability rule.

Uploaded by

yonas
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
46 views8 pages

Enhanced Intrusion Detection System Usin

This document summarizes a research paper that proposes an enhanced intrusion detection system using feature selection and ensemble learning algorithms. The paper aims to select the most relevant features from a dataset to improve the accuracy of the detection model while reducing complexity. It applies feature selection techniques to the NSL-KDD dataset and evaluates different classification algorithms. Experimental results show that the proposed feature selection method improves system accuracy with fewer features. When combining feature selection with ensemble learning algorithms like Random Forest and PART, the best results are achieved using a voting learning algorithm with the product probability rule.

Uploaded by

yonas
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

International Journal of Computer Science and Information Security (IJCSIS),

Vol. 16, No. 2, February 2018

Enhanced Intrusion Detection System using Feature


Selection Method and Ensemble Learning Algorithms

Manal Abdullah Arwa Alshannaq


Faculty of Computing and Information Technology Faculty of Computing and Information Technology
King Abdul-Aziz University King Abdul-Aziz University
Jeddah, Saudi Arabia Jeddah, Saudi Arabia
[email protected] [email protected]

Asmaa Balamash Soad Almabdy


Faculty of Computing and Information Technology Faculty of Computing and Information Technology
King Abdul-Aziz University King Abdul-Aziz University
Jeddah, Saudi Arabia Jeddah, Saudi Arabia
[email protected] [email protected]

Abstract— The main goal of Intrusion Detection Systems (IDSs) is sensitive for instance in detection of forest fires [1], power
to detect intrusions. This kind of detection system represents a transmission as well as distribution [2], localization [3],
significant tool in traditional computer based systems for ensuring applications of the military [4], Critical-infrastructures (CIs) [5]
cyber security. IDS model can be faster and reach more accurate and Underwater Wireless Sensor Networks (Underwater WSNs)
detection rates, by selecting the most related features from the [6].
input dataset. Feature selection is an important stage of any IDs to
select the optimal subset of features that enhance the process of the Lack of proper security measures can lead to launching of
training model to become faster and reduce the complexity while different types of attacks in environments that are hostile. These
preserving or enhancing the performance of the system. In this kinds of attacks can interrupt the WSNs from working normally
paper, we proposed a method that based on dividing the input and can defeat the deployment’s purpose. Consequently,
dataset into different subsets according to each attack. Then we security is a significant networks feature. The shortage of means
performed a feature selection technique using information gain makes the creators use primitives of security which are
filter for each subset. Then the optimal features set is generated by traditional such as encryption and one-way functions cautiously.
combining the list of features sets that obtained for each attack. Detection of intrusion is seen as the defense’s second line which
Experimental results that conducted on NSL-KDD dataset shows matches the security primitives. For practicality in implementing
that the proposed method for feature selection with fewer features, WSNs, intrusions detection ideas need to be lightweight,
make an improvement to the system accuracy while decreasing the
scalable as well as distributed. This paper proposes such
complexity. Moreover, a comparative study is performed to the
approaches in the detection of anomaly intrusion in WSNs. In
efficiency of technique for feature selection using different
classification methods. To enhance the overall performance, this kind of context, it is very important to make sure that there
another stage is conducted using Random Forest and PART on is the protection of the sensor network from threats emanating
voting learning algorithm. The results indicate that the best from cyber−security. Regrettably, the achievement of this
accuracy is achieved when using the product probability rule. objective is a bit of a challenge due to features number of WSNs,
highest important one being: inadequate computational
Keywords-Intrusion Detection Systems, NSL-KDD, Feature resources, inhibiting the execution of robust mechanisms that are
Selection, Supervised Learning, Classification. cryptographic; and their distribution in environments that are
wild and unattended, where it is possible for the enemy to access
I. INTRODUCTION the sensor nodes physically, for instance, reading cryptographic
Wireless sensor networks (WSNs) comprise of tiny sensor keys straight from the memory. The fast technology
nodes or devices that have radio, processor, memory; battery as development over the Internet makes the security of a computer
well as sensor hardware. The widespread deployment of these serious issue. Currently, Intelligence which is artificial, data
sensor nodes makes it possible for environmental monitoring. mining as well as machine learning algorithms are exposed to a
These small devices are resource inhibited in terms of the speed broad investigation in ID with stress on enhancing the detection
of the processor, the range of the radio, memory as well as accuracy as well as create a model that is immune for IDS. In
power. This nature of resource inhibition makes designers addition to detection abilities, IDSs also offers extra
design systems that are application specific. While the Wireless mechanisms, for instance, diagnosis as well as prevention.
Sensor Networks are not protected, and the transmitted medium Wireless sensor networks’ IDSs architectures are presently
is wireless, this raises the vulnerability to attacks. WSNs are being examined and various solutions have been recommended
being gradually embraced also in applications which are very in the research.

48 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

This paper concentrates on building IDs for WSN. To robust and effective. The methods that were used combined with
construct an Intrusion Detection System model quicker with classifier are Info Gain, Correlation, Relief and Symmetrical
more correct rates of detection, choice of features that are vital Uncertainty. Their experimental work was divided into two
from the input dataset is extremely important. Learning parts: The first one is building multiclass classifier based on
process’s feature selection while designing the model indicates various decision tree techniques such as ID3, CART, REP Tree,
a decrease in computational rate and improves precision. The REP Tree and C4.5. The second one is applying feature selection
main objective of this paper is determining the greatest suitable technique on the best model obtained which was here C4.5.
features to use in the identification of attack in a dataset of NSL Their experimental analysis was conducted using WEKA tool.
KDD as well as WEKA [7] tool is used for analysis. Different The results showed that C4.5 with Info Gain had better results
performance metrics are used to assess the performance of each and achieved highest accuracy of 99.68% with only 17 features.
classifier such as: precision, recall, F-measure, false positive However, in case of using11 features, Symmetrical Uncertainty
rate, overall accuracy (ACC) and ROC curve. NSL KDD dataset achieved 99.64% accuracy.
[8] is a common dataset for revealing of the anomaly,
particularly for identifying the intrusion. This dataset comprises Deshmukh, 2014 [11], developed IDS using Naive Bayes
of forty-one features that resemble different types of the network classifier with different pre-processing methods. Authors used
traffic. The network traffic is divided into dual classes, one being NSL-KDD dataset and WEKA for their experimental analysis.
the normal class while the other is referred to as the anomaly They compared their results with other classification algorithms
class. The anomaly class usually depicts intrusions or attacks such as NB TREE and AD Tree. The results showed that with
that originate from the network at the time of taking records for respect to the TP rate of all algorithms, the execution time of
the network traffic. In relation to these attacks, the NSL KDD Naïve Bayes is less.
dataset is additionally categorized into four main attack Noureldien Yousif, 2016 [12], examined the performance of
classifications such as the DoS, in addition to probing. Further seven supervised machine learning algorithms in detecting the
classifications comprise of users to root (U2R), as well as remote DoS attacks using NSL-KDD dataset. The experiments were
to local (R2L). The DoS attack renders the unavailability of conducted by using for training step the Train+20 percent file
crucial services to genuine users through the bombardment of and for testing using Test-21 file. they used 10-fold cross
the attack packets that are found on the computing and also on validation in test and evaluate the methods to confirm that
network resources. Instances of DoS attacks contain backland, techniques will achieve on undetected data. Their results showed
and smurf. Moreover, teardrop, plus neptune attacks are also that Random Committee was the best algorithm for detecting
examples of such attacks. Due to the high levels of the risks that smurf attack with accuracy of 98.6161%. At the average rate, the
are found in other types of the DoS attacks that relate to PART algorithm was the best for detecting the Dos Attacks,
computer expenses, the paper primarily dealt on the DoS attacks, however, Input Mapped algorithm was the worst.
as stated in the 2014 document [9]. A DoS attack is viewed as a
major concern for authentic operators retrieving services Jabbar and Samreen, 2016 [13], have presented a novel
through the Internet. DoS attacks render the unattainability of approach for ID using alternating decision trees (ADT) to
services to users through limiting network and also the system classify the various types of attacks while it is usually used for
resources. While a lot of investigation has been performed by binary classification problems. The results showed that their
dint of network security professionals to defeat the DoS attack proposed model produced higher detection rate and reduces the
concerns, DoS attacks are still on the rise and have a more false alarm rate in classification of IDS attacks.
significant detrimental influence as time passes. Paulauskas and Auskalnis, 2017 [14], analyses the initial
The organization of the paper as following. Section 2 data pre-processing influence on attack detection accuracy by
presents an intrusion detection overview, reviews related work. using of ensemble, that are depend on the idea of combining
Section 3 describes IDS proposed model, and Sect. 4 is analysis multiple weaker learners to create a stronger learner, model of
the experimental results obtained. Finally, Section 5 states the four different classifiers: J48, C5.0, Naïve Bayes and PART.
conclusions. Min-Max normalization as well as Z-Score standardization was
applied in pre-processing stage. They compared their proposed
II. LITERATURE REVIEW model with and without pre-processing techniques using more
than one classifier. Their results showed that their proposed
Intrusion detection system uses machine learning algorithms
classifier ensemble model produces more accurate results. After
or classifiers to learn system normal or abnormal behavior and
they presented their results, they were warned not to use only the
build models that help to classify new traffic. Developing an
NSL-KDDTrain+ dataset for both training and testing because
optimal machine learning based detection systems directs
even without pre-processing methods, it leads to get 99% of
research to examine the performance of a single machine
accuracy. Therefore, NSL-KDDTest+ dataset must be used for
learning algorithm or multiple algorithms to all four major attack
model assessment. In this case the performance of the real model
categories rather than to a single attack category. Some of the
can be tested to detect a new type of attack.
algorithms and methods used by the researchers in this filed will
be mentioned. Also, we will try to focus on the researches that Wang, 2017 [15], suggested an SVM based intrusion
used NSL-KDD for analyzing their experimental results. detection technique that considers pre-processing data utilizing
converting the usual attributes by the logarithms of the marginal
Hota and Shrivas, 2014 [10], proposed a model that used
density ratios that exploits the classification information that is
different feature selection techniques to remove the irrelevant
included in each feature. This resulting in data that has high
features in the dataset and developed a classifier that is more
quality and concise which in turn achieved a better detection

49 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

performance in addition to reducing the training time required accurateness and lessen computational expenses in contrary to
for the SVM detection model. the state-of-the-art techniques.
Yin, et al., 2017 [16], have explored how to model an IDS Ikram and Cherukuri,2017 [21], proposed an ID model using
based on deep learning approach using recurrent neural Chi-Square attribute selection and multi-class support vector
networks (RNN-IDS) because of its potential of extracting better machine (SVM). The main idea behind this model is to construct
representations for the data and create better models. They pre- a multi class SVM which has not been adopted for IDS so far to
processed the dataset using Numericalization technique because decrease the training and testing time and increase the individual
the input value of RNN-IDS should be a numeric matrix. The classification accuracy of the network attacks.
results showed that RNN-IDS has great accuracy rate and
detection rate with a low false positive rate compared with In [22], Khammassi and Krichen have applied a wrapper
traditional classification methods. methods based on a genetic algorithm as a search strategy and
logistic regression as a learning algorithm for network IDSs to
Feature selection as a vital part of any IDS can assist make choice the best subset of features. The proposed approach is
the procedure of training the model less multifaceted and faster based on three stages: a pre-processing phase, a feature selection
while preserving or even enhancing the total performance of the phase, and a classification stage the experiment will be
system. Shahbaz et al. [17] suggested an efficient algorithm for conducted on the KDD99 dataset and the UNSW-NB15 dataset.
feature selection by considered the correlation between the The results showed that accuracy of classification equal to 99.90
behavior class label and a subset of attribute to resolve the %, 0.105 % FAR and 99.81% DR with a subset of only 18
problem of dimensionality lessening and to defining good features for the KDD99 dataset. Furthermore, the selected subset
features. The outcomes revealed that the proposed model has provides a good DR for DoS category with 99.98%. The
considerably minimal training time while preserving accuracy obtained results for the UNSW-NB15 provided the lowest FAR
with precision. Additionally, several feature selection methods with 6.39% and a good classification accuracy compared to the
are tested with varying classifiers regarding the detection rate. other mentioned approaches with a subset composed of 20
The comparison outcomes reveal that J48 classifier features.
accomplishes well with the proposed feature selection method.
From this inspiration, we are trying to find out which of
Similarly, the study in [18] proposed a new intelligent IDS classification algorithms that we select will give better results
that works on reduced number of features. First, authors perform after selecting the features that have a strong correlation in the
feature ranking on the basis of information-gain and correlation. training dataset. In this work, researchers will try to conduct
Feature reduction is then done by combining ranks obtained some experiments to differentiate and discover the normal and
from both information gain and correlation using a novel abnormal behavior.
approach to identify useful and useless features. These reduced
features are then fed to a feed forward neural network for III. PROPOSED IDS METHODOLOGY
training and testing on KDD99 dataset. The method uses pre- The main goal of the research, is to build a framework of
processing to eliminate redundant and irrelevant data from the intrusion detection with minimum number of features in the
dataset in order to improve resource utilization and reduce time dataset. The previous researches showed that only a subset of
complexity. The performance of the feature reduced system is these features is related to ID. So, the aim is to reduce the data
actually better than system without feature reduction. According set dimensionality to build a better classifier in a reasonable
to the feature optimization selection problems of the rare attack time. The proposed approach consists of four main phases: The
categories detection the researchers in [19] used the cascaded first phase is to select the related features for each attack using
SVM classifiers to classify the non-rare attack categories and feature selection method. Then combining the different features
using BN classifiers to classify rare attack categories, combining to obtain the optimal set of features for all attacks. The final set
with cascaded GFR feature selection method (CGFR) The of features is fed to the classification stage. Finally, the model is
experimental results showed that the CGFR feature selection is tested using a test dataset. The framework of the proposed
effective and accurate in IDS. methodology is shown in Fig. 1.
Redundant as well as irrelevant characteristics in data have A. Selecting the Related Features for Each Attack
resulted in a constant problem in network traffic classification.
While the network intrusion system deals with a large
To combat this concern, Ambusaidi et al. [20] offered a
amount of raw data, the feature selection is becoming a basic
supervised filter-based feature selection algorithm that
step in building such system. Feature selection is related to a
methodically picks the ideal feature for categorization. The
number of methods and techniques that are used to eliminate the
Flexible Mutual Information Feature Selection (FMIFS) that has
irrelevant and redundant features. The dimensionality of the data
been proposed to lessen the redundancy among features. FMIFS
set has a big effect in the model complexity that leads to low
is then combined with the Least Square Support Vector Machine
classification accuracy, and high computational cost and time.
based IDS(LSSVM) technique to develop an IDS. The role of
The aim of these methods also is to select the optimal features
the model is appraised by means of three intrusion identification
which will enhance the model’s performance. There are two
datasets, that is to say, KDD Cup 99, NSL-KDD plus Kyoto
general categories of methods for feature selection, filter
2006+ datasets. The appraisal outcomes revealed that
methods and wrapper methods [23]. In the Filter algorithms an
characteristic selection algorithm gives other essential
characteristics for LSSVM-IDS to accomplish enhanced

50 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

Figure 1. Framework of The Proposed Model of IDS

independent measure is utilized (such as, information, distance, The selection features stage for each attack is divided into
or consistency) which are used to estimate the relation of a set of three main steps as follows:
features, while wrapper algorithms use of one of learning
algorithms to make the evaluation of the feature’s value. In this Step1: The training dataset is divided into 22 datasets.
study, Information Gain (IG) will be used to select the subset of Each dataset file contains the records of one attack records
related features. IG is often cost less and faster than the wrapper merged with the normal records. If the whole dataset is used
methods. without splitting, then the selection features method will be
biased to the most frequent attacks. So, this step is essential
Information gain is computed for each individual attribute in to obtain more accurate results.
the training dataset related to one class. If the ranked value is
high that means a feature is highly distinctive this class. Step2: Each file then is used as an input to IG method to
Otherwise if the value is less than the predetermined threshold, select the most relevant features of that attack. For example,
it will be removed from the feature space. To obtain a better the spy attack has the related features ranked as shown in
threshold value, the distribution of the IG values is examined and Table 1.
tested with different threshold values on the training dataset. Step3: A ranked feature list is generated, and according
The IG of a feature t, overall classes is known by equation to some thresholds, a number of features are eliminated.
(1). From the list in Table I, it can be noticed that the most
relevant features for spy attack are features 38 and 39, if we
take the threshold equal to 0.003. So, we can take the best
= log two features and eliminate the others.
TABLE I. SPY RANKED RELATED FEATURES
+ \ log \
Feature
Ranked Value Feature Name
Number
+ ̅ \ ̅ log \ ̅ 0.004029 38 dst_host_serror_rate
0.0036057 39 dst_host_srv_serror_rate
Where: 0.0018171 3 Service
• represents (i) category. 0.0012618 18 num_shells
• P( ): probability that a random instance document 0.0011184 15 su_attempted
belongs to class . 0.0008256 19 num_access_files
0.0001008 2 protocol_type
• P(t) and P( ̅ probability of the occurrence of the
feature w in a randomly selected document.
B. Combining the Different Set of Features for All Attacks
• P( |t): probability that a randomly selected document
belongs to class if document has the feature w. In this step, a combined list of features for all attacks is
generated from the obtained subsets. For some attacks the
• m is the number of classes.
highest rank of the first three features are selected. But for
another set of attacks, like land attack, one feature has been

51 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

taken, since it’s rank is equal to 1, while the ranks for other • Random Forest: is related to a machine learning
features were very low. That means this feature can fully method which makes a combination between decision
discriminate this attack. tree and ensemble methods. The input of the forest that
represent the features are picked randomly to build the
C. Classification of the Training Dataset
composed trees. The generation process of the forest
The final combined subset is used as an input to the constructs a collection of trees with controlled variance.
classification stage. The results of three different classifiers have majority voting or weighted voting can be used to decide
been considered to make the comparative study. These the resulting prediction.
classifiers are J48, Random-Forest (RF) and Partial Decision
List (PART). After conducting the experiments, the best two • Partial Decision List (PART): PART is an algorithm
classifiers results are chosen. The next step, is to use the vote of decision–list based on partial decision tree, joining
ensemble method to enhance the performance of the model. the advantages of both classifier C4.5 and PIPPER. A
pruned decision tree is created for all existing instances,
• J48 classifier: C4.5 (J48) is an algorithm developed by for the leaf node building a rule corresponding with the
Ross Quinlan that used to generate a decision tree. This largest coverage, after that discarding the tree and
algorithm becomes a popular in classification and Data continuing.
Mining. The gain ratio method is used in this algorithm
as a criterion for splitting the data set. Some • Ensemble classifier: An ensemble classifier consists of
normalization techniques are applied to the information the combination of multiple weak machine learning
gain using a “split information” value. algorithms (known as weak learners) to improve the
classification performance. The combination of weak
67,343

D I ST R I B U T I O N O F AT TAC KS I N N S L - K D DT R A I N +
41,214

3,633

3,599

2,931

2,646

1,493

956

892

890

201

53

30

20

18

11

10

2
Figure 2. Distribution of Attacks in NSL-KDDTrain+

D I ST R I B U T I O N O F AT TAC KS I N N S L - K D DT EST +
9,711
4,657
1,231
996
944
737
735
685
665
359
331
319
293
S N M P G E T A T T A … 178
157
141
133
73
41
B U F F E R _ O V E R … 20
18
17
15
14
13
13
12
9
7
4
3
2
2
2
2
2
2
1
PS
NMAP

NAMED

IMAP
APACHE2

MAILBOMB
BACK
SNMPGUESS

SENDMAIL
HTTPTUNNEL

XLOCK

PHF

SQLATTACK
NORMAL
NEPTUNE

MSCAN

SATAN
PROCESSTABLE

SAINT

PORTSWEEP
IPSWEEP

POD

LAND

FTP_WRITE
ROOTKIT
MULTIHOP

XTERM
TEARDROP

XSNOOP

WORM

PERL
UDPSTORM
WAREZMASTER

SMURF

LOADMODULE
GUESS_PASSWD

Figure 3. Distribution of Attacks in NSL-KDDTest+

52 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

learners can be based on different strategies such as


majority vote, boosting, or bagging. = × %

D. Testing the Model


In this stage, a test dataset KDD-Test is used to evaluate the Where:
model which has been generated by the vote ensemble method. • TP: related to the true positive.
The test dataset file is different from the training dataset and has
• FP: related to the false positive.
an extra number of attacks. After that the performance
evaluation of the model is conducting using some measures such • FN: related to the false negative.
as accuracy, and area under the ROC. C. Results Analysis
IV. RESULTS AND ANALYSIS After making many experiments on the combined list. The
optimal number of combined features is equal to 28 features.
In this section, experiments results analysis is discussed. All These features as well as its number in the DS are listed in Table
experiments were conducted using platform of Windows with III.
configuration of Intel® core™ i7 CPU 2.70 GHZ, 8 GB RAM.
WEKA tool was used to evaluate the method and perform TABLE III. THE FINAL SELECTED FEATURES
feature selection. In order to select the optimal training
parameters, a 10-fold cross validation (CV) is performed on the
training dataset. Feature Number Feature Name

A. Dataset Description 1 duration


2 protocol_type
All experiments are carried out on NSL-KDD datasets [8].
NSL-KDD is a refined version of the KDD’99 dataset. It 3 services
overcomes some inherent problems in the original KDD dataset. 4 flag
Redundant records in the training set have been removed so that 5 src_bytes
the classifiers produce unbiased results. There is no duplicate 6 dst_bytes
data in the improved testing set. Therefore, the biased influence 7 land
on the performance of the learners has been significantly 8 wrong_fragment
reduced. Each connection in this dataset contains 41 features. 9 urgent
Researchers in this work carry out the experiments using the 10 hot
KDDTrain and KDDTest data. The different attacks are listed in
11 num_failed_logins
Table II. The Distribution of Attacks in NSL-KDDTrain+ and
NSL-KDDTest+ files are shown in Fig 2 and Fig 3. 13 Num_compromised
14 Root_shell
TABLE II. ATTACKS IN NSL_KDD TRAINING DATASET
17 num_file_creations
18 num_shells
Attack Type Attack Name
19 num_access_files
DOS Neptune, Smurf, Pod, Teardrop, Land, Back
26 srv_serror_rate
Probe Port-sweep, IP-sweep, Nmap, Satan 29 same_srv_rate
R2L Guess-password, Ftp-write, Imap, Phf, 30 diff_srv_rate
Multihop, spy, warezclient, Warezmaster
31 srv_diff_host_rate
U2R Buffer-overflow, Load-module, Perl, Rootkit 32 dst_host_count
33 dst_host_srv_count
B. Evaluation Metrics 34 dst_host_same_srv_rate
The performance evaluation of the proposed model, used 36 dst_host_same_src_port_ra
different performance metrics such as: precision (equation 2), 37 dst_host_srv_diff_host_rat
recall (equation 3), F-measure (equation 4), true negative rate, 38 dst_host_serror_rate
false positive rate and overall accuracy (ACC) (equation 5) that 39 dst_host_srv_serror_rate
known as correctly classified instances (CC). In addition, 41 dst_host_srv_rerror_rate
presented Received Operating Characteristics (ROC) of the
system. The ROC curve is computed by drawing the relation
In Table IV, comparing the accuracy and different evaluation
between true positive rate and false positive rate in y-axis and x-
metrics with two sets of attributes against using the all dataset
axis, respectively.
with 41 attributes according to PART classifier with two test
= option cross validation and NSL-KDD Test +. As observed, for
+
the accuracy is shown. The performance of proposed technique
compared in terms of using cross validation test and testing
=
+ dataset. The result shows that high accuracy with (99.7984%) is
obtained when using set of 19 feature with cross validation test,
× ×
=
+

53 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

while using 28 features, the accuracy is (86.66%) when using


NSL-KDD Test + dataset.
On the other hand, the results of the comparison between the
performance of three classification algorithms with the proposed
method, and both CV and testing are presented in Table V.
As a comparison, we used various popular classifiers
algorithms. These classifiers are J48, Random-Forest (RF) and
Partial Decision List (PART). The highest testing accuracy with
(86.66%) is achieved by PART algorithm, whereas the highest
obtained accuracy from CV with (99.78%) by using RF. Fig. 4
shows a comparison of classification algorithms in term of
accuracy with test option cross validation and NSL-KDD Test Figure 5. Final ROC Area for each Class for CV and NSL-KDDTest+
+. According to these results, the best two classifiers (PART and
RF) have been chosen to manipulate the voting ensemble V. CONCLUSION AND FUTURE WORK
algorithm. Table VI demonstrates the performance of using IDS is used to secure the computer based systems against a
voting learning algorithm for Random Forest and PART to lot of cyber-attacks. Feature selection at the beginning stage of
improve the obtained accuracy for the system of intrusion machine learning approach has proven to enhance the detection
detection. It was noticed that, when Random Forest and PART performance. In the research, we have proposed feature selection
classifiers are used under different combination methods, the approach using information gain methods that was calculated for
accuracy of the model is enhanced. Table VI shows also that the each attack in the NSL-KDD dataset to identify the optimal
accuracy in CV is the same while using the three rules. But when feature set for each presented attack and select these features
the supplied test dataset is being used, a different behavior is according to some thresholds. Then combining the feature list
noticed for the three rules. The best accuracy is achieved when for all attacks. The experiment result shows that the highest
using the product probability rule. Finally, the area under the accuracy obtained when using Random Forest and PART
ROC curves as shown in Fig. 5 is calculated for each attack classifiers under combination methods namely the product
classes in the dataset based on cross validation and NSL-KDD probability rule.
Test. The results also show that, the ROC values for DoS and
probe attacks are almost the same in the two test options, but the As a future work, it is suggested to use the adaptive boost
values fluctuate with R2L and U2R attacks. learning algorithm in the feature selection stage instead of using
IG. This will increase the efficiency of the detection system.
ACCURACY RESULTS OF THREE
CLASSIFIERS USING CROSS VALIDATION
AND NSL-KDDTEST+
(28 ATTRIBUTES)
Testing-ACC Cross Validation
99.85%

99.78%
99.74%

86.66%
86.51%
86.08%

J48 RANDOM-FOREST PART

Figure 4. Accuracy Results of Three Classifiers

TABLE IV. RESULTS WITH DIFFERENT NUMBER OF FEATURES USING PART

Feature Correctly Incorrectly F- ROC


Test Option Accuracy TP FP Precision Recall
set Classified Classified Measure Area
Cross Validation 125719 254 99.7984 % 0.998 0.001 0.998 0.998 0.998 0.999
19
NSL-KDD Test + 16231 2563 86.3627 % 0.864 0.124 0.794 0.864 0.814 0.856
28 Cross Validation 125701 272 99.7841 % 0.998 0.001 0.998 0.998 0.998 0.999
NSL-KDD Test + 16287 2507 86.6606 % 0.867 0.108 0.850 0.867 0.823 0.880
Cross Validation 125714 259 99.7944 % 0.998 0.001 0.998 0.998 0.998 0.999
41
NSL-KDD Test + 16283 2511 86.6394 % 0.866 0.124 0.881 0.866 0.818 0.857

54 https://sites.google.com/site/ijcsis/
ISSN 1947-5500
International Journal of Computer Science and Information Security (IJCSIS),
Vol. 16, No. 2, February 2018

TABLE V. CROSS-VALIDATION AND TEST RESULTS OF THREE CLASSIFIERS

Classifier Correctly Incorrectly F- ROC


Test Option Accuracy TP FP Precision Recall
Name Classified Classified Measure Area
Cross Validation 125644 329 99.7388 % 0.997 0.002 0.997 0.997 0.997 0.999
J48
NSL-KDD Test + 16178 2616 86.0807 % 0.861 0.119 0.774 0.861 0.814 0.840
Random- Cross Validation 125785 188 99.8508 % 0.999 0.001 0.998 0.999 0.998 1.000
Forest NSL-KDD Test + 16259 2535 86.5117% 0.865 0.112 0.831 0.865 0.819 0.943
Cross Validation 125701 272 99.7841 % 0.998 0.001 0.998 0.998 0.998 0.999
PART
NSL-KDD Test + 16287 2507 86.6606 % 0.867 0.108 0.850 0.867 0.823 0.880

TABLE VI. CROSS-VALIDATION AND TEST RESULTS USING VOTE METHOD WITH (RF+PART)
Combination Correctly Incorrectly F- ROC
Test Option Accuracy TP FP Precision Recall
Rule Classified Classified Measure Area
Majority Cross Validation 125743 230 99.8174 % 0.998 0.001 0.998 0.998 0.998 0.999
Voting NSL-KDD Test + 16292 2502 86.6872 % 0.867 0.108 0.850 0.867 0.823 0.847
Product Cross Validation 125737 225 99.8127 % 0.998 0.001 0.998 0.998 0.998 0.999
probability NSL-KDD Test + 16294 2496 86.6979 % 0.867 0.108 0.851 0.867 0.823 0.884
Average Cross Validation 125743 230 99.8174 % 0.998 0.001 0.998 0.998 0.998 1.000
probability NSL-KDD Test + 16292 2502 86.6872 % 0.867 0.108 0.850 0.867 0.823 0.947

no. 4, pp. 89–92, 2016.


[13] M. A. Jabbar and S. Samreen, “Intelligent network intrusion detection
REFERENCES using alternating decision trees,” in 2016 International Conference
on Circuits, Controls, Communications and Computing (I4C), 2016,
[1] P. D ́ıaz-Ram ́ırez, A., Tafoya, L.A., Atempa, J.A., Mej ́ıa-Alvarez, pp. 1–6.
“Wireless sensor networks and fusion information methods for [14] N. Paulauskas and J. Auskalnis, “Analysis of data pre-processing
forest fire detection,” Procedia Technol. 3, pp. 69–79, 2012. influence on intrusion detection using NSL-KDD dataset,” in 2017
[2] A. Isaac, S., Hancke, G., Madhoo, H., Khatri, “A survey of wireless Open Conference of Electrical, Electronic and Information Sciences
sensor network applications from a power utility’s distribution (eStream), 2017, pp. 1–5.
perspective,” AFRICON 2001, pp. 1–5, 2011. [15] H. Wang, J. Gu, and S. Wang, “An effective intrusion detection
[3] B. . Mao, G., Fidan, B., Anderson, “Wireless sensor network framework based on SVM with feature augmentation,” Knowledge-
localization techniques. Computer Networks,” vol. 10, no. 51, pp. Based Syst., vol. 136, no. Supplement C, pp. 130–139, 2017.
2529–2553, 2007. [16] C. Yin, Y. Zhu, J. Fei, and X. He, “A Deep Learning Approach for
[4] V. Durisic, M., Tafa, Z., Dimic, G., Milutinovic, “A survey of military Intrusion Detection Using Recurrent Neural Networks,” IEEE
applications of wireless sensor networks,” in 2012 Mediterranean Access, vol. 5, pp. 21954–21961, 2017.
Conference on Embedded Com- puting, MECO, 2012, pp. 196–199. [17] M. B. Shahbaz, Xianbin Wang, A. Behnad, and J. Samarabandu, “On
[5] L. Afzaal, M., Di Sarno, C., Coppolino, L., D’Antonio, S., Romano, “A efficiency enhancement of the correlation-based feature selection for
resilient architecture for forensic storage of events in critical intrusion detection systems,” 2016 IEEE 7th Annu. Inf. Technol.
infrastructures.,” in 2012 IEEE 14th International Symposium on Electron. Mob. Commun. Conf., pp. 1–7, 2016.
High-Assurance Systems Engineering, HASE, 2012, pp. 48–55. [18] Akashdeep, I. Manzoor, and N. Kumar, “A feature reduced intrusion
[6] D. Wahid, A., Kim, “Connectivity-based routing protocol for detection system using ANN classifier,” Expert Syst. Appl., vol. 88,
underwater wireless sensor networks,” in 2012 International pp. 249–257, 2017.
Conference on ICT Convergence, ICTC, 2012, pp. 589–590. [19] Y. Sun and F. Liu, “A & ascaded ) eature 6 election $ pproach in 1
[7] I. H. Witten, E. Frank, M. A. Hall, and C. J. Pal, Data Mining: Practical etwork , ntrusion â€TM etection,” pp. 119–124, 2015.
machine learning tools and techniques. Morgan Kaufmann, 2016. [20] M. Ambusaidi, X. He, P. Nanda, and Z. Tan, “Building an intrusion
[8] M. Tavallaee, E. Bagheri, W. Lu, and A. A. Ghorbani, “A detailed detection system using a filter-based feature selection algorithm,”
analysis of the KDD CUP 99 data set,” in IEEE Symposium on IEEE Trans. Comput., vol. PP, no. 99, p. 1, 2016.
Computational Intelligence for Security and Defense Applications, [21] I. Sumaiya Thaseen and C. Aswani Kumar, “Intrusion detection model
CISDA 2009., 2009, pp. 1–6. using fusion of chi-square feature selection and multi class SVM,”
[9] P. Institute, “2014 Global report on the cost of cyber crime,” 2014. J. King Saud Univ. - Comput. Inf. Sci., vol. 29, no. 4, pp. 462–472,
[10] H. S. Hota and A. K. Shrivas, “Decision Tree Techniques Applied on 2017.
NSL-KDD Data and Its Comparison with Various Feature Selection [22] C. Khammassi and S. Krichen, “A GA-LR wrapper approach for feature
Techniques,” in Advanced Computing, Networking and Informatics- selection in network intrusion detection,” Comput. Secur., vol. 70,
Volume 1: Advanced Computing and Informatics Proceedings of the pp. 255–277, 2017.
Second International Conference on Advanced Computing, [23] F. Amiri, M. R. Yousefi, C. Lucas, A. Shakery, and N. Yazdani,
Networking and Informatics (ICACNI-2014), 2014, pp. 205–211. “Mutual information-based feature selection for intrusion detection
[11] D. H. Deshmukh, T. Ghorpade, and P. Padiya, “Intrusion detection systems,” J. Netw. Comput. Appl., vol. 34, no. 4, pp. 1184–1199,
system by improved preprocessing methods and Na #x00EF;ve 2011.
Bayes classifier using NSL-KDD 99 Dataset,” in 2014 International
Conference on Electronics and Communication Systems (ICECS),
2014, pp. 1–7. AUTHORS PROFILE
[12] I. M. Y. Noureldien A. Noureldien, “Accuracy of Machine Learning
Algorithms in Detecting DoS Attacks Types,” Sci. Technol., vol. 6, Authors Profile …

55 https://sites.google.com/site/ijcsis/
ISSN 1947-5500

You might also like