0% found this document useful (0 votes)
12 views6 pages

Ayub 2020

This paper presents a study on the vulnerabilities of Intrusion Detection Systems (IDS) using adversarial machine learning, specifically focusing on a Model Evasion Attack against a Multilayer Perceptron (MLP) network. The authors demonstrate that adversaries can significantly reduce the accuracy of IDS by crafting adversarial samples that the model misclassifies as benign. Experimental results indicate that the MLP model achieved high accuracy but is still susceptible to evasion attacks, highlighting the need for improved defenses against such threats.

Uploaded by

jawad hamza
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
12 views6 pages

Ayub 2020

This paper presents a study on the vulnerabilities of Intrusion Detection Systems (IDS) using adversarial machine learning, specifically focusing on a Model Evasion Attack against a Multilayer Perceptron (MLP) network. The authors demonstrate that adversaries can significantly reduce the accuracy of IDS by crafting adversarial samples that the model misclassifies as benign. Experimental results indicate that the MLP model achieved high accuracy but is still susceptible to evasion attacks, highlighting the need for improved defenses against such threats.

Uploaded by

jawad hamza
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 6

2020 54th Annual Conference on Information Sciences and Systems (CISS)

Model Evasion Attack on Intrusion Detection


Systems using Adversarial Machine Learning
Md. Ahsan Ayub∗ , William A. Johnson† , Douglas A. Talbert‡ , and Ambareen Siraj§
Department of Computer Science
2020 54th Annual Conference on Information Sciences and Systems (CISS) 978-1-7281-4085-8/20/$31.00 ©2020 IEEE 10.1109/CISS48834.2020.1570617116

Tennessee Technological University


Cookeville, USA
{mayub42∗ , wajohnson43† }@students.tntech.edu
{dtalbert‡ , asiraj§ }@tntech.edu

Abstract—Intrusion Detection Systems (IDS) have a long his- deviate from normal profile [10], [39]. In our study, we focus
tory as an effective network defensive mechanism. The systems on network-based IDS, also referred as NIDS, and Artificial
alert defenders of suspicious and / or malicious behavior detected Neural Network (ANN) as our machine learning algorithm.
on the network. With technological advances in AI over the
past decade, machine learning (ML) has been assisting IDS Despite success of machine learning for intrusion detection,
to improve accuracy, perform better analysis, and discover the advent of Adversarial Machine Learning has recently
variations of existing or new attacks. However, applications of ML emerged as a significant threat to the effectiveness of such
algorithms have some reported weaknesses and in this research, applications. An adversary can exploit vulnerabilities in the
we demonstrate how one of such weaknesses can be exploited
against the workings of the IDS. The work presented in this paper
machine learning algorithm itself or the trained ML model
is twofold: (1) we develop a ML approach for intrusion detection to compromise network defense [16]. There are various ways
using Multilayer Perceptron (MLP) network and demonstrate this can be achieved, such as, Membership Inference Attack
the effectiveness of our model with two different network-based [36], Model Inversion Attack [11], Model Poisoning Attack
IDS datasets; and (2) we perform a model evasion attack against [25], Model Extraction Attack [42], Model Evasion Attack [3],
the built MLP network for IDS using an adversarial machine
learning technique known as the Jacobian-based Saliency Map
Trojaning Attack [22], etc. The range of these attacks typically
Attack (JSMA) method. Our experimental results show that depends on the level of access an adversary has to the trained
the model evasion attack is capable of significantly reducing model. For example, an adversary may have perfect knowledge
the accuracy of the IDS, i.e., detecting malicious traffic as about the type of the model used as well as its workings or
benign. Our findings support that neural network-based IDS is s/he may have no knowledge about the model at all. Our focus
susceptible to model evasion attack, and attackers can essentially
use this technique to evade intrusion detection systems effectively.
in this research is on demonstrating a Model Evasion Attack
for IDS whereby an adversary can evade the ML model for
Index Terms—Adversarial Machine Learning, Evasion Attack, network-based IDS by crafting adversarial samples. If s/he
Intrusion Detection System, Neural Network is successful, the attacker may be able to gain access to the
network with malicious traffic and cause significant harm.
I. I NTRODUCTION
The following are the main contributions of the paper:
The use of Machine Learning (ML) in Intrusion Detection
System (IDS) is widespread and has demonstrated remarkable • We construct a Multilayer Preceptron (MLP) model, a
performance as a robust and effective defense mechanisms popular Neural Network topology, to perform binary
[37], [41]. An IDS provides detection capabilities over ma- classification over benign and attack traffic in a network-
licious traffic by generating alerts with network logs such based anomaly IDS. In our experiments, we achieve more
that further intelligence can be derived as and when ned- than 99% accuracy for all experimental datasets used.
eded. Based on its placement in the network infrastructure, • We demonstrate Model Evasion Attack against the built
a network-based IDS monitors the communication that trav- MLP model in a white-box setting, where the accuracy
els into and out of the network, while a host-based IDS of the attacked model drops significantly and discuss
scans a particular host (e.g., server), to notify the network possible countermeasures to prevent this type of attacks.
administrator for possible security threats. The two types of The rest of the paper is organized as follows: Section 2
IDSs are (1) Signature-based IDS that analyze network traffic provides an overview of the Model Evasion Attack. Section 3
is for known malicious signatures and (2) Anomaly-based first explains the datasets we use in our experiments, followed
IDS that compares the network traffic against a user’s known by the evaluation of our twofold research by discussing the
patterns and raises an alert if it deviates from the pattern. construction of a Multilayer Perceptron (MLP) network as
Researchers have leveraged various ML based classifiers, such well as its classification results and then describing the design
as, Artificial Neural Networks, Decision Trees, Support Vector of our attack and its effectiveness. We conclude this section
Machine (SVM), Fuzzy Logic, and Bayesian Networks to with a description of some possible countermeasures against
detect malicious traffic as well as discover unseen attacks that this attack. Section 4 provides an overall discussion of our

978-1-7281-4085-8/20/$31.00 ©2020 IEEE

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.
experiments, followed by relevant work in section 5. Section the weights within the model. As in the case of the Black Box
6 summarizes the paper, its contributions, and future work to model, the adversary can query the target model such that any
further improve upon this research. information the model leaks, can potentially be exploited.
In all cases, the adversary is only able to alter the data
II. M ODEL E VASION ATTACK
during the testing phase of the classification. Our attack
The goal of Model Evasion Attack is to cause the machine depends on the knowledge about certain parameters used in
learning model to misclassify observations during the testing the trained model but does not depend on the training dataset.
phase (as shown in Fig. 1). Applied to a network-based Because of this, we consider our attack scenario to be white
IDS, an adversary attempts to evade detection by altering the box. We also assume that the attacker is able to modify the
malicious instances in such a way that the IDS misclassifies test instance in such a way as to modify any of the features
this behavior as benign. To elaborate, there are four different seen by the IDS. Additionally, the general intuition behind
ways this can take place [26]: Confidence Reduction, where performing a successful model evasion attack is to define a
reducing the confidence score output leads to misclassifica- loss function that the adversary aims to maximize or minimize
tion; Misclassification, where an adversary tries to alter the for each sample to results in misclassification [3], [19]. It is
correct output classification to a class than the original class; important to note that, our experimentation is based on a white-
Target Misclassification (our approach), where the adversary box setting and do not evaluate the attack’s effectiveness based
produces a sample that fools the model to classify the behavior on other attack scenarios (e.g., black-box setting).
as a target class; and Source / Target Misclassification, where
the attacker makes the output class classification of a specific III. E VALUATION
adversarial sample to be a specific target class.
We first describe the datasets that we use for evaluation,
followed by the description of the target model and our
Testing Phase experimental setup. We then present the results of model
adv(X)
evasion attacks against different datasets.

Misclassification A. Dataset
Adversarial Adversary
Samples CICIDS 2017. Released by the Canadian Institute for Cy-
bersecurity in 2017, this dataset closely resembles real-world
data [35]. The IDS logs were recorded over five days with a
total 51.1 GB of packet capture (PCAP) files1 built upon the
Training Machine Learning
Dataset Model abstract behavior of 25 users based on the HTTP, HTTPS, FTP,
test(X)
SSH, and email protocols. With 12 different victim machines
Classification
and 2 attacker machines, this labelled dataset features common
Legitimate
Samples User
attacks, such as, Web based Brute Force, XSS and SQL
Injection, DoS, DDoS, Infiltration, Heart-bleed, Bot, and Scan.
The dataset is available online for public use2 .
Fig. 1. Diagram of the model evasion attack against a trained machine learning
For our experiments, we sampled 950, 000 records in total.
model during the testing phase. Each record consists of 80 continuous features with a binary
labelled class of benign or attack.
To clarify the capabilities of an adversary, we outline three
different knowledge scenarios [29] - TRAbID 2017. Viegas et al. [43] produced a network based
intrusion database in a controlled and reproducible environ-
White Box. The adversary has perfect knowledge of the target ment. To depict a real-world use case, the dataset includes
classification model including the type of the classifier used client-server communication. Legitimate traffic was generated
and its structure. S/he also knows all parameters of the model by the client requesting services available in the server, such
that are required to perform prediction as well as all or part as, HTTP, SMTP, SSH, SNMP, and DNS, while the attacker
of the training dataset and its features. from a client machine launched attacks to the same server.
Black Box. The adversary has zero knowledge of the target The type of the attacks primary included different categories
model. It may be known that the model performs classification, of DoS (e.g., SYN flood, ICMP flood, etc.) and Scan (e.g.,
but the adversary does not have access to the training data, SYN scan, ACK scan, etc.).
model structure or type, or any parameters of the model. S/he For our experimentation, we collected 18, 000 records in
is able to offset this lack of information by querying the model total. Unlike CICIDS 2017, each record in this case consists
for potential information leakage. of 43 continuous features. The dataset is also labelled with
a binary class of benign or attack and is available online
Grey Box. The adversary has an incomplete knowledge of the
target model and knows the features considered by the model 1 https://fileinfo.com/extension/pcap

and its type. S/he does not have any part of the training set or 2 https://www.unb.ca/cic/datasets/ids-2017.html

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.
2020 54th Annual Conference on Information Sciences and Systems (CISS)

for public use3 . It is important to note that both datasets are the feature values are in the range of 0 to 1, which is also
balanced and suitable for binary classification tasks. referred as MinMax Scalar. Then, we split the data in 80%
training and 20% testing instances. Thus, we have approx-
B. Building the ML Model for IDS
imately 760, 000 and 14, 400 training records from CICIDS
We now describe the construction of the target machine and TRAbID dataset respectively. Additionally, we perform
learning model for network-based intrusion detection system stratified splits of training and testing instances to preserve
(IDS) and its performance in detecting attack traffic. the same percentage for each target class as in the complete
Multilayer Perceptron (MLP) Network. We use a Multilayer set provided in the dataset.
Perceptrons (MLP) network, which is a widely used Neural Experimental Results. In this section, we describe our ex-
Network topology. With w as the real vector of weights, b ∈ R perimental results in terms of Accuracy, which denotes the
as the bias, and h as the transfer function, the decision function extent of correct predictions by the model; Precision, which
of the MLP can be formally defined: is a measurement of the ratio of the true positive records to all
f (x) = w · h(vi · x + di ) + b positively labelled instances; Recall, which is the ratio of the
true positive instances to all instances that should have been
where, (vi , di ) ∈ Rn × R is a representation of the weight labelled positive; and F1 score, which is the harmonic mean
of the i-th hidden unit [8]. An MLP network is usually of precision and recall [30].
constructed with three or more layers, that is, one input layer, P
True Positive
one or more hidden layer, and one output layer. We select one Precision Score = P P
hidden layer and build a fully connected network (i.e., each True Positive + False Positive
node in the input layer is connected with a certain weight to
P
True Positive
every node in the hidden layer) as shown in Fig. 2. Recall Score = P P
True Positive + False Negative
Precision Score × Recall Score
w11
F1 Score = 2 ·
x1 h1 Precision Score + Recall Score
1
w2
For both datasets, MLP model was succesfull in accurately
31
w

detecting benign traffic as well as attack traffic. Fig. 4 shows


w1
0

w
12
that the model performed at 99.5% accuracy for CICIDS 2017
w22 w20 dataset while at 99.8% accuracy for TRAbID 2017 dataset.
x2 h2 O
To further illustrate our results with the described param-
32
w

30
w

eters (i.e., precision, recall, and F1 ) scores, Table I shows


that the obtained results are very close in all settings. We
w 13 3
w2

w33
demonstrate the performance of benign and attack traffic
x3 h3 prediction individually with weighted average.

Fig. 2. A simple Multilayer Perceptron (MLP) network with one input layer TABLE I
X = (x1 , x2 , x3 ), one hidden layer H = (h1 , h2 , h3 ), and one output O. P ERFORMANCE OF THE MLP N ETWORK ON THE L EGITIMATE I NSTANCES .
Precision Recall F1 Support
Particular
CICIDS TRAbID CICIDS TRAbID CICIDS TRAbID CICIDS TRAbID
We use a Rectified Linear Unit (ReLU) activation function Benign 0.9957 0.9978 0.9951 0.9973 0.9954 0.9975 110, 246 1, 832
Attack 0.9932 0.9973 0.9951 0.9978 0.9954 0.9975 79, 553 1, 8317
[32], where the function and its derivative are monotonic with Weighted Avg. 0.9946 0.9975 0.9946 0.9975 0.9945 0.9975 189, 799 3, 663
a range of 0 to ∞, in the input layer as well as hidden layer
of the MLP network. Since we derive the probability value of We train both model for 100 epochs with a batch size of 64.
the binary class in the output layer, we use Sigmoid activa- From Fig. 3, we notice that the MLP network tends to overfit
tion function, where the S-shaped function is differentiable after 13 epochs for CICIDS 2017 and 16 epochs for TRAbID
with a range of 0 and 1 [40]. To compile the model, we 2017. However, we preserve the generalization of the model
use an Adam optimization algorithm, a first-order gradient- by triggering an early stopping task based on the validation
based optimization of stochastic objective functions [18], and loss. Additionally, we define a delay of 2 epochs to verify that
Binary Cross Entropy (BCE) loss function [21]. To control the there were no signs of improvement after the initial indication.
generalization ability of a perception, we incorporate the use
of early stopping during training. We then monitor validation C. Adversarial Machine Learning to Evade IDS
loss to trigger this action. We also select 10% of the training In this section, we report on how we have used the built
records as the validation set to perform this task. MLP model as our target model to launch the model evasion
We base our analysis on both the CICIDS 2017 and attack using Adversarial Machine Learning.
TRAbID 2017 datasets and tune our model similar to what
Attack Design. The purpose of the IDS evasion attack is to
we discussed before. We first scale the dataset so that all
generate data samples in such a way that it confuses the trained
3 https://secplab.ppgia.pucpr.br/?q=trabid MLP model to classify malicious data as benign. In other

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.
2020 54th Annual Conference on Information Sciences and Systems (CISS)

Accuracy over Epoch Loss over Epoch Accuracy over Epoch 0.40
Loss over Epoch
0.12 Train 1.00 Train
Validation Validation

0.99 0.35

0.10 0.98
0.30

0.98 0.25
0.08 0.96
Accuracy

Accuracy
Loss

Loss
0.20

0.97 0.06
0.94
0.15

0.04 0.10
0.96 0.92

0.05
Train 0.02 Train
Validation 0.90 Validation
0.00
0 1 2 3 4 5 6 7 8 9 10 11 12 13 0 1 2 3 4 5 6 7 8 9 10 11 12 13 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
Epoch Epoch Epoch Epoch

Fig. 3. Accuracy over Epoch and Loss over Epoch curves for the MLP generated on CICIDS 2017 (on left) and TRAbID 2017 dataset (on right).

performance drop in terms of accuracy for CICIDS 2017 and


TRAbID 2017 dataset respectively. This demonstrates that
the trained MLP model fails to correctly predict adversarial
samples as attack traffic. It also signifies that an adversary
can craft a malicious network traffic such that the network-
based IDS classifies it as benign. Thus, the defense mechanism
is successfully evaded by exploiting the legitimated traffic
communication between the client and the server.

TABLE II
P ERFORMANCE OF M ODEL E VASION ATTACK ON THE MLP N ETWORK .
Accuracy Accuracy Performance
Dataset # Instances
Legitimate Instances Adversarial Instances Drop
(a) CICIDS 2017 (b) TRAbID 2017 CICIDS 190, 291 99.5% 78.09% 21.52%
TRAbID 3, 694 99.8% 69.99% 29.87%
Fig. 4. Confusion matrix graphs for MLP generated on two different datasets.

It is worthwhile to mention that there are other methodolo-


words, we aim to inject adversarial instances to the model gies to simulate similar types of attacks (e.g., Fast Gradient
during testing and obtain benign outputs. More formally, we Sign Method (FGSM) [12]); however, the JSMA method
desire to craft adversarial sample X ∗ by adding a perturbation possesses the ability to generate adversarial samples with a
to the legitimate sample X (i.e., X ∗ ← X + δX, where δ is lesser degree of distortion. Additionally, the FSGM method
denoted as perturbation), such that F (X ∗ ) = Y ∗ 6= Y . has been found to be ill-suited for the IDS setting [33].
In this study, we base our attack analysis on the assumption Implementation. We implement our MLP model using Keras
that the only knowledge the adversary has is about the pa- [7], utilizing the Python machine learning tool Scikit-Learn
rameters the model uses to predict the binary class (benign or [28] to execute data processing tasks, and used Matplotlib
attack). This is close to a real-world setting because it is not library [17] to generate all the graphs in this paper. To
trivial for the adversary to easily infer knowledge about the implement the attack, we used CleverHans, a Python library
model architecture trained for network-based IDS. On a related to benchmark machine learning systems’ vulnerability to ad-
note, Tramèr et al. [2016] did show that it is possible to extract versarial examples [24]. Finally, our implementation of the
internal information of the MLP network architecture [42]. attack was tested with TensorFlow version 1.13.1 [1]. Our
Our attack design is primarily focused on creating adversar- implementation has been made open source for the community
ial test samples based on Jacobian-based Saliency Map Attack with MIT License and is available online4 .
(JSMA) [26]. With only a small perturbation in the legiti-
mate data samples, the JSMA generates adversarial samples Countermeasures. Following is a discussion of a few potential
based on the Saliency Map method, as described in [38]. We countermeasures to prevent the model evasion attack that we
leverage the saliency map to search through the legitimate test used on our MLP network. Papernot et al. [2016] leveraged the
instances, closely observe the sensitivity information to choose distillation method, introduced by Hinton et al. [15], to reduce
a perturbation δX among the input dimensions that is likely the size of the deep neural network and thus the computing
to fool the built MLP model, and make iterative changes to resources, as a defense strategy to reduce the network’s
produce an adversarial sample set. Following these steps, we vulnerability to adversarial sample generation [27]. The paper
exploit legitimate test instances collected from both datasets. pointed out that reducing the amplitude of the adversarial
gradient would enable the model to generalize better since
Attack Effectiveness. We evaluate the effectiveness of our at- crafting adversarial samples becomes easier when adversarial
tack by testing the trained MLP model with the adversarial test
samples. As shown in Table II, we notice 21.52% and 29.87% 4 https://github.com/TnTech-CEROC/adversarial ml ids

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.
2020 54th Annual Conference on Information Sciences and Systems (CISS)

gradients are high. Thus, this increases the resilience of the typically include Support Vector Machine (SVM) with dif-
model to adversarial samples. Other mitigation techniques for ferentiable kernels and neural networks. Other methodologies
model evasion attack include methods to tighten the decision have been devised to perform attacks on a broader spectrum
boundary of a classification algorithm so that benign features of classification algorithm, including [45] which uses genetic
cannot easily be applied to malicious samples [3], training the algorithms, and in [26] which devises a Forward Derivative
classifier over malicious samples generated with adversarial based on Jacobian Matrices.
knowledge [34], and removing features from the model that We demonstrate the model evasion attack on the network-
are not immediately necessary [2], [46]. One of our future based IDS; however, there are other applications where this
work will be to validate and compare the effectiveness of such type of attack was employed such as, spam filtering [13], [20],
countermeasures applied to our research. [23], visual recognition [5], [14], [31], and malware detection
[3], [9], [44]. This research adds intrusion detection to the list.
IV. D ISCUSSION
VI. C ONCLUSION
We evaluate our MLP network trained over two different
datasets and are confident that its accuracy with other network- In our study, we first build a supervised machine learning
based IDS dataset will be similarly effective. We then analyze model to detect and classify benign and attack traffic using
model evasion attack on the presented Multilayer Perceptron two different network-based intrusion detection system (IDS)
(MLP) model for the target class misclassification. In this way, datasets: CICIDS 2017 [35] and TRAbID 2017 [43]. We
we fool the model into misclassifing attack records as benign construct a Multilayer Perceptron (MLP) network to perform
and hence evade the network defense. Reducing the output binary classification task and achieved outstanding detection
confidence of the predicted class by the classifier would be results, i.e., 99.5% and 99.8% accuracy for CICIDS and
another avenue to perform this kind of evasion attack, which TRAbID, respectively. We then apply model evasion attack
we leave for future work. from the adversarial machine learning suite to demonstrate that
We base our experimentation on white-box setting and do it is possible to evade intrusion detection systems effectively.
not evaluate the attack’s effectiveness based on other attack We consider the trained MLP model as our target model. To
scenarios (e.g., black-box setting). We validate the effective- implement our attack, we select the Jacobian-based Saliency
ness of our attack in terms of the drop in accuracy when tested Map Attack (JSMA) method in a white-box setting, where an
with the adversarial samples crafted with small perturbations adversary has perfect knowledge over the parameters required
on the input dimensions and the Jacobian-based Saliency Map for the model to perform prediction. In other words, the
Attack (JSMA) method. Another avenue of our future work adversary crafts adversarial samples with small perturbation to
includes gathering knowledge of sensitive features that cause the legitimate testing samples and attempts to fool the model
flipping the prediction of the class. during the testing phase. We demonstrate success in our attack
and validated its effectiveness in terms of 22.52% and 29.87%
V. R ELEVANT W ORK accuracy drop in performance for CICIDS and TRAbID,
Adversarial machine learning has become a topic of much respectively. This signifies that evading network defense is
interest in the cybersecurity space. This is largely because possible without much effort unless proper countermeasures
classification algorithms have been successful in solving the are undertaken, as discussed.
problems of malware detection. There are different kinds of ACKNOWLEDGEMENT
adversarial machine learning techniques that allow attackers
The work reported in this paper has been entirely supported
to subvert these classification algorithms in malicious ways.
by Cybersecurity Education, Research & Outreach Center
One such methodology is known as model evasion attack that
(CEROC) at Tennessee Technological University.
allows an adversary to alter an adversarial sample such that it
is misclassified as benign. Pitropakis et al. [2019] provided R EFERENCES
a detailed taxonomy on the model evasion attack to well [1] Martin Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis,
understand different types of the applications, the architecture Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving,
of the models, and the used datasets [29]. Michael Isard, Manjunath Kudlur, Josh Levenberg, Rajat Monga, Sherry
Moore, Derek G. Murray, Benoit Steiner, Paul Tucker, Vijay Vasudevan,
Model evasion attack is often done via gradient descent Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. Ten-
over the discrimination function of the classifier [3], [4], sorflow: A system for large-scale machine learning. In 12th USENIX
[6]. By applying gradient descent over the discrimination Symposium on Operating Systems Design and Implementation (OSDI
16), pages 265–283, 2016. URL: https://www.usenix.org/system/files/
function of the classifier, these methodologies are able to conference/osdi16/osdi16-abadi.pdf.
identify traits of benign samples, such that these traits may be [2] Arjun Nitin Bhagoji, Daniel Cullina, Chawin Sitawarin, and Prateek
applied to malicious samples to force misclassification. Much Mittal. Enhancing robustness of machine learning systems via data
transformations. In 2018 52nd Annual Conference on Information
like [33], we also leverage the Jacobian-based Saliency Map Sciences and Systems (CISS), pages 1–5. IEEE, 2018.
Attack (JSMA) method to evade the model’s detection using [3] Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim
different datasets. Gradient descent methodologies are not Šrndić, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. Evasion attacks
against machine learning at test time. In Joint European conference on
without their weaknesses. They specifically target classifiers machine learning and knowledge discovery in databases, pages 387–
with differentiable discrimination functions. Such classifiers 402. Springer, 2013.

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.
2020 54th Annual Conference on Information Sciences and Systems (CISS)

[4] Battista Biggio and Fabio Roli. Wild patterns: Ten years after the rise of [26] Nicolas Papernot, Patrick McDaniel, Somesh Jha, Matt Fredrikson,
adversarial machine learning. Pattern Recognition, 84:317–331, 2018. Z Berkay Celik, and Ananthram Swami. The limitations of deep learning
[5] Nicholas Carlini and David Wagner. Adversarial examples are not in adversarial settings. In 2016 IEEE European Symposium on Security
easily detected: Bypassing ten detection methods. In Proceedings of and Privacy (EuroS&P), pages 372–387. IEEE, 2016.
the 10th ACM Workshop on Artificial Intelligence and Security, pages [27] Nicolas Papernot, Patrick McDaniel, Xi Wu, Somesh Jha, and Anan-
3–14. ACM, 2017. thram Swami. Distillation as a defense to adversarial perturbations
[6] Lingwei Chen, Yanfang Ye, and Thirimachos Bourlai. Adversarial ma- against deep neural networks. In 2016 IEEE Symposium on Security
chine learning in malware detection: Arms race between evasion attack and Privacy (SP), pages 582–597. IEEE, 2016.
and defense. In 2017 European Intelligence and Security Informatics [28] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion,
Conference (EISIC), pages 99–106. IEEE, 2017. O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vander-
[7] François Chollet. Keras, 2016. URL: https://github.com/fchollet/keras. plas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duch-
[8] Ronan Collobert and Samy Bengio. Links between perceptrons, mlps esnay. Scikit-learn: Machine learning in Python. Journal of Machine
and svms. In Proceedings of the twenty-first international conference Learning Research, 12:2825–2830, 2011.
on Machine learning, page 23. ACM, 2004. [29] Nikolaos Pitropakis, Emmanouil Panaousis, Thanassis Giannetsos, Eleft-
[9] Luca Demetrio, Battista Biggio, Giovanni Lagorio, Fabio Roli, and herios Anastasiadis, and George Loukas. A taxonomy and survey of
Alessandro Armando. Explaining vulnerabilities of deep learning to attacks against machine learning. Computer Science Review, 34:100199,
adversarial malware binaries. arXiv preprint arXiv:1901.03583, 2019. 2019.
[10] Bo Dong and Xue Wang. Comparison deep learning method to [30] David Martin Powers. Evaluation: from precision, recall and f-measure
traditional methods using for network intrusion detection. In 2016 to roc, informedness, markedness and correlation. 2011.
8th IEEE International Conference on Communication Software and [31] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised
Networks (ICCSN), pages 581–585. IEEE, 2016. representation learning with deep convolutional generative adversarial
[11] Matt Fredrikson, Somesh Jha, and Thomas Ristenpart. Model inversion networks. arXiv preprint arXiv:1511.06434, 2015.
attacks that exploit confidence information and basic countermeasures. [32] Prajit Ramachandran, Barret Zoph, and Quoc Le. Searching for
In Proceedings of the 22nd ACM SIGSAC Conference on Computer and activation functions. 2018. URL: https://arxiv.org/pdf/1710.05941.pdf.
Communications Security, pages 1322–1333. ACM, 2015. [33] Maria Rigaki. Adversarial deep learning against intrusion detection
[12] Ian Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining classifiers, 2017.
and harnessing adversarial examples. In International Conference on [34] Paolo Russu, Ambra Demontis, Battista Biggio, Giorgio Fumera, and
Learning Representations, 2015. URL: http://arxiv.org/abs/1412.6572. Fabio Roli. Secure kernel machines against evasion attacks. In
[13] Michael Großhans, Christoph Sawade, Michael Brückner, and Tobias Proceedings of the 2016 ACM workshop on artificial intelligence and
Scheffer. Bayesian games for adversarial regression problems. In security, pages 59–69. ACM, 2016.
International Conference on Machine Learning, pages 55–63, 2013. [35] Iman Sharafaldin, Arash Habibi Lashkari, and Ali A Ghorbani. To-
[14] Jamie Hayes and George Danezis. Machine learning as an adversar- ward generating a new intrusion detection dataset and intrusion traffic
ial service: Learning black-box adversarial examples. arXiv preprint characterization. In ICISSP, pages 108–116, 2018.
arXiv:1708.05207, 2, 2017. [36] Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov.
[15] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge Membership inference attacks against machine learning models. In 2017
in a neural network. arXiv preprint arXiv:1503.02531, 2015. IEEE Symposium on Security and Privacy (SP), pages 3–18. IEEE, 2017.
[16] Ling Huang, Anthony D Joseph, Blaine Nelson, Benjamin IP Rubinstein, [37] Nathan Shone, Tran Nguyen Ngoc, Vu Dinh Phai, and Qi Shi. A deep
and J Doug Tygar. Adversarial machine learning. In Proceedings of the learning approach to network intrusion detection. IEEE Transactions on
4th ACM workshop on Security and artificial intelligence, pages 43–58. Emerging Topics in Computational Intelligence, 2(1):41–50, 2018.
ACM, 2011. [38] Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside
[17] J. D. Hunter. Matplotlib: A 2d graphics environment. convolutional networks: Visualising image classification models and
Computing in Science & Engineering, 9(3):90–95, 2007. saliency maps. arXiv preprint arXiv:1312.6034, 2013.
http://dx.doi.org/10.1109/MCSE.2007.55 doi:10.1109/MCSE. [39] Jayveer Singh and Manisha J Nene. A survey on machine learning
2007.55. techniques for intrusion detection systems. International Journal of
[18] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic Advanced Research in Computer and Communication Engineering,
optimization. arXiv preprint arXiv:1412.6980, 2014. 2(11):4349–4355, 2013.
[19] Pavel Laskov and Marius Kloft. A framework for quantitative security [40] Maxwell Stinchcombe and Halbert White. Universal approximation
analysis of machine learning. In Proceedings of the 2nd ACM workshop using feedforward networks with non-sigmoid hidden layer activation
on Security and artificial intelligence, pages 1–4. ACM, 2009. functions. In IJCNN International Joint Conference on Neural Networks,
[20] Bo Li and Yevgeniy Vorobeychik. Feature cross-substitution in adversar- 1989.
ial classification. In Advances in neural information processing systems, [41] Tuan A Tang, Lotfi Mhamdi, Des McLernon, Syed Ali Raza Zaidi, and
pages 2087–2095, 2014. Mounir Ghogho. Deep learning approach for network intrusion detection
[21] Jenny Liu. Global optimization techniques using cross-entropy and in software defined networking. In 2016 International Conference on
evolution algorithms. Master’s Thesis, Department of Mathematics, Wireless Networks and Mobile Communications (WINCOM), pages 258–
University of Queensland, 2004. 263. IEEE, 2016.
[22] Yingqi Liu, Shiqing Ma, Yousra Aafer, Wen-Chuan Lee, Juan Zhai, [42] Florian Tramèr, Fan Zhang, Ari Juels, Michael K Reiter, and Thomas
Weihang Wang, and Xiangyu Zhang. Trojaning attack on neural Ristenpart. Stealing machine learning models via prediction apis. In
networks. 2017. 25th {USENIX} Security Symposium ({USENIX} Security 16), pages
[23] Daniel Lowd and Christopher Meek. Adversarial learning. In Pro- 601–618, 2016.
ceedings of the eleventh ACM SIGKDD international conference on [43] Eduardo K Viegas, Altair O Santin, and Luiz S Oliveira. Toward a
Knowledge discovery in data mining, pages 641–647. ACM, 2005. reliable anomaly-based intrusion detection in real-world environments.
[24] Nicolas Papernot, Fartash Faghri, Nicholas Carlini, Ian Goodfellow, Computer Networks, 127:200–216, 2017.
Reuben Feinman, Alexey Kurakin, Cihang Xie, Yash Sharma, Tom [44] Huang Xiao, Battista Biggio, Gavin Brown, Giorgio Fumera, Claudia
Brown, Aurko Roy, Alexander Matyasko, Vahid Behzadan, Karen Ham- Eckert, and Fabio Roli. Is feature selection secure against training data
bardzumyan, Zhishuai Zhang, Yi-Lin Juang, Zhi Li, Ryan Sheatsley, poisoning? In International Conference on Machine Learning, pages
Abhibhav Garg, Jonathan Uesato, Willi Gierke, Yinpeng Dong, David 1689–1698, 2015.
Berthelot, Paul Hendricks, Jonas Rauber, and Rujun Long. Technical [45] W Xu, Y Qi, and D Evans. Automatically evading classifiers: A case
report on the cleverhans v2.1.0 adversarial examples library. arXiv study on pdf malware classifiers. ndss, 2016.
preprint arXiv:1610.00768, 2018. [46] Fei Zhang, Patrick PK Chan, Battista Biggio, Daniel S Yeung, and
[25] Nicolas Papernot, Patrick McDaniel, Ian Goodfellow, Somesh Jha, Fabio Roli. Adversarial feature selection against evasion attacks. IEEE
Z Berkay Celik, and Ananthram Swami. Practical black-box attacks transactions on cybernetics, 46(3):766–777, 2015.
against machine learning. In Proceedings of the 2017 ACM on Asia
conference on computer and communications security, pages 506–519.
ACM, 2017.

Authorized licensed use limited to: Auckland University of Technology. Downloaded on May 26,2020 at 20:26:54 UTC from IEEE Xplore. Restrictions apply.

You might also like