Explainable Artificial Intelligence Applications I
Explainable Artificial Intelligence Applications I
Explainable Artificial Intelligence Applications I
This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
ABSTRACT This survey presents a comprehensive review of current literature on Explainable Artificial
Intelligence (XAI) methods for cyber security applications. Due to the rapid development of Internet-
connected systems and Artificial Intelligence in recent years, Artificial Intelligence including Machine
Learning and Deep Learning has been widely utilized in the fields of cyber security including intrusion
detection, malware detection, and spam filtering. However, although Artificial Intelligence-based
approaches for the detection and defense of cyber attacks and threats are more advanced and efficient
compared to the conventional signature-based and rule-based cyber security strategies, most Machine
Learning-based techniques and Deep Learning-based techniques are deployed in the ‘‘black-box’’ manner,
meaning that security experts and customers are unable to explain how such procedures reach particular
conclusions. The deficiencies of transparencies and interpretability of existing Artificial Intelligence
techniques would decrease human users’ confidence in the models utilized for the defense against cyber
attacks, especially in current situations where cyber attacks become increasingly diverse and complicated.
Therefore, it is essential to apply XAI in the establishment of cyber security models to create more
explainable models while maintaining high accuracy and allowing human users to comprehend, trust, and
manage the next generation of cyber defense mechanisms. Although there are papers reviewing Artificial
Intelligence applications in cyber security areas and the vast literature on applying XAI in many fields
including healthcare, financial services, and criminal justice, the surprising fact is that there are currently no
survey research articles that concentrate on XAI applications in cyber security. Therefore, the motivation
behind the survey is to bridge the research gap by presenting a detailed and up-to-date survey of XAI
approaches applicable to issues in the cyber security field. Our work is the first to propose a clear roadmap
for navigating the XAI literature in the context of applications in cyber security.
INDEX TERMS Artificial intelligence, cyber security, deep learning, explanation artificial intelligence,
intrusion detection, machine learning, malware detection, spam filtering.
1
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
Security (ENISA) [6], safe and trustworthy cyberspace is Implementing Artificial Intelligence in applications of cyber
expected to become even more crucial in the new social and security has been researched in recent years and many
economic norms formed by the COVID-19 epidemic. These previous surveys reviewed the existing work in this field. On
figures and events demonstrate the serious facts that the the other hand, the trends of applying XAI to provide more
Internet and connected networks and devices have suffered explainable and transparent services for areas including
more cybercriminals and cyber attacks nowadays. healthcare and image analysis are popular in research as well.
Therefore, a stable and secure cyber security computer However, to the best of our knowledge, although there are
system must be established to ensure the information privacy, some other excellent survey papers available on the topics of
accessibility, and integrity transmitted within the Internet. XAI and cyber security independently, there is a lack of a
Nevertheless, the conventional signature-based and rule- comprehensive survey paper focusing on the review of
based cyber defensive mechanisms are facing challenges solutions based on XAI across a wide variety of cyber
within the increasing quantities of information spread over security applications. This survey also concludes with special
the Internet [7]. On the other hand, cyber hackers are always deep analytical insights based on their opinions. These
striving to keep one step ahead of law enforcement by findings reveal several holes that may be filled using XAI
generating new, smart, and intricate attacking techniques and methods, indicating the overall future direction of research in
implementing technological advances including Artificial this domain.
Intelligence to make their adversarial behaviors more In general, this survey intends to provide a comprehensive
sophisticated and efficient [8]. As a consequence, researchers review of state-of-art XAI applications in the cyber security
in cyber security have begun to investigate Artificial area. The research motivations behind this work are listed as
Intelligence-based approaches especially Machine Learning followings:
and Deep Learning rather than traditional (non-AI) (1) To review different techniques and categorizations of
cybersecurity techniques including Game theory, Rate XAI.
Control, and Autonomous systems to enhance the (2) To review existing challenges and problems of XAI.
performance of cyber defensive systems. (3) To identify the frameworks and available datasets for the
Although Artificial Intelligence techniques, especially XAI-based cyber defensive mechanism.
Machine Learning and Deep Learning algorithms could (4) To review the latest successful XAI-based systems and
provide impressive performances on benchmark datasets in a applications in the cyber security domain.
number of cyber security domain applications such as (5) To identify challenges and research gaps of XAI
Intrusion detection, spam e-mail filtering, Botnet detection, applications in cyber security.
fraud detection, and malicious application identification [9]. (6) To identify the key insights and future research
Despite the excellent performance of Machine Learning and directions for applying XAI in the cyber security area.
Deep Learning algorithms, they can commit errors, some of
which are more expensive than conventional cyber defensive B. PREVIOUS SURVEYS
approaches. On the other hand, cyber security developers XAI and cyber security have been reviewed mostly
have sometimes sought higher accuracy at the price of separately in previous surveys. However, crossovers have
interpretability, making their models more intricate and emerged between the two domains. This survey presented a
difficult to grasp [10]. This lack of explainability has been comprehensive introduction of different XAI techniques
disclosed by the European Union’s General Data Protection applied in cyber defensive systems. Our work also provided
Regulation, preserving the capacity to comprehend the logic comprehensive XAI categorizations and analyzed details
behind an Artificial Intelligence algorithmic decision that about the existing challenges and frameworks of XAI for
negatively impacts individuals [11]. Accordingly, to be able cyber security. Cyber security datasets available for XAI
to believe the decisions of cyber security systems, Artificial models and the cyber threats faced by XAI models are
Intelligence must be transparent and interpretable. To satisfy discussed in this paper as well. Table 1 contrasts our study
these kinds of demands, several strategies have been with currently available surveys and reviewing articles.
proposed to make Artificial Intelligence decisions more Many existing surveys only analyzed AI applications, either
intelligible to humans. And these explainable techniques are Machine Learning or Deep Learning, in the cyber security
usually shortened as “XAI”, which have already been area, whereas other authors review XAI methods for a
implemented in many application domains such as healthcare, narrow set of cyber security applications. And some
Natural Language Processing, and financial services [12]. reviewers could not describe the background of XAI and
And the objective of this research paper is to focus on the cyber security in detail. Furthermore, most articles discuss
applications of XAI in different fields in the context of cyber
security.
A. RESEARCH MOTIVATION
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
TABLE 1. Comparison of existing surveys with our work (legend: √ means included; N/A means not included; ≈ means partially included)
Our Paper √ √ √ √ √ √ √ √ √ √ √
30 2022
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
comprehensive discussion of XAI applications in cyber research state-of-art in the areas of XAI applications in cyber
security from different perspectives. The existing challenges, security. Therefore, to collect the research articles reviewed,
key insights, and future directions of this area are highlighted the following criteria were established:
in Section VI, which is followed by the conclusion. And the 1) A thorough search was carried out whereas
conclusion would be the last section, which is Section VII. different academic search engines illustrated in
Table 2 were utilized to collect the relevant
TABLE 2. Research searching database engines.
papers.
2) The searching keywords for this survey paper were
Searching Engines Database Address constituted as 2 aspects: “XAI” and “Cyber
Security”. To create the search string, all potential
Springer https://link.springer.com/
Taylor & Francis https://taylorandfrancis.com/ pertinent synonyms of the given terms were
Semantic Scholar https://www.semanticscholar.org/ discovered in different databases and the percentage
ACM Digital Library https://dl.acm.org/ of reviewed papers from sources was depicted in
ResearchGate https://www.researchgate.net/
Figure 3. The following synonyms may be pertinent
Google Scholar https://scholar.google.com/
IEEE Xplore https://ieeexplore.ieee.org to the subject: “Cyber Security”, “Cyber Physical”,
Elsevier https://www.elsevier.com/ “Cyber Attack”, “Cyber Threat”, Network Security”,
Research Rabbit https://researchrabbitapp.com/ “Cyber Crime”, “XAI”, “Explainable Artificial
Intelligence”, “Interpretable Artificial Intelligence”,
“Explainable Machine Learning (XML)”, and
“Transparent Artificial Intelligence”.
3) Only researches published between 2011 and 2022
were selected to report on the most recent trends in
the application of XAI techniques in cyber security
for this research. Besides, papers published after
2017 were given higher attention and occupied a
large proportion of all reviewed publications, as
shown in Figure 4.
4) Only publications written in the English language
were included in this review and duplicated studies
were excluded.
5) Only papers objecting to cyber security vulnerability
domains were reviewed in this survey paper whereas
FIGURE 3. Percentage of Reviewed Papers from Sources.
researches proposing Machine Learning-based
systems, Deep Learning-based systems, XAI-based
mechanisms, and AI-based mechanisms would be
extracted.
The procedure of choosing articles was instantaneous and
consisted of two steps: firstly, the searching results were
initially chosen based on the selection criteria by scanning
the publications' titles and abstracts; secondly, the documents
chosen in the initial phase were thoroughly read to create a
shortlist of articles published that would be chosen based on
the inclusion and exclusion criteria.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
TABLE 3. Summary of XAI challenges. architecture. In Recital 71, the word ‘ ‘ explanation’’ is
mentioned, outlining the human right to contest the decision
Challenges Reference Descriptions made following such an evaluation and to get an explanation
[73] The necessity to develop defense mechanisms of the decision. Furthermore, Martin [85] investigated
against attacks especially for building 6G whether and to what degree people have a legal right to an
industries.
[74] The application of ML and DL in adversarial
explanation of automated decision-making under EU law,
XAI security circumstances. Be aware of the input data. particularly when AI systems are involved.
[75] Criticized some post-hoc explanation 4) THE TRADE-OFF BETWEEN INTERPRETABILITY
methods such as LIME and SHAP by fooling AND ACCURACY
these techniques.
[76] Discussed the DeepFool tool targeting DNN
The Explainability and performance (predictive accuracy) of
models and offered several methods against a model are generally shown to be in trading-off with each
it. other [90]. In fact, there is a demand for explainable models
[77] Outlined the fact that there is no accepted that can attain high performance because the algorithms that
system for determining the XAI system’s
priority.
currently perform the best are frequently the least explainable
[78] Proposed strong concerns about choosing the (for example, deep learning) [53].
XAI performance best technique for explainability Despite simple models being frequently favored for their
evaluation [80] Proposed a definition of feature relevance in ease of explaining [91], these models’ explainability may be
Boolean functions and a testing environment
compromised in cases when highly engineered or heavy
[81] Presented a framework for evaluating XAI
algorithms based on the CLEVR visual dimensional features are used [86].
question answering task. Amann et al. [87] adopted a multidisciplinary approach to
[82] Proposed concerns about the role of XAI in analyze the relevance of explainability for medical AI from
marketing AI applications. different perspectives, showing the necessity to apply XAI in
[83] The European Commission (EC) has also
published ethical guidelines for Trustworthy
clinical practice even though the primary objective is to give
Legal and privacy AI and highlighted privacy. patients the finest care possible [88].
issues [84] GDPR of the EU outlined the human right to
contest the decision made and got an IV. XAI FRAMEWORK AND DATASETS FOR CYBER
explanation of the decision. SECURITY
[85] Discussed what degree people have a legal
right to an explanation of automated decision-
making under EU law A. XAI FRAMEWORK FOR CYBER SECURITY
[53] Outlined the fact that the algorithms that In this section, based on the publications we have carefully
The trade-off currently perform the best are frequently the read in this survey, we provide a general XAI framework
between least explainable such as Deep Learning. diagram for cyber security applications. And the conceptual
interpretability [86] Pointed out that models’ explainability may
and accuracy be compromised in cases when highly
framework diagram for XAI applications in cyber security is
engineered or heavy dimensional features are illustrated in Figure 7. This diagram is considered to be as
used general as it can be to show the processes of applying XAI in
[87] Adopted a multidisciplinary approach to the cyber area domains. There are several stages in this
analyze the relevance of explainability for
medical AI from different perspectives workflow whereas certain sample instances are presented in
[88] Argued the necessity to apply XAI in clinical each stage.
practice The framework workflow starts by determining the types
of cyber security tasks, including malware detection, spam
3) LEGAL AND PRIVACY ISSUES detection, and fraud detection, which are defined by the types
Besides the above described technical challenges, XAI faces of cyber attacks facing. The corresponding data such as
significant legal and privacy issues as well. In numerous emails, network traffic, and application activities will be
instances, including some well-known court cases, a history collected and processed in the next stages. Then features
of biased legal and privacy issues was made by XAI systems representing significant characteristics will be extracted and
[89]. fed to train different Artificial Intelligence models depending
Arun [82] proposed concerns about the role of XAI in on specific situations. Cyber security test samples will be
influencing the privacy calculus of individuals, especially the analyzed and made decisions after the models have been
privacy concerns of customers in marketing AI applications. trained. Users can get decisions and explanations explicitly
The European Commission (EC) has also published ethical from self-interpretable models whereas the predictions made
guidelines for Trustworthy AI as a legal document [83], by black-box modes require explanations of XAI models to
highlighting the respect for privacy, quality and integrity of make the users requesting for the cyber security tasks
data, and access to data. satisfied. It is noticeable that this diagram is only a general
The General Data Protection Regulation (GDPR) [84] of workflow of XAI applied in cyber security areas, and the
the EU has added clarification to its information security details may differ for different tasks specifically.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
FIGURE 7 The conceptual framework diagram for XAI applications in cyber security.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
TABLE 4. Some public available datasets in the context of cyber attacks categories.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
For neural network-based detecting mechanisms, Shamik decision would be distributed certain values to each set
et al. [154] proposed a framework explaining how a deep respectively, showing the contribution of different sets of
neural network generalizes real-world testing set in different features to the detection results. The detection rates of TCP
layers. The gradients and weights of different layers of the flow and HTTP models reach 98.16% and 99.65% while the
MalConv architecture [155] and emberMalConv [156] are false positive rates are 5.14% and 1.84%.
analyzed to identify different parts’ contributions to the An explainable fast, and accurate approach for detecting
classification. High gradient values were found in the header Android malware called PAIRED was illustrated by
of the files while there are peaks elsewhere, demonstrating Mohammed et al. in [161]. The proposed detection system
that these parts are mostly responsible for classification achieved lightweight by reducing the number of features by a
results. Besides, two filters A and B learned two different factor of 84% and deploying classifiers that are not resource-
sets of features, the accuracy and F1-Score can achieve intensive. 35 static features were extracted and explained
91.2% and 90.7% respectively when model B was replaced later by SHAP methods. In the experiment, PAIRED
by model A. malware detection system was able to retain a very high
Hamad et al. [157] developed a pre-trained Inception-v3 accuracy of 97.98% while processing data in just 0.8206µs
CNN-based transfer learned model to analyze malware in by testing with the CICMalDroid2020 dataset with the
IoT devices. To better understand the features learned by the extracted 35 features.
CNN models, Gradient weighted class activation mapping Martin et al. [162] presented a novel way to find locations
(Grad-CAM) is utilized to generate cumulative heatmaps and in an Android app's opcode sequence that the CNN model
explain the models visually. Besides, t-distributed stochastic considered crucial and that might help with malware
neighbor embedding (t-SNE) is used to verify the density of detection. CNN was demonstrated to assign a high priority in
the features in the proposed CNN models. Achieved by the locations similar to those highlighted by LIME as the state-
suggested methods, the detection accuracies were 98.5% and of-the-art for highlighting feature relevance on the
96.9% on the available testing dataset with SoftMax benchmark Drebin [101] dataset. And satisfying
classifier and RF classifier respectively. experimental results were produced as well, including
Anli et al. [158] suggested a technique for extracting rules accuracy = 0.98, precision =0.98, recall = 0.98, and F1-Score
from a deep neural network so that the rules can be used to = 0.97.
identify mobile malware behaviors. To represent the rules 2) SPAM
discovered between the inputs and outputs of each hidden Due to the increasing number of Internet users, spam has
layer in the deep neural network, an input-hidden tree and a become a major problem for Internet users in recent years
single hidden-output tree for each hidden layer were [163]. According to [164], while over 306.4 billion emails
established. Then the hidden-output tree can tell the most were sent and received per day in 2021, spam emails
important hidden layer which could specify the related input- accounted for more than 55 percent of all emails sent in 2021,
hidden tree. The experimental results illustrated accuracy, meaning that unsolicited email messages accounted for
precision, recall, and F-Measure of the proposed method nearly half of all email traffic.
were 98.55%, 97.93%, 98.27%, and 98.04% respectively. Recently, AI-based systems can be regarded as an efficient
Giacomo et al. [159] offered a way for assessing deep option to tackle the spam issue primarily because of their
learning models for malware classification using image data. ability to evolve and tune themselves [165]. However, due to
It uses data from a Grad-CAM and makes an effort to extend the privacy and legal specialties of spam, users can ask many
the evaluation of the training phase of the models being questions about AI models, especially the black-box ML and
studied and provide visual information to security analysts. DL models [166]. For instance, a curious spam recipient can
Besides, this technique extends the use of the Grad-CAM and, have an interest in understanding the utilized AI models and
in addition to the cumulative heatmap, automates the analysis ask the following questions:
of the heatmaps, assisting security analysts in debugging the 1) Why is Message classified as spam by Model?
model without having any prior knowledge of the
issue/pattern in question. Over a testing dataset of more than 2) What distinguishes spam from no spam?
8,000 samples classified into 7 families, the proposed model 3) How does Model distinguish spam from no spam?
tested in the experimental study had a test accuracy of 97%. 4) How does Model work distinguishing an alternative
However, the limitation of this approach is the morphed spam filter Model′ used in the past?
version of the malicious sample belonging to the family can 5) How does Model work?
evade antimalware detection. These proposed questions can be answered by the
TrafficAV, an effective and explainable detection implementation of XAI algorithms and XAI algorithms
framework of mobile malware behavior using network traffic can be used to complement ML models with desired
was proposed by Shanshan et al. [160]. This framework properties, such as explainability and transparency [167].
provided explainability to users by defining four sets for each And many works of literature have studied this area to
feature extracted from the malware HTTP request and every enhance the trust of the AI-based spam filters.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
Julio et al. [168] conducted a highly exploratory the botnet detecting systems’ trust and prevent automation
investigation on fake spam news detection with ML bias when users have too much trust in the systems’ output.
algorithms from a large and diverse set of features. SHAP In [178], HATMA et al. proposed a novel model for
method was deployed to explain why some are classified botnet DGA detection. Five Machine Learning algorithms
as fake news whereas others are not by representative were utilized and tested with datasets of 55 botnet families.
models of each cluster. Novel features related to the source Random Forest achieved the best accuracy of 96.3% and
domain of the fake news are proposed and demonstrated outperformed previous works as well. Open-source
five times more frequencies appeared in the detection intelligence (OSINT) and XAI techniques including SHAP
models than in other features. Besides, only 2.2 percent of and LIME were combined in this work to provide an antidote
the models have a detection performance higher than 0.85 for skepticism toward the model’s output and enhance the
in terms of AUC, which highlights how difficult it is to system trust. Besides, the limitations of the proposed
identify bogus news. frameworks were the temporal complexity involved in
The legally required trade-off between accuracy and calculating the characteristics and the model's low resistance
explainability was discussed and demonstrated in the to Mask botnet assaults.
context of spam classification by Philipp et al. in [169] as Shohei et al. [179] presented a novel two-step clustering
well. A dataset of 5574 SMS messages [170] was used to approach based on DBSCAN to cluster botnets and classify
support the argument that it is equally important to select their categories. Important features were represented and
the appropriate model for the task at hand in addition to explained by combining subspace clustering and frequent
concentrating on making complex models understandable. pattern mining from 2 different real-world flow datasets:
In this work, under circumstances, that which just a small MAWI [180] and ISP. 60 bot groups from 61,167 IP
quantity of annotated training data is available, very addresses were categorized from the MAWI dataset whereas
simple models, such as Naive Bayes, can outperform more 295 bot groups from 408,118 IP addresses from the ISP
complicated models, such as Random Forests. dataset. And the cluster results of botnets were self-explained
HateXplain, a benchmark dataset for hate speech spam by using a dendrogram.
that considers bias and explainability from many angles Visualization tools are also used to give better
was introduced by Binny et al. in [171]. Several models explanations about the reasons for labeling an account as
including CNN-GRU [172], BiRNN [173], and BiRNN- botnet or legitimate. Michele et al. [181] suggested ReTweet-
Attention [174] were used and tested on this dataset Tweet (RTT), a small but informative scatterplot
whereas explainability-based metrics such as Intersection- representation to make it simpler to explore a user's
Over-Union (IOU), comprehensiveness, and sufficiency retweeting activities. While the proposed botnet detection
were utilized to evaluate the model interpretability. method Retweet-Buster (RTbust) based on Variational
Experimental results showed that models that succeed at autoencoders (VAEs) and long short-term memory (LSTM)
classification may not always be able to explain their network unsupervised feature extraction approaches were
conclusions in a way that is believable and accurate. The utilized in a black-box nature, the visualization tool RTT can
limitations behind this benchmark dataset are that external still be employed economically after RTbust has been
contexts that would be relevant to the classification task, applied to comprehend the traits of those accounts that have
such as the profile bio, user gender, and post history were been classified as bots.
not considered and the proposed dataset contained English Some researchers suggested the necessity to reduce the
language only. number of the required features for botnet classification to
3) BOTNET overcome the scalability and computation resource problems
A botnet attack is known as a group of connected computers and provide more reliable explanations in botnet detection
working together to carry out harmful and repetitive actions systems. In [182], Hayretdin et al. utilized Principal
to corrupt and disrupt the resources of a victim, such as Component Analysis (PCA) for feature dimension reduction
crashing websites [175]. As shown in Figure 9, a typical Decision Tree classifier preserved the original features and
botnet’s lifecycle contains 5 phases, including Initial clearly illustrated how the classifier determined the labels.
Injection, Secondary Injection, Connection, Malicious Therefore, An analyst for cyber security can quickly
Activities, and Maintenance and Updating. comprehend an attack or typical behavior and utilize this
The market for global botnet detection is anticipated to understanding to further interpret a security event or incident.
expand from US$207.4 million in 2020 to US$965.6 million With the rise of deep learning (DL), several pilot studies
in 2027, at a compound annual growth rate (CAGR) of 24.0 have been created to understand the behavior of botnet traffic.
percent from 2021 to 2027, according to [176]. And Imperva However, It is difficult for users to understand and put their
Research Labs [177] also found that botnets constituted 57% trust in the outcomes of present DL models because of neural
of all attacks against e-commerce websites in 2021. These networks’ poor decision-making and lack of transparency
statistics indicate that developing AI-based systems for compared to other approaches. To address this issue, Partha
detecting botnets is necessary. Besides, XAI can contribute to et al. [183] carried out in-depth tests using both synthetic and
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
Yongchun et al. [194] proposed a Hierarchical Explainable with the most attention would be regarded as the most
Network (HEN) to represent user behavior patterns, which important content contributing to the final decision.
could help with fraud detection while also making the Paulo et al. [200] utilized LIME and EBM explanation
inference process more understandable. Furthermore, a techniques based on malicious URLs for a phishing
transfer framework was suggested for knowledge transfer experiment on a publicly available dataset Ebbu2017 [201].
from source domains with sufficient and mature data to the EBM, Random Forest, and SVM classifiers rated accuracy of
target domain to address the issue of cross-domain fraud 0.9646, 0.9732, and 0.9469 respectively on the tested
detection. database. The empirical evidence supported that the models
A novel fraud detection algorithm called FraudMemory could accurately categorize URLs as phishing or legitimate,
was proposed in [195] by Kunlin et al. This methodology and they also added explainability to these machine learning
used memory networks to enhance both performance and models, improving the final classification outcome.
interpretability while using a novel sequential model to Visual explanations of the phishing detection system
capture the sequential patterns of each transaction. Besides, attracted attention in the work of Yun et al. [202] as well.
memory components were incorporated in FraudMemory to The proposed phishing website detection method Phishpedia
possess high adaptability to the existence of the concept drift. solved the challenging issues of logo detection and brand
The precision and AUC of the FraudMemory model were recognition in phishing website detection. Both high
0.968 and 0.969 respectively and performed better than any accuracy and little runtime overhead are attained via
other methods for comparison including SVM, DNN, RF, Phishpedia. And most crucially, unlike conventional methods
and GRU. such as EMD, PhishZoo, and LogoSENSE, Phishpedia does
Based on a real-world dataset and a simulated dataset, not demand training on any specific phishing samples.
Zhiwen and Jianbin [196] proposed an explainable Moreover, Phishpedia was implemented with the CertStream
classification approach within the multiple instance learning service, and in just 30 days, we found 1,704 new genuine
(MIL) framework that deployed the AP clustering method in phishing websites, far more than other solutions. In addition,
the self-training LSTM model to obtain a precise explanation. 1,133 of these were not flagged by any engines in VirusTotal.
The experimental results indicated that the presented Rohit et al. [203] proposed an anti-phishing method that
methodology surpassed the other 3 benchmark classifiers utilizes persuasion cues and investigated the effectiveness of
including AP, SVM, and RF in both 2 datasets. Only a few persuasion cues. Three machine learning models were
classification methods that can produce a straightforward developed with pertinent gain persuasion cues, loss
casual explanation is the one used in this study. persuasion cues, and combined gain and loss persuasion cues,
Wei et al. [197] proposed a Deep Learning-based behavior respectively, to respond to the research questions. We then
representation framework for clustering to detect fraud in compare the results with a baseline model that does not take
financial services, called FinDeepBehaviorCluster. Time the persuasion cues into account. The findings demonstrate
attention-based Bi-LSTM was used to learn the embedding that the three phishing detection models incorporating
of behavior sequence data whereas handcrafted features were pertinent persuasion cues considerably outperform the
deployed to provide explanations. Then a GPU-optimized baseline model in terms of F1-score by a range of 5% to 20%,
HDBSCAN algorithm called pHDBSCAN is used for making them effective tools for phishing email detection. In
clustering transactions with similar behaviors. The proposed addition, the use of the theoretical perspective can aid in the
pHDBSCAN has demonstrated comparable performance to creation of models that are comprehensible and can
the original HBDSCAN in experiments on two real-world understand black-box models.
transaction data sets but with hundreds of times greater 6) NETWORK INTRUSION
computation efficiency. An unauthorized infiltration into a computer in your
5) PHISHING company or an address in your designated domain is referred
Phishing refers to fake email messages that look to be sent by to as a network intrusion. On the other hand, Network
a well-known company. The intention is to either download Intrusion Detection Systems (NIDSs) are defined as
malicious software onto the victim's computer or steal monitoring network or local system activity for indications of
sensitive data from it, including credit card numbers and unusual or malicious behavior that violates security or
login credentials. Phishing is a form of online fraud that is accepted practices [36]. Recently, many works have adopted
gaining popularity [198]. ML and DL algorithms for building efficient NIDSs. In
Yidong et al. [199] proposed a multi-modal hierarchical addition, cyber security experts also consider introducing
attention model (MMHAM) that, for phishing website explainability to the black-box AI systems to make the
detection, jointly learned the deep fraud cues from the three NISDs more robust and many have tried with XAI [204].
main modalities of website content including URLs, text, and Pieter et al. [204] proposed a two-staged pipeline for
image. Extracted features from different contents would be robust network intrusion detection, which deployed XGBoost
aligned representations in the attention layer. This in the first phase and Autoencoder in the second phase.
methodology is self-explained because content distributed SHAP method was implemented to explain to the first stage
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
model whereas the explanation results were utilized in the understandability of intrusion detection alerts. The proposed
second stage to train the autoencoder. Experiments in the framework will help cyber analysts make better decisions
public corpus NSL-KDD [105] showed that the proposed because false positives will be quickly eliminated. Five
pipeline can outperform many state-of-the-art efforts in terms functional modules were identified in FAIXID framework:
of accuracy, recall, and precision with 93.28%, 97.81%, and the pre-modeling explainability model, the modeling module,
91.05% respectively on the NSL-KDD dataset while adding the post-modeling explainability module, the attribution
an extra layer of explainability. module, and the evaluation module. XAI algorithms
ROULETTE, an explainable network intrusion detection including Exploratory Data Analysis (EDA), Boolean Rule
system for neural attention multi-output classification of Column Generation(BRCG), and Contrastive Explanations
network traffic data was introduced by Giuseppina et al. in Method (CEM) were deployed in the pre-modeling
[205]. Experimentations were performed on two benchmark explainability model, the modeling module, and the post-
datasets, NSL-KDD [105] and UNSW-NB15 [113] to modeling explainability module respectively to provide
demonstrate the effectiveness of the proposed neural model cybersecurity analysts comprehensive and high-quality
with attention. The additional attention layer enables users to explanations about the detection decisions made by the
observe specific network traffic characteristics that are most framework. On the other hand, collecting analysts’ feedback
useful for identifying particular intrusion categories. Two through the evaluation module to enhance the explanation
heatmaps depicting the ranked average feature relevance of models by data cleaning also proved effective in this work as
the flow characteristics in the attention layer of the above 2 well.
datasets were provided to show the explanation. Shraddha et al. [211] proposed a system where the
Zakaria et al. [206] designed a novel Deep Learning and relations between features and system outcome, instance-
XAI-based system for intrusion detection in IoT networks. wise explanations, and local and global explanations aid to
Three different explanation methods including LIME, SHAP, get relevant features in decision making were identified to
and RuleFit were deployed to provide local and global help users to comprehend the patterns that the model has
explanations for the single output of the DNN model and the learned by looking at the generated explanations. If the
most significant features conducted to the intrusion detection learned patterns are incorrect, they can alter the dataset or
decision respectively. Experiments were operated on NSL- choose a different set of features to ensure that the model
KDD [105] and UNSW-NB15 [113] datasets and the learns the correct patterns. XAI methods including SHAP,
performance results indicated the proposed framework's LIME, Contrastive Explanations Method (CEM), ProtoDash,
effectiveness in strengthening the IoT IDS's interpretability and Boolean Decision Rules via Column Generation (BRCG)
against well-known IoT assaults and assisting cybersecurity were implemented at different stages of the framework so
professionals in better comprehending IDS judgments. that the neural network not being a black box. The
Yiwen et al. [207] presented an intrusion detection system experiment was performed on the dataset NSL-KDD [105]
aimed at detecting malicious traffic intrusion in networks and the proposed framework was applied to generate
such as flood attacks and Ddos attacks. This method was explanations from different perspectives.
XAI-based and deployed both neural networks and tree The Decision Tree algorithm was utilized by Basim et al.
models. It is noticeable that this approach decreased the in [212] to enhance trust management and was compared
number of convolution layers in the neural work to enhance with other Machine Learning algorithms such as SVM. By
the model’s explainability whereas the accuracy performance applying the Decision Tree model for the network intrusion
of the model was not sacrificed. XGBoost was implemented of benchmark dataset NSL-KDD [105], three tasks were
to process the prediction outputs of the neural network and performed: ranking the features, decision tree rule extraction,
the processed results would be fed to LIME and SHAP for and comparison with the state-of-the-art algorithms. The
further explanations. ranking of network features was listed and it is noticeable
A novel intrusion detection system known as BiLSTM- that not all features contributed to the decision of intrusion.
XAI was presented by S. Sivamohan et al. in [208]. Krill Besides, the advantages of the Decision Tree algorithm
herd optimization (KHO) algorithm was implemented to compared with other popular classifiers, being
generate the most significant features of two network computationally cheaper and easy to explain were also
intrusion datasets, NSL-KDD [105] and Honeypot [209], to demonstrated in this work.
reduce the complexities of BiLSTM model and thus enhance Syed et al. [213] suggested an Intrusion Detection System
the detection accuracy and explainability. The obtained that used the global explanations created by the SHAP and
detection rate of Honeypot is 97.2% and the NSL-KDD Random Forest joint framework to detect all forms of
dataset is 95.8% which was superior and LIME and SHAP malicious intrusion in network traffic. The suggested
were deployed to explain the detection decisions. framework was composed of 2 stages of Random Forest
Hong et al. [210] suggested a network intrusion detection classifiers and one SHAP stage. SHAP provided explanations
framework called FAIXID making use of XAI and data for the outcome of the initial Random Forest classifier and
cleaning techniques to enhance the explainability and one decision of the first Random Forest classifier with low
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
credibility would be reassessed by the secondary classifier. decision trees can provide a possible explanation for the
This three-stage based architecture can increase user trust clusters, this does not necessarily reflect how the model
while filtering out all cloaked dangerous network data by classifies this data, especially when there are numerous
introducing transparency to the decision-making process. equally valid explanations.
CSE-CIC IDS 2018 [214] dataset was utilized to evaluate the EXPLAIN, a feature-based and contextless DGAs
performance of the proposed framework and the presented multiclass classification framework was introduced by
architecture produced accuracy rates of 98.5 percent and 100 Arthur et al. in [219] and compared with several state-of-the-
percent, respectively on the test dataset and adversarial art classifiers such as RNN, CNN, SVM, RF, and ResNet
samples. based on real-world datasets including DGArchive [220] and
Tahmina et al. [215] proposed an XAI-based Machine University Network [221]. After the ResNet-based
Learning system to detect malicious DoH traffic within DNS techniques, the best model, EXPLAIN-OvRUnion, used 76
over HTTPS protocol. publicly available CIRA-CIC- features and achieves the best F1-score. Moreover, Only 28
DoHBrw-2020 dataset [216] was utilized in the testing of the features were used by EXPLAIN-OvRRFE-PI and
proposed Balanced and Stacked Random Forest framework EXPLAIN-RFRFE-PI, which outperformed all feature-based
and other Machine Learning algorithms including Gradient strategies put out in previous work by a significant margin.
Boosting and Generic Random Forest. The suggested Additionally, they outperformed the deep learning-based
approach in this work got slightly greater precision (99.91 algorithms M-Endgame, M-Endgame.MI, and M-NYU in
percent), recall (99.92 percent), and F1 score (99.91 percent) terms of F1-scores as well.
over other methods for comparison. Additionally, feature To address the issues of DGAs classification including
contributions to the detection results were also highlighted which traffic should be trained in which network and when,
with the help of the SHAP algorithm. The limitation of this and how to measure resilience against adversarial assaults,
framework would be the inconsideration of DGA-related Arthur et al. [222] proposed two ResNets-based DGAs
DoH traffic from other HTTPS traffic. detection classifiers, one for binary classification and the
7) DOMAIN GENERATION ALGORITHMS (DGA) other for multiclass classification. Experiments on real-world
DGAs are a type of virus that is frequently used to generate a datasets demonstrated that the proposed classifier performed
huge number of domain names that can be utilized for at least comparably to the best state-of-the-art algorithms for
evasive communication with Command and Control (C2) the binary classification test with a very low false positive
servers. It is challenging to prohibit harmful domains using rate, and significantly outperformed the competition in the
common approaches like blacklisting or sink-holing due to extraction of complex features. In addition, for the multiclass
the abundance of unique domain names. A DGA's dynamics classification problem, the ResNet-based classifier performed
widely used a seeded function. Deterring a DGA strategy better than previous work in attributing AGDs to DGAs for
presents a hurdle because an administrator would need to the multiclass classification problem, achieving an
recognize the virus, the DGA, and the seed value to filter out improvement of nearly 5 percent in F1-score while requiring
earlier dangerous networks and subsequent servers in the 30 percent less training time than the next best classifier. In
sequence. The DGA makes it more challenging to stop the explainability analysis, it was also highlighted that some
unwanted communications because a skilled threat actor can of the self-learned properties employed by the deep learning-
sporadically switch the server or location from which the based systems.
malware automatically calls back to the C2 [217]. Therefore, 8) DENIAL-OF-SERVICE (DOS)
blacklisting and other conventional malware management The Internet is seriously threatened by denial-of-service
techniques fall short in combating DGA attacks and many (DoS) assaults, and numerous protection measures have been
machine learning classifiers have been suggested. These suggested to address the issue. DoS attacks are ongoing
classifiers allow for the identification of the DGA responsible attacks in which malicious nodes produce bogus messages to
for the creation of a given domain name and consequently obstruct network traffic or drain the resources of other nodes
start targeted remedial actions. However, it's challenging to [223]. As the DoS attacks become increasingly complicated
assess the inner logic due to the black box aspect and the in the past years, conventional Intrusion Detection Systems
consequent lack of confidence makes it impossible to use (IDS) are finding it increasingly challenging to identify these
such models. newer, more sophisticated DoS attacks because they use
Franziska et al. [218] proposed a visual analytics more complicated patterns. To identify malicious DoS
framework that offers clear interpretations of the models assaults, numerous machine learning and deep learning
created by deep learning model creators for the classification models have been deployed. Additionally, for the goal of
of DGAs. The activations of the model's nodes were model transparency, XAI methods that investigate how
clustered, and decision trees were utilized to illuminate these features contribute to or impact an algorithm-based choice
clusters. The users can examine how the model sees the data can be helpful [224].
at different layers in conjunction with a 2D projection. A
drawback of the proposed strategy is that although the
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
TABLE 6. Details of XAI applications in defending mechanisms against different categories of cyber attacks.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
images (fMRI) and identify neurological biomarkers was integration into IoT and AI-enabled smart city applications
proposed by Xiaoxiao et al. [232]. Motivated by the can help to address black-box model difficulties and offer
requirements for transparency and explainability in medical transparency and explainability components for making
image analysis, the proposed BrainGNN framework included useful data-driven decisions for smart city applications.
ROI-selection pooling layers (R-pool) that highlight Smart city applications are usually utilized in high-risk and
prominent ROIs (nodes in the graph) so that which ROIs are privacy-sensitive scenarios. Therefore, it is crucial to
crucial for prediction could be determined. By doing so, the establish an effective XAI approach to give authorities
advantage of the BrainGNN framework could be the additional information about the justification, implications,
allowance of users to interpret significant brain regions in potential throughput, and an in-depth explanation of
multiple ways. background procedures to aid in final decision-making [236].
The chain of reasoning behind Computer Aided Roland et al. [237] introduced a tree-based method
Diagnostics (CAD) is attracting attention to build trust in Gradient Boosted Regression Trees (GBRT) model in
CAD decisions from complicated data sources such as conjunction with the SHAP-value framework to identify and
electronic health records, magnetic resonance imaging scans, analyze major patterns of meteorological determinants of
cardiotocography, etc. To address this issue, Julian et al. [233] PM1 species and overall PM1 concentrations. SIRTA [238],
presented a new algorithm, Adaptive-Weighted High a ground-based atmospheric observatory dataset for cloud
Importance Path Snippets (Ada-WHIPS) to explain and aerosol was utilized to experiment and the location for
AdaBoost classification with logical and simple rules in the establishing this dataset was in the city of Paris. The findings
context of CAD-related data sets. The weights in the of this study show that shallow MLHs, cold temperatures,
individual decision nodes of the internal decision trees of the and low wind speeds play distinct roles during peak PM1
AdaBoost model are redistributed especially by Ada-WHIPS. events in winter. Under high-pressure synoptic circulation,
A single rule that dominated the model's choice is then northeastern wind input frequently intensifies these
discovered using a straightforward heuristic search of the conditions.
weighted nodes. Moreover, according to experiments on nine One of the most demanded bus lines of Madrid was
CAD-related data sets, Ada-WHIPS explanations typically analyzed by Leticia et al. in [239] to make the smart city
generalize better (mean coverage 15 percent to 68 percent) transport network more efficient by predicting bus passenger
than the state of the art while being competitive for demand. The proposed method created an interpretable
specificity. model from the Long Short Term Memory (LSTM) neural
A novel human-in-the-loop XAI system, XAI-Content network that enhances the generated XAI model's linguistic
based Image Retrieval (CBIR), was introduced by Deepak et interpretability without sacrificing precision using a surrogate
al. in [234] to retrieve video frames from minimally invasive model and the 2-tuple fuzzy linguistic model. The public
surgery (MIS) videos that are comparable to a query image transportation business can save money and energy by using
based on content. MIS video frames were processed using a passenger demand forecasting to plan its resources most
self-supervised deep learning algorithm to extract semantic effectively. This methodology can also be used in the future
features. The search results were then iteratively refined to forecast passenger demand for other forms of
using an iterative query refinement technique, which utilized transportation (air, railway, marine).
a binary classifier that has been trained online using user Georgios et al. [240] proposed explainable models for
feedback on relevance. The saliency map, which provided a early prediction of certification in Massive Open Online
visual description of why the system deems a retrieved image Courses (MOOCs) for Smart City Professionals. MOOCs
to be similar to the query image, was produced using an XAI have grown significantly over the past few years due to
technique. The proposed XAI-CBIR system was tested using Covid-19 and tend to become the most common type of
the publicly available Cholec80 dataset, which contains 80 online and remote higher education. Several Machine
films of minimally invasive cholecystectomy procedures. Learning classification techniques such as Adaptive Boosting,
2) XAI FOR CYBER SECURITY OF SMART CITIES Gradient Boosting, Extremely Randomized Trees, Random
As increasingly data-driven artificial intelligence services Forest, and Logistic Regression were utilized to build
such as IoT, blockchain, and deep learning are incorporated corresponding predictive models using PyCaret. And the
into contemporary smart cities, smart cities are able to offer XAI method SHAP summary plot was employed to the
intelligent services for energy, transportation, healthcare, and classifiers including LightGBM, GB, and RF. Furthermore,
entertainment to both city locals and visitors by real-time new classification models based only on the two most
environmental monitoring [235]. However, smart city important features in each step gained from the SHAP
applications not only gather a variety of information from summary plot. And the experimental results showed that the
people and their social circles that are sensitive to privacy, effectiveness of all methods was slightly improved for all
but also control municipal services and have an impact on metrics.
people's life, cyber security, cyber crime, and privacy 3) XAI FOR CYBER SECURITY OF SMART FARMING
problems about smart cities arise. To address this issue, XAI
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
Smart farming refers to the use of cutting-edge technology in extremely sensitive areas such as Money Laundering
agriculture, including IoT, robots, drones, sensors, and detection and Corporate Mergers and Acquisitions to not
geolocation systems. Big data, cloud computing, AI, and only have a highly accurate and robust model but also to be
augmented reality are the engines of smart farming as well. able to produce helpful justifications to win a user's faith in
However, the addition of several communication modules the automated system.
and AI models leaves the system open to cyber-security risks Swati et al. [246] proposed a belief-rule-based automated
and threats to the infrastructure for smart farming [241]. And AI decision-support system for loan underwriting (BRB).
cyber attacks can harm nations' economies that heavily rely This system can take into account human knowledge and can
on agriculture. However, due to the black box nature of most employ supervised learning to gain knowledge from prior
AI models, users cannot understand the connections between data. Factual and heuristic rules can both be accommodated
features. This is crucial when the system is designed to by BRB's hierarchical structure. The significance of rules
simulate physical farming events with socioeconomic effects triggered by a data point representing a loan application and
like evaporation [242]. Therefore, many researchers are the contribution of attributes in activated rules can both be
working on the implementation potentials of XAI applied in used to illustrate the decision-making process in this system.
smart farming cyber security. The textual supplied to rejected applicants as justification for
Nidhi et al. [242] presented an IoT and XAI-based declining requesters’ loan applications might have been
framework to detect plant diseases such as rust and blast in started by the progression of events from the factual-rule-
pearl millet. Parametric data from the pearl millet farmland at base to the heuristic-rule-base.
ICAR, Mysore, India was utilized to train the proposed A novel methodology for producing plausible
Custom-Net Deep Learning Models, reaching a classification counterfactual explanations for the Corporate Mergers and
accuracy of 98.78% which is similar to state-of-the-art Acquisitions (M&A) Deep Transformers system was
models including Inception ResNet-V2, Inception-V3, presented by Linyi et al. [247]. The proposed transformer-
ResNet-50, VGG-16, and VGG-19 and superior to them in based classifier made use of the regularization advantages of
terms of reducing the training time by 86.67%. Additionally, adversarial training to increase model resilience. More
the Grad-CAM is used to display the features that the significantly, a masked language model for financial text
Custom-Net extracted to make the framework more categorization that improved upon prior methods to measure
transparent and explainable. the significance of words and guarantee the creation of
To thoroughly assess the variables that can potentially credible counterfactual explanations was developed. When
explain why agricultural land is used for plantations of wheat, compared to state-of-art methods including SVM, CNN,
maize, and olive trees, Viana et al. [243] implemented an ML BiGRU, and HAN, the results show greater accuracy and
and agnostic-model approach to show global and local explanatory performance.
explanations of the most important variables. Machine An interactive, evidence-based method to help customers
Learning model Random Forest and XAI approach LIME understand and believe the output produced by AI-enabled
were deployed for analysis and approximately 140 variables algorithms was generated for analyzing customer
related to agricultural socioeconomic, biophysical, and transactions in the smart banking area by Ambreen [248]. A
bioclimatic factors were gathered. By applying the proposed digital dashboard was created to make it easier to engage
framework, it is found that the three crop plantations in the with algorithm results and talk about how the suggested XAI
research area's usage of agricultural land were explained by method can greatly boost data scientists' confidence in their
five major factors: drainage density, slope, soil type, and the ability to comprehend the output of AI-enabled algorithms.
ombrothermic index anomaly (for humid and dry years). In the proposed model, a Probabilistic Neural Network (PNN)
4) XAI FOR CYBER SECURITY OF SMART FINANCIAL was utilized to classify the multi-class scenario of bank
SYSTEM transaction classification.
The financial system has been rapidly altered by AI models, 5) XAI FOR CYBER SECURITY OF HUMAN-COMPUTER
which offer cost savings and improved operational efficiency INTERACTION (HCI)
in fields like asset management, investment advice, risk HCI enables people to comprehend and engage with
forecasting, lending, and customer service [244]. On one technology by establishing an effective channel of
hand, the ease of using AI in these smart financial systems communication. And HCI's primary goal is to create
provides efficiency for all parties involved, but on the other interactions that take users' wants and abilities into account
hand, the risk of cyberattacks on them is growing [249]. In the field of HCI, security and privacy have long
exponentially. Attackers have traditionally been motivated been significant research concerns, where Usable Security
primarily by money, making smart financial systems their top has arisen as an interdisciplinary research area. On the other
choice of target. To combat the finance crime targeting smart hand, HCI and AI emerge together in such a way that AI
financial systems, one of the primary priorities in the smart imitates human behavior to create intelligent systems,
financial domain should be the implementation of XAI [245]. whereas HCI tries to comprehend human behavior to modify
The reason behind this issue is that it is essential in these the machine to increase user experience, safety, and
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
efficiency. However, from an HCI standpoint, there is no past few years, AI has made significant progress in providing
assurance that an AI system's intended users will be able to effective performance in smart transportation systems, the
comprehend it. And according to the user-centered design XAI methods are still required as XAI could make it possible
(UCD), a design must offer an understandable AI that cyber- for the smart transportation system to monitor transportation
attacks the requirements and skills of the intended users (e.g., details such as drivers’ behaviour, accicent causes, and
knowledge level). Therefore, the final objective of XAI in vechicles’ conditions.
HCI should be to guarantee that target users can comprehend A Machine Learning approach to detect misbehaving
the outcomes, assisting them in becoming more efficient vehicles in the Vehicular Adhoc Networks (VANET) was
decision-makers [250]. proposed by Harsh et al. [256]. In the smart VANET, the
Gaur et al. [251] utilized XAI methods including LIME performance of each vehicle depends upon the information
and SHAP in conjunction with machine learning algorithms from other autonomous vehicles (AVs). Therefore, the
including Logistic Regression(80.87%), Support Vector misinformation from misbehaving vehicles would damage
Machine(85.8%), K-nearest Neighbour(87.24%), Multilayer the entire VANET as a whole and detecting misbehaving
Perceptron(91.94%), and Decision Tree(100%) to build a would be significant to build a stable and safe VANET
robust explainable HCI model for examining the mini-mental system. Vehicular reference misbehavior (VeReMi) dataset
state for Alzheimer’s disease. It is worth mentioning that the [257] was utilized in an ensemble learning using Random
most significant features contributing to the Alzheimer's Forest algorithm and a decision tree-based algorithm and
disease examing were different for the LIME-based accuracy and F1 score of 98.43% and 98.5% were achieved
framework and the SHAP-based framework. In contrast to respectively.
nWBV's dominance of the LIME features, MMSE makes a Shideh et al. [258] described a transportation energy
significant contribution to Shapely values. model (TEM) that forecasts home transportation energy use
To fill the gap few publications on artistic image using XAI technique LIME. Data from Household Travel
recommendation systems give an understanding of how users Survey (HTS), which is utilized to train the artificial neural
perceive various features of the system, including domain network accurately, has been deployed in TEM and high
expertise, relevance, explainability, and trust, Vicente et al. validation accuracy (83.4%) was developed. For certain
[252] examed several aspects of the user experience with a traffic analysis zones (TAZs), the significance and impact
recommender system of artistic photos from algorithmic and (local explanation) of HTS inputs (such as household travel,
HCI perspectives. Three different recommender interfaces demographics, and neighborhood data) on transportation
and two different Visual Content-based Recommender energy consumption are studied. The explainability of the
(VCBR) algorithms were employed in this research. proposed TEM framework can help the home transportation
Q. Vera et al. [253] presented a high-level introduction of energy distribution in two ways, including describing the
the XAI algorithm's technical environment, followed by a local inference mechanisms on individual (household)
selective examination of current HCI works that use human- predictions and assessing the model's level of confidence can
centered design, evaluation, and provision of conceptual and be done using a broad grasp of the model.
methodological tools for XAI. Human-centered XAI was C. Bustos et al. [259] provided an automated scheme for
highlighted in this research, and the emerged research reducing traffic-related fatalities by utilizing a variety of
communities of human-centered XAI were introduced in the Computer Vision techniques (classification, segmentation,
context of HCI. and interpretability techniques). An explainability analysis
6) XAI FOR CYBER SECURITY OF SMART based on image segmentation and class activation mapping
TRANSPORTATION on the same images, as well as an adaptation and training of a
The emergence of cutting-edge technologies including Residual Convolutional Neural Network to establish a danger
software-defined networks (SDNs), IIoT, Blockchain, AI, index for each specific urban scene, are all steps in this
and vehicular ad hoc networks (VANETs) has increased process. This computational approach results in a fine-
operational complexity while smoothly integrating smart grained map of risk levels across a city as well as a heuristic
transportation systems [254]. However, it can experience for identifying potential measures to increase both pedestrian
security problems that leave the transportation systems open and automobile safety.
to intrusion. In addition, security concerns in transportation
technology affect the AI model [255]. Major transportation C. CYBER THREATS TARGETING XAI AND DEFENSIVE
infrastructures such as Wireless Sensor Networks (WSN), APPROACHES
Vehicle-to-everything communication (V2X), VMS, and In the above sections, the applications of XAI in different
Traffic Signal Controllers (TSC) have either already been areas to defend against different cyber threats have been
targeted or are still susceptible to hacking. To defend against discussed. Nevertheless, although XAI could be effective in
these cyber attacks and prevent the potential cyber threats on protecting other areas and models by providing transparency
the smart transportation system, AI-enabled intrusion and explainability, XAI models themselves would face cyber
detection systems are introduced recently. Although In the threats as well. Both the AI models deployed and the
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
explainability part could be vulnerable to cyber attacks. they want, effectively masks the biases of any given classifier.
Some cyber attackers even utilize the explainable Extensive experimental testing using real data from the
characteristics to attack the XAI model. Therefore, we deem criminal justice and credit scoring fields showed that the
it necessary to review the cyber threats targeting XAI and proposed fooling method was successful in producing
corresponding defensive approaches against them in this adversarial classifiers that can trick post-hoc explanation
review. procedures, including LIME and SHAP, with LIME being
Apart from the different parts that conventional AI models found to be more susceptible than SHAP. In detail, it was
need to protect, including samples, learning models, and the demonstrated how highly biased (racist) classifiers created by
interoperation processes, the explainable part of XAI-based the proposed fooling framework can easily deceive well-
models should be paid attention to as well. The following liked explanation techniques like LIME and SHAP into
researches describe some cyber attacks targeting XAI models producing innocent explanations which do not reflect the
using different approaches from different perspectives. underlying biases using extensive evaluation with numerous
A novel black box attack was developed by Aditya et al. real-world datasets (including COMPAS [264]).
[260] to examine the consistency, accuracy, and confidence Simple, model-agnostic, and intrinsic Gradient-based NLP
security characteristics of gradient-based XAI algorithms. explainable approaches are considered faithful compared
The proposed black box attack focused on two categories of with other state-of-art XAI approaches including SHAP and
attack: CI and I attack. While I attack attempts to attack the LIME. However, Junlin et al. [265] show how the gradients-
single explainer without affecting the classifier's prediction based explanation methods can be fooled by creating a
given a natural sample, the CI attack attempts to FACADE classifier that could be combined with any
simultaneously compromise the integrity of the underlying particular model having deceptive gradients. Although the
classifier and explainer. It is demonstrated that the gradients in the final model are dominated by the customized
effectiveness of the attack on various gradient-based FACADE model, the predictions are comparable to those of
explainers as well as three security-relevant data sets and the original model. They also demonstrated that the proposed
models through empirical and qualitative evaluation. method can manipulate a variety of gradient-based analysis
Thi-Thu-Huong et al. [261] proposed a robust adversarial methods: saliency maps, input reduction, and adversarial
image patch (AIP) that alters the causes of interpretation perturbations all misclassify tokens as being very significant
model prediction outcomes and leads to incorrect deep neural and of low importance.
networks (DNNs) model predictions, such as gradient- On the other hand, to defend against these cyber threats
weighted class activation mapping. Four tests pertaining to targeting XAI models, researchers also developed several
the suggested methodology were carried out on the ILSVRC defensive approaches, divided into three main categories:
image dataset. There are two different kinds of pre-trained modifying the training process and input data, modifying the
models (i.e., feature and no feature layer). The Visual model network, and sing auxiliary tools.
Geometry Group 19-Batch Normalization (VGG19-BN) and Gintare et al. [266] assessed how JPG compression affects
Wide Residual Networks models, in particular, were used to the categorization of adversarial images. Experimental tests
test the suggested strategy (Wide ResNet 101). Two more demonstrated that JPG compression could undo minor
pre-trained models: Visual Geometry Group 19 (VGG19) adversarial perturbations brought forth by the Fast-Gradient-
and Residual Network (ResNext 101 328d), were also Sign technique. JPG compression could not undo the
deployed whereas masks and heatmaps from Grad-CAM adversarial perturbation, nevertheless, if the perturbations are
results were utilized to evaluate the results. more significant. In this situation, neural network classifiers'
Tamp-X, a unique approach that manipulates the strong inductive bias cause inaccurate yet confident
activations of powerful NLP classifiers was suggested by misclassifications.
Hassan et al. [262], causing cutting-edge white-box and Ji et al. [267] present DeepCloak, a defense technique.
black-box XAI techniques to produce distorted explanations. DeepCloak reduces the capacity an attacker may use to
Two steps were carried out to evaluate state-of-art XAI generate adversarial samples by finding and eliminating
methods, including the white-box InteGrad andSmoothGrad, pointless characteristics from a DNN model, increasing the
and the black-box—LIME and SHAP. The first step was to robustness against such adversarial attacks. In this work, the
randomly mask keywords and observe their impact on NLP mask layer, inserted before processing the DNN model,
classifiers whereas the second step was to tamper with the encoded the discrepancies between the original images and
activation functions of the classifiers and evaluate the outputs. related adversarial samples, as well as between these images
Additionally, three cutting-edge adversarial assaults were and the output features of the preceding network model layer.
utilized to test the tampered NLP classifiers and it was found Pouya et al. [268] Defense-GAN, a novel defense
that the adversarial attackers have a much tougher time technique leveraging GANs to strengthen the resilience of
fooling the tampered classifiers. classification models against adversarial black-box and
Slack et al. [263] provided a unique scaffolding method white-box attacks. The proposed approach was demonstrated
that, by letting an antagonistic party create any explanation to be successful against the majority of frequently thought-of
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
attack tactics without assuming a specific assault model. On could be utilized to measure the performance of the proposed
two benchmark computer vision datasets, we empirically mechanisms. However, when applying XAI methods in the
demonstrate that Defense-GAN consistently offers cyber security domains, measurements to evaluate the
acceptable defense while other approaches consistently accuracy and completeness of explanations from the XAI
struggled against at least one sort of assault. systems are required. In general, the evaluation
measurements of XAI systems should be able to assess the
VI. ANALYSIS AND DISCUSSION quality, value, and satisfaction of explanations, the
A. CHALLENGES OF USING XAI FOR CYBER enhancement of the users’ mental model brought about by
SECURITY model explanations, and the impact of explanations on the
We have reviewed the state-of-art XAI techniques utilized in effectiveness of the model as well as on the users’ confidence
the defense of different cyber attacks and the protection of and reliance. Unfortunately, the findings derived from the
distinct industrial cyber security domains. It is noticeable that above reviews of this survey demonstrate the challenge that:
although XAI could be a powerful tool in the application of more generic, quantifiable XAI system evaluation
different cyber security domains, XAI faces certain measurements are required to support the community's
challenges in its application of cyber security. And in this suggested XAI explainability measuring techniques and tools.
section, we will discuss these challenges. Popular XAI explanation evaluation measurements can be
1) DATASETS divided into two main categories: user satisfaction and
An overview of the famous and commonly used datasets of computational measurements. However, user satisfaction-
different cyber attacks and distinct industries was provided in based evaluation approaches are dependent on user feedback
Table 4 and Table 5 respectively. However, there is a severe or interview, which may cause privacy issues for many cyber
issue with the most used cyber security datasets, i.e. many security problems. On the other hand, for computational
datasets are not updated in certain directions. For instance, measurements, many researchers utilize inherently
the most famous datasets about spam detection were more interpretable models [56] (e.g., linear regression and decision
than 10 years before. This phenomenon may be caused by trees) to compare with the generated explanations.
privacy and ethical issues. Therefore, the most recent Nevertheless, there are no benchmark comparison models for
categories of cyber attacks were not included in the public this evaluation approach, and the users’ understanding of the
cyber attack datasets, which would lead to inefficiency in the explanation could not be reflected. Besides, the XAI
training of the XAI applications in the establishment of cyber evaluation systems lack measurements focusing on some
attack defensive mechanisms. Although the industrial other significant factors of the cyber security domain
datasets in areas such as healthcare, smart agriculture, and including computational resources as well as computational
smart transportation include more recent samples than the power. In conclusion, it is necessary to take into account a set
datasets for cyber attacks, these datasets should be updated as of agreed-upon standard explainability evaluation metrics for
well because cyber attacks are becoming more sophisticated comparison to make future improvements for XAI
and diverse these days. Another issue with the currently applications in cyber security.
available datasets is that these datasets usually lack a large 3) CYBER THREATS FACED BY XAI MODELS
volume of data available for the training of XAI methods, As we discussed in Section V, although XAI methods can
which will decrease both the performance and the provide transparency and explainability to AI-enabled
explainability of the XAI approaches. Another reason behind systems to prevent cyber threats, the current XAI models are
this situation is that some of the information related to cyber facing many cyber attacks targeting the vulnerabilities of the
attacks and cyber industries is redundant and unbalanced. explanation approaches, which is extremely dangerous for
Other than that, the heterogeneity of samples collected in the cyber security systems as they always require a high level
these datasets is a challenge for the XAI models as well. The of safety. For instance, many researchers [263] [264] have
number of features and categories varies for each dataset and proved the fact that it is possible to fool some of the most
some datasets are composed of human-generated cyber popular XAI explanation methods such as LIME and SHAP,
attacks rather than exhibiting real-world and latest attacks. which are also frequently deployed in the XAI application of
These problems highlight the challenge that the most recent cyber security areas. It is demonstrated that the explanations
benchmark datasets with a massive amount of data for generating processes of those state-of-art XAI methods might
training and testing and a balanced and equal number of be counter-intuitive. Other than that, in the practical
attack categories are still to be identified. industrial cyber security domains, such as XAI-enabled face
2) EVALUATION authentication systems. Although in Section V, we have
Evaluation measure for XAI systems is another important discussed several defensive methods against cyber threats
factor in the application of XAI approaches for cyber security. targeting XAI systems, most defensive approaches focus on
When evaluating the performance of the established XAI- the protection of the performance of the prediction results of
based cyber security systems, several conventional XAI models rather than the explanation results. However, for
evaluation metrics including F1-Score, Precision, and ROC XAI-based cyber security systems, the explainability of the
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
models is significant to maintain the transparency and implemented in cyber security. Selecting different
efficiency of the entire system and prevent the cyber attacks explanation approaches for distinct Machine
as well. Learning or Deep Learning algorithms in different
4) PRIVACY AND ETHICAL ISSUES cyber security tasks would influence the
In addition to the aforementioned technical challenges, performance and explainability of XAI models
privacy and ethical issues are also crucial challenges when significantly. Other than that, the tuning process of
implementing XAI in cyber security. During the system life parameters and model structures of the established
cycle, XAI models must explicitly take privacy concerns into XAI model is another crucial consideration as well.
account. It is commonly agreed that respecting every person's 4) The model defense could be highlighted in
right to privacy is essential, especially in some very sensitive particular for cyber security tasks as they are the
areas of cyber security, for instance, authentication, e-mails, main targets for cyber attackers. Especially for
and password. Moreover, XAI systems naturally fall within XAI-based cyber security mechanisms, the decision
the general ethical concern of potential discrimination (such model, security data as well as the explanation
as racism, sexism, and ageism) by AI systems. In theory, process should be protected to prevent cyber threats.
identical biases may be produced by any AI model that is 5) Privacy awareness is another insight that XAI
built using previously collected data from humans. It is methods could provide for the cyber security system.
important to take precautions to ensure that there is no Giving end users of cyber security systems a way to
discrimination, bias, or unfairness in the judgments made by evaluate their data privacy is a significant objective
the XAI system and the explanations that go along with them. in the application of XAI. End-users could learn
The ethical bias of XAI systems should be eliminated in through XAI explanations about what user data is
terms of justification as well as explainability, in particular in used in algorithmic decision-making.
specific domains of cyber security applications. For privacy
issues, because the data are gathered from security-related C. FUTURE RESEARCH DIRECTIONS
sources, the privacy and security-related concerns increase. 1) HIGH-QUALITY DATASETS
Therefore, it is essential to guarantee that data and models are The quantity and quality of the available datasets have a
protected from adversarial assaults and being tampered with significant impact on how well XAI methods work for the
by unauthorized individuals, which means that only cyber security system, and the biases and constraints of the
authorized individuals should be permitted access to XAI datasets used to train the models have an impact on how
models. accurate the decisions and explanations are. On the other
hand, as we discussed in the above sections, the existing
B. KEY INSIGHTS LEARNED FROM USING XAI FOR available cyber security datasets could not reflect the most
CYBER SECURITY
recent cyber attacks due to privacy and ethical issues. Data
In this section, some key insights learned from using XAI for
from real networks or the Internet typically contain sensitive
cyber security will be discussed based on the review in the
information, such as personal or business details, and if made
above sections. The main insights for the XAI
publicly available, they may disclose security flaws in the
implementation in cyber security systems can be itemized as
network from which they originated. Additionally, the
follows:
imbalance of both volumes and features of the datasets would
1) User trust and reliance should be satisfied. By influence the establishment of the XAI-based cyber security
offering explanations, an XAI system can increase system negatively as well. Therefore, the construction of both
end users' trust in the XAI-based cyber security high-quality and up-to-date datasets available for XAI
system. Users of an XAI system can test their applications for cyber security could be a possible future
perception of the system's correctness and reliability. research direction.
Users become dependent on the system as a result
2) TRADE-OFF BETWEEN PERFORMANCE AND
of their trust in the XAI-based cyber security EXPLAINABILITY
system. It is essential for cyber security experts to maintain the trade-
2) Model visualization and inspection should be off between performance and explainability aspects of the
considered. Cyber security experts could benefit newly introduced XAI-enabled cyber security systems. It is
from XAI system visualization and explainability to noticeable that although for some self-explainable XAI
inspect model uncertainty and trustworthiness. approaches, for instance, Decision Tree, the model is quite
Additionally, identifying and analyzing XAI model transparent and users could understand the decision-making
and system failure cases is another crucial process easier, the performance of those approaches could
component of model visualization and inspection. not always be satisfying. On the other hand, the AI
3) Model tuning and selection are crucial factors to algorithms that now often perform best (for example, Deep
ensure the efficiency of the XAI model Learning) are the least explainable, causing a demand for
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
explainable models that can achieve high performance. Some even more severe as both the decisions and the explanations
researchers have exploited this area, including authors of related to users should be preserved. As a result, there is a
[269] significantly reduce the trade-off between efficiency conflict between using big data for security and safeguarding
and performance by introducing XAI for DNN into existing it. Data must be guaranteed to be safe from adversarial
quantization techniques. And authors of [270] demonstrated assaults and manipulation by unauthorized users and
that the wavelet modifications provided could lead to legitimate users should also be able to access the data.
significantly smaller, simplified, more computationally Therefore, the protection of data and generated explanations
efficient, and more naturally interpretable models, while of XAI systems could be a future research direction as well.
simultaneously keeping performance. However, there is a
lack of research focusing on the trade-off of performance and VII. CONCLUSION
explainability of XAI approaches applied in cyber security. XAI is a powerful framework to introduce explainability and
3) USER-CENTERED XAI transparency to the decisions of conventional AI models
The human understandability of XAI approaches has become including Deep Learning and Machine Learning. On the
the focus of some recent studies to find new potential for its other hand, cyber security is an area where transparency and
application in areas of cyber security. As we mentioned in explainability are required to defend against cyber security
the above sections, user satisfaction with the generated threats and analyze generated security decisions. Therefore,
explanation is a significant component of the XAI in this paper, we presented a comprehensive survey of state-
approaches to explainability evaluation. However, in areas of of-art research regarding XAI for cyber security applications.
cyber security, the questionnaire and feedback of users are We concluded the basic principles and taxonomies of state-
limited to some degree due to security concerns. Therefore, of-art XAI models with essential tools, such as a general
how to generate user-centered XAI systems for cyber framework and available datasets. We also investigated the
security end users in terms of user understanding, user most advanced XAI-based cyber security systems from
satisfaction, and user performance without violating the different perspectives of application scenarios, including XAI
security issues could be a future research direction. applications in defending against different categories of
cyber attacks, XAI for cyber security in distinct industrial
4) MULTIMODAL XAI
applications, and cyber threats targeting XAI models and
Multimodal information of text, video, audio, and images in
corresponding defensive approaches. Some common cyber
the same context can all be easily understood by people. The
attacks including malware, spam, fraud, DoS, DGAs,
benefit of multimodality is its capacity to gather and combine
phishing, network intrusion, and botnet were introduced. The
important and comprehensive data from a range of sources,
corresponding defensive mechanisms utilizing XAI against
enabling a far richer depiction of the issue at hand. In some
them were presented. The implementation of XAI in various
cyber security industrial areas, such as healthcare, medical
industrial areas namely in smart healthcare, smart financial
decisions are primarily driven by a variety of influencing
systems, smart agriculture, smart cities, smart transportation,
variables originating from a plurality of underlying signals
and Human-Computer Interaction were described
and information bases, which highlights the need for
exhausively. Distinct approaches of cyber attacks targeting
multimodality at every stage. On the other hand, due to the
XAI models and the related defensive methods were
application of XAI in these areas, multimodal XAI could be
introduced as well. In continuation to these, we pointed out
developed in near future.
and discussed some challenges, key insights and research
5) ADVERSARIAL ATTACKS AND DEFENSES directions of XAI applications in cyber security. We hope
As we discussed in this review, although XAI could be that this paper could serve as a reference for researchers,
applied in cyber security to prevent cyber attacks, the XAI developers, and security professionals who are interested in
model performance and explainability could be attacked as using XAI models to solve challenging issues in cyber
well. Other than that, the adversarial inputs to the sample security domains.
data should be paid attention to as well. Some researchers
[263] have already developed powerful tools to fool the state-
of-art XAI methods including LIME and SHAP. However, REFERENCES
although the cyber threats and corresponding defensive [1] CISA, “What is Cybersecurity? | CISA,” What is Cybersecurity?
mechanisms focusing on the performance of AI models have https://www.cisa.gov/uscert/ncas/tips/ST04-001 (accessed Jul. 01,
2022).
been studied recently, the adversarial attacks and defenses [2] D. S. Berman, A. L. Buczak, J. S. Chavis, and C. L. Corbett, “A
against the explainability of XAI models still require further Survey of Deep Learning Methods for Cyber Security,” Information,
research. vol. 10, no. 4, Art. no. 4, Apr. 2019, doi: 10.3390/info10040122.
[3] “Number of internet users worldwide 2021,” Statista.
6) PROTECTION OF DATA https://www.statista.com/statistics/273018/number-of-internet-users-
In cyber security areas, confidentiality and protection of data worldwide/ (accessed Jul. 01, 2022).
[4] “2021 Cyber Attack Trends Mid-Year Report | Check Point
are significant issues as privacy and ethical issues are
Software.” https://pages.checkpoint.com/cyber-attack-2021-
highlighted recently. For XAI-based systems, the situation is trends.html (accessed Jul. 01, 2022).
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
[5] “Cyberattack disrupts unemployment benefits in some states,” [23] R. A. Alves and D. Costa, “A Survey of Random Forest Based
Washington Post. Accessed: Jul. 02, 2022. [Online]. Available: Methods for Intrusion Detection Systems,” ACM Computing Surveys
https://www.washingtonpost.com/politics/cyberattack-disrupts- (CSUR), May 2018, doi: 10.1145/3178582.
unemployment-benefits-in-some-states/2022/06/30/8f8fe138-f88a- [24] J. Li, “Cyber security meets artificial intelligence: a survey,”
11ec-81db-ac07a394a86b_story.html Frontiers Inf Technol Electronic Eng, vol. 19, no. 12, pp. 1462–1474,
[6] “Threat Landscape,” ENISA. Dec. 2018, doi: 10.1631/FITEE.1800573.
https://www.enisa.europa.eu/topics/threat-risk-management/threats- [25] I. Ahmed, G. Jeon, and F. Piccialli, “From Artificial Intelligence to
and-trends (accessed Jul. 02, 2022). Explainable Artificial Intelligence in Industry 4.0: A Survey on
[7] D. Gümüşbaş, T. Yıldırım, A. Genovese, and F. Scotti, “A What, How, and Where,” IEEE Transactions on Industrial
Comprehensive Survey of Databases and Deep Learning Methods Informatics, vol. 18, no. 8, pp. 5031–5042, 2022, doi:
for Cybersecurity and Intrusion Detection Systems,” IEEE Systems 10.1109/TII.2022.3146552.
Journal, vol. 15, no. 2, pp. 1717–1731, Jun. 2021, doi: [26] A. Kuppa and N.-A. Le-Khac, “Adversarial XAI Methods in
10.1109/JSYST.2020.2992966. Cybersecurity,” IEEE Transactions on Information Forensics and
[8] S. Zeadally, E. Adi, Z. Baig, and I. A. Khan, “Harnessing Artificial Security, vol. 16, pp. 4924–4938, 2021, doi:
Intelligence Capabilities to Improve Cybersecurity,” IEEE Access, 10.1109/TIFS.2021.3117075.
vol. 8, pp. 23817–23837, 2020, doi: [27] S. Mane and D. Rao, “Explaining Network Intrusion Detection
10.1109/ACCESS.2020.2968045. System Using Explainable AI Framework.” arXiv, Mar. 12, 2021.
[9] S. M. Mathews, “Explainable Artificial Intelligence Applications in doi: 10.48550/arXiv.2103.07110.
NLP, Biomedical, and Malware Classification: A Literature [28] “Survey of AI in Cybersecurity for Information Technology
Review,” in Intelligent Computing, Cham, 2019, pp. 1269–1292. doi: Management | IEEE Conference Publication | IEEE Xplore.”
10.1007/978-3-030-22868-2_90. https://ieeexplore.ieee.org/document/8813605 (accessed Jul. 05,
[10] “Explainable Artificial Intelligence for Tabular Data: A Survey | 2022).
IEEE Journals & Magazine | IEEE Xplore.” [29] S. Mahdavifar and A. A. Ghorbani, “Application of deep learning to
https://ieeexplore.ieee.org/document/9551946 (accessed Jul. 02, cybersecurity: A survey,” Neurocomputing, vol. 347, pp. 149–176,
2022). Jun. 2019, doi: 10.1016/j.neucom.2019.02.056.
[11] B. Goodman and S. Flaxman, “European Union Regulations on [30] G. Srivastava et al., “XAI for Cybersecurity: State of the Art,
Algorithmic Decision-Making and a ‘Right to Explanation,’” AI Challenges, Open Issues and Future Directions.” arXiv, Jun. 02,
Magazine, vol. 38, no. 3, Art. no. 3, Oct. 2017, doi: 2022. doi: 10.48550/arXiv.2206.03585.
10.1609/aimag.v38i3.2741. [31] “AI-Driven Cybersecurity: An Overview, Security Intelligence
[12] “A Systematic Review of Human–Computer Interaction and Modeling and Research Directions | SpringerLink.”
Explainable Artificial Intelligence in Healthcare With Artificial https://link.springer.com/article/10.1007/s42979-021-00557-0
Intelligence Techniques | IEEE Journals & Magazine | IEEE (accessed Jul. 05, 2022).
Xplore.” https://ieeexplore.ieee.org/document/9614151 (accessed Jul. [32] M. Humayun, M. Niazi, N. Jhanjhi, M. Alshayeb, and S. Mahmood,
02, 2022). “Cyber Security Threats and Vulnerabilities: A Systematic Mapping
[13] H. Jiang, J. Nagra, and P. Ahammad, “SoK: Applying Machine Study,” Arab J Sci Eng, vol. 45, no. 4, pp. 3171–3189, Apr. 2020,
Learning in Security - A Survey,” Nov. 2016. doi: 10.1007/s13369-019-04319-2.
[14] “A Survey of Data Mining and Machine Learning Methods for [33] K. Shaukat, S. Luo, V. Varadharajan, I. A. Hameed, and M. Xu, “A
Cyber Security Intrusion Detection | IEEE Journals & Magazine | Survey on Machine Learning Techniques for Cyber Security in the
IEEE Xplore.” https://ieeexplore.ieee.org/document/7307098 Last Decade,” IEEE Access, vol. 8, pp. 222310–222354, 2020, doi:
(accessed Jul. 05, 2022). 10.1109/ACCESS.2020.3041951.
[15] D. Kwon, H. Kim, J. Kim, S. C. Suh, I. Kim, and K. J. Kim, “A [34] A. Bécue, I. Praça, and J. Gama, “Artificial intelligence, cyber-
survey of deep learning-based network anomaly detection,” Cluster threats and Industry 4.0: challenges and opportunities,” Artif Intell
Comput, vol. 22, no. 1, pp. 949–961, Jan. 2019, doi: Rev, vol. 54, no. 5, pp. 3849–3886, Jun. 2021, doi: 10.1007/s10462-
10.1007/s10586-017-1117-8. 020-09942-2.
[16] A. P. Veiga, “Applications of Artificial Intelligence to Network [35] I. Kok, F. Y. Okay, O. Muyanli, and S. Ozdemir, “Explainable
Security.” arXiv, Mar. 27, 2018. doi: 10.48550/arXiv.1803.09992. Artificial Intelligence (XAI) for Internet of Things: A Survey.”
[17] D. Ucci, L. Aniello, and R. Baldoni, “Survey of machine learning arXiv, Jun. 07, 2022. doi: 10.48550/arXiv.2206.04800.
techniques for malware analysis,” Computers & Security, vol. 81, pp. [36] M. Macas, C. Wu, and W. Fuertes, “A survey on deep learning for
123–147, Mar. 2019, doi: 10.1016/j.cose.2018.11.001. cybersecurity: Progress, challenges, and opportunities,” Computer
[18] P. Mishra, V. Varadharajan, U. Tupakula, and E. S. Pilli, “A Networks, vol. 212, p. 109032, Jul. 2022, doi:
Detailed Investigation and Analysis of Using Machine Learning 10.1016/j.comnet.2022.109032.
Techniques for Intrusion Detection,” IEEE Communications Surveys [37] 14:00-17:00, “ISO/IEC 27032:2012,” ISO.
& Tutorials, vol. 21, no. 1, pp. 686–728, 2019, doi: https://www.iso.org/cms/render/live/en/sites/isoorg/contents/data/sta
10.1109/COMST.2018.2847722. ndard/04/43/44375.html (accessed Jul. 05, 2022).
[19] C. Rudin, “Stop Explaining Black Box Machine Learning Models [38] “What is a Cyber Attack?,” Check Point Software.
for High Stakes Decisions and Use Interpretable Models Instead,” https://www.checkpoint.com/cyber-hub/cyber-security/what-is-
arXiv:1811.10154 [cs, stat], Sep. 2019, Accessed: Apr. 28, 2022. cyber-attack/ (accessed Jul. 05, 2022).
[Online]. Available: http://arxiv.org/abs/1811.10154 [39] “Cybersecurity Market worth $345.4 billion by 2026.”
[20] Z. Lv, Y. Han, A. K. Singh, G. Manogaran, and H. Lv, https://www.marketsandmarkets.com/PressReleases/cyber-
“Trustworthiness in Industrial IoT Systems Based on Artificial security.asp (accessed Jul. 05, 2022).
Intelligence,” IEEE Transactions on Industrial Informatics, vol. 17, [40] M.-A. Clinciu and H. Hastie, “A Survey of Explainable AI
no. 2, pp. 1496–1504, 2021, doi: 10.1109/TII.2020.2994747. Terminology,” in Proceedings of the 1st Workshop on Interactive
[21] C. S. Wickramasinghe, D. L. Marino, K. Amarasinghe, and M. Natural Language Technology for Explainable Artificial Intelligence
Manic, “Generalization of Deep Learning for Cyber-Physical System (NL4XAI 2019), 2019, pp. 8–13. doi: 10.18653/v1/W19-8403.
Security: A Survey,” in IECON 2018 - 44th Annual Conference of [41] O. Biran and C. V. Cotton, “Explanation and Justification in
the IEEE Industrial Electronics Society, 2018, pp. 745–751. doi: Machine Learning : A Survey Or,” undefined, 2017, Accessed: Jul.
10.1109/IECON.2018.8591773. 08, 2022. [Online]. Available:
[22] P. A. A. Resende and A. C. Drummond, “A Survey of Random https://www.semanticscholar.org/paper/Explanation-and-
Forest Based Methods for Intrusion Detection Systems,” ACM Justification-in-Machine-Learning-%3A-Biran-
Comput. Surv., 2018, doi: 10.1145/3178582. Cotton/02e2e79a77d8aabc1af1900ac80ceebac20abde4
[42] T. Speith, “A Review of Taxonomies of Explainable Artificial
Intelligence (XAI) Methods,” in 2022 ACM Conference on Fairness,
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
Accountability, and Transparency, New York, NY, USA, Jun. 2022, Gradient-based Localization,” Int J Comput Vis, vol. 128, no. 2, pp.
pp. 2239–2250. doi: 10.1145/3531146.3534639. 336–359, Feb. 2020, doi: 10.1007/s11263-019-01228-7.
[43] S. Han, M. Xie, H.-H. Chen, and Y. Ling, “Intrusion Detection in [62] S. Wachter, B. Mittelstadt, and C. Russell, “Counterfactual
Cyber-Physical Systems: Techniques and Challenges,” IEEE Explanations Without Opening the Black Box: Automated Decisions
Systems Journal, vol. 8, no. 4, pp. 1052–1062, 2014, doi: and the GDPR.” Rochester, NY, Oct. 06, 2017. doi:
10.1109/JSYST.2013.2257594. 10.2139/ssrn.3063289.
[44] R. Donida Labati, A. Genovese, V. Piuri, F. Scotti, and S. [63] M. Ibrahim, M. Louie, C. Modarres, and J. Paisley, “Global
Vishwakarma, “Computational Intelligence in Cloud Computing,” in Explanations of Neural Networks: Mapping the Landscape of
Recent Advances in Intelligent Engineering: Volume Dedicated to Predictions.” arXiv, Feb. 06, 2019. doi: 10.48550/arXiv.1902.02384.
Imre J. Rudas’ Seventieth Birthday, L. Kovács, T. Haidegger, and A. [64] H. Liu, Q. Yin, and W. Y. Wang, “Towards Explainable NLP: A
Szakál, Eds. Cham: Springer International Publishing, 2020, pp. Generative Explanation Framework for Text Classification.” arXiv,
111–127. doi: 10.1007/978-3-030-14350-3_6. Jun. 11, 2019. doi: 10.48550/arXiv.1811.00196.
[45] R. A. Nafea and M. Amin Almaiah, “Cyber Security Threats in [65] M. Danilevsky, K. Qian, R. Aharonov, Y. Katsis, B. Kawas, and P.
Cloud: Literature Review,” in 2021 International Conference on Sen, “A Survey of the State of Explainable AI for Natural Language
Information Technology (ICIT), Jul. 2021, pp. 779–786. doi: Processing.” arXiv, Oct. 01, 2020. doi: 10.48550/arXiv.2010.00711.
10.1109/ICIT52682.2021.9491638. [66] J. V. Jeyakumar, J. Noor, Y.-H. Cheng, L. Garcia, and M. Srivastava,
[46] “Black Box Attacks on Explainable Artificial Intelligence(XAI) “How Can I Explain This to You? An Empirical Study of Deep
methods in Cyber Security | IEEE Conference Publication | IEEE Neural Network Explanation Methods,” in Advances in Neural
Xplore.” https://ieeexplore.ieee.org/abstract/document/9206780 Information Processing Systems, 2020, vol. 33, pp. 4211–4222.
(accessed Jul. 08, 2022). Accessed: Jul. 09, 2022. [Online]. Available:
[47] K. D. Ahmed and S. Askar, “Deep Learning Models for Cyber https://proceedings.neurips.cc/paper/2020/hash/2c29d89cc56cdb191
Security in IoT Networks: A Review,” International Journal of c60db2f0bae796b-Abstract.html
Science and Business, vol. 5, no. 3, pp. 61–70, 2021. [67] W. Jin, X. Li, and G. Hamarneh, “Evaluating Explainable AI on a
[48] J. Gerlings, A. Shollo, and I. Constantiou, “Reviewing the Need for Multi-Modal Medical Imaging Task: Can Existing Algorithms
Explainable Artificial Intelligence (xAI).” arXiv, Jan. 26, 2021. doi: Fulfill Clinical Requirements?” arXiv, Mar. 12, 2022. doi:
10.48550/arXiv.2012.01007. 10.48550/arXiv.2203.06487.
[49] T. Perarasi, S. Vidhya, L. Moses M., and P. Ramya, “Malicious [68] J. Lu, D. Lee, T. W. Kim, and D. Danks, “Good Explanation for
Vehicles Identifying and Trust Management Algorithm for Enhance Algorithmic Transparency.” Rochester, NY, Nov. 11, 2019. doi:
the Security in 5G-VANET,” in 2020 Second International 10.2139/ssrn.3503603.
Conference on Inventive Research in Computing Applications [69] L. Amgoud and H. Prade, “Using arguments for making and
(ICIRCA), Jul. 2020, pp. 269–275. doi: explaining decisions,” Artificial Intelligence, vol. 173, no. 3, pp.
10.1109/ICIRCA48905.2020.9183184. 413–436, Mar. 2009, doi: 10.1016/j.artint.2008.11.006.
[50] G. Jaswal, V. Kanhangad, and R. Ramachandra, AI and Deep [70] M. Wu, M. Hughes, S. Parbhoo, M. Zazzi, V. Roth, and F. Doshi-
Learning in Biometric Security: Trends, Potential, and Challenges. Velez, “Beyond Sparsity: Tree Regularization of Deep Models for
CRC Press, 2021. Interpretability,” Proceedings of the AAAI Conference on Artificial
[51] “What is GDPR, the EU’s new data protection law?,” GDPR.eu, Intelligence, vol. 32, no. 1, Art. no. 1, Apr. 2018, doi:
Nov. 07, 2018. https://gdpr.eu/what-is-gdpr/ (accessed Jul. 08, 2022). 10.1609/aaai.v32i1.11501.
[52] C. T. Wolf, “Explainability scenarios: towards scenario-based XAI [71] H. Lakkaraju, E. Kamar, R. Caruana, and J. Leskovec, “Faithful and
design,” in Proceedings of the 24th International Conference on Customizable Explanations of Black Box Models,” in Proceedings
Intelligent User Interfaces, New York, NY, USA, Mar. 2019, pp. of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, New
252–257. doi: 10.1145/3301275.3302317. York, NY, USA, Jan. 2019, pp. 131–138. doi:
[53] A. Barredo Arrieta et al., “Explainable Artificial Intelligence (XAI): 10.1145/3306618.3314229.
Concepts, taxonomies, opportunities and challenges toward [72] G. Fidel, R. Bitton, and A. Shabtai, “When Explainability Meets
responsible AI,” Information Fusion, vol. 58, pp. 82–115, Jun. 2020, Adversarial Learning: Detecting Adversarial Examples using SHAP
doi: 10.1016/j.inffus.2019.12.012. Signatures.” arXiv, Sep. 08, 2019. doi: 10.48550/arXiv.1909.03418.
[54] D. V. Carvalho, E. M. Pereira, and J. S. Cardoso, “Machine Learning [73] W. Guo, “Explainable Artificial Intelligence for 6G: Improving
Interpretability: A Survey on Methods and Metrics,” Electronics, vol. Trust between Human and Machine,” IEEE Communications
8, no. 8, Art. no. 8, Aug. 2019, doi: 10.3390/electronics8080832. Magazine, vol. 58, no. 6, pp. 39–45, Jun. 2020, doi:
[55] V. Arya et al., “One Explanation Does Not Fit All: A Toolkit and 10.1109/MCOM.001.2000050.
Taxonomy of AI Explainability Techniques.” arXiv, Sep. 14, 2019. [74] F. Hussain, R. Hussain, and E. Hossain, “Explainable Artificial
doi: 10.48550/arXiv.1909.03012. Intelligence (XAI): An Engineering Perspective.” arXiv, Jan. 10,
[56] M. T. Ribeiro, S. Singh, and C. Guestrin, “‘Why Should I Trust 2021. doi: 10.48550/arXiv.2101.03613.
You?’: Explaining the Predictions of Any Classifier.” arXiv, Aug. 09, [75] D. Slack, S. Hilgard, E. Jia, S. Singh, and H. Lakkaraju, “Fooling
2016. doi: 10.48550/arXiv.1602.04938. LIME and SHAP: Adversarial Attacks on Post hoc Explanation
[57] A. Altmann, L. Toloşi, O. Sander, and T. Lengauer, “Permutation Methods.” arXiv, Feb. 03, 2020. doi: 10.48550/arXiv.1911.02508.
importance: a corrected feature importance measure,” Bioinformatics, [76] N. Papernot et al., “Technical Report on the CleverHans v2.1.0
vol. 26, no. 10, pp. 1340–1347, May 2010, doi: Adversarial Examples Library.” arXiv, Jun. 27, 2018. doi:
10.1093/bioinformatics/btq134. 10.48550/arXiv.1610.00768.
[58] R. Ying, D. Bourgeois, J. You, M. Zitnik, and J. Leskovec, [77] D. Gunning, M. Stefik, J. Choi, T. Miller, S. Stumpf, and G.-Z. Yang,
“GNNExplainer: Generating Explanations for Graph Neural “XAI—Explainable artificial intelligence,” Science Robotics, vol. 4,
Networks.” arXiv, Nov. 13, 2019. doi: 10.48550/arXiv.1903.03894. no. 37, p. eaay7120, Dec. 2019, doi: 10.1126/scirobotics.aay7120.
[59] S. M. Lundberg and S.-I. Lee, “A Unified Approach to Interpreting [78] C. Mars, R. Dès, and M. Boussard, “The three stages of Explainable
Model Predictions,” in Advances in Neural Information Processing AI: How explainability facilitates real world deployment of AI,” Jan.
Systems, 2017, vol. 30. Accessed: Jul. 09, 2022. [Online]. Available: 2020.
https://proceedings.neurips.cc/paper/2017/hash/8a20a8621978632d7 [79] L. Longo, R. Goebel, F. Lecue, P. Kieseberg, and A. Holzinger,
6c43dfd28b67767-Abstract.html Explainable Artificial Intelligence: Concepts, Applications,
[60] R. Iyer, Y. Li, H. Li, M. Lewis, R. Sundar, and K. Sycara, Research Challenges and Visions. 2020, p. 16. doi: 10.1007/978-3-
“Transparency and Explanation in Deep Reinforcement Learning 030-57321-8_1.
Neural Networks.” arXiv, Sep. 17, 2018. doi: [80] “Evaluation of Post-hoc XAI Approaches Through Synthetic
10.48550/arXiv.1809.06061. Tabular Data | Foundations of Intelligent Systems,” Guide
[61] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and Proceedings. https://dl.acm.org/doi/abs/10.1007/978-3-030-59491-
D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via 6_40 (accessed Jul. 10, 2022).
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
[81] L. Arras, A. Osman, and W. Samek, “CLEVR-XAI: A benchmark [101] “Drebin: Effective and Explainable Detection of Android Malware
dataset for the ground truth evaluation of neural network in Your Pocket – NDSS Symposium.” https://www.ndss-
explanations,” Information Fusion, vol. 81, pp. 14–40, May 2022, symposium.org/ndss2014/programme/drebin-effective-and-
doi: 10.1016/j.inffus.2021.11.008. explainable-detection-android-malware-your-pocket/ (accessed Jul.
[82] A. Rai, “Explainable AI: from black box to glass box,” J. of the 13, 2022).
Acad. Mark. Sci., vol. 48, no. 1, pp. 137–141, Jan. 2020, doi: [102] T. A. Almeida, J. M. G. Hidalgo, and A. Yamakami, “Contributions
10.1007/s11747-019-00710-5. to the study of SMS spam filtering: new collection and results,” in
[83] E. LEMONNE, “Ethics Guidelines for Trustworthy AI,” Proceedings of the 11th ACM symposium on Document engineering,
FUTURIUM - European Commission, Dec. 17, 2018. New York, NY, USA, Sep. 2011, pp. 259–262. doi:
https://ec.europa.eu/futurium/en/ai-alliance-consultation (accessed 10.1145/2034691.2034742.
Jul. 10, 2022). [103] V. Metsis, I. Androutsopoulos, and G. Paliouras, “Spam Filtering
[84] European Parliament. Directorate General for Parliamentary with Naive Bayes - Which Naive Bayes?,” Jan. 2006.
Research Services., The impact of the general data protection [104] M. S. I. Mamun, M. A. Rathore, A. H. Lashkari, N. Stakhanova, and
regulation on artificial intelligence. LU: Publications Office, 2020. A. A. Ghorbani, “Detecting Malicious URLs Using Lexical
Accessed: Jul. 10, 2022. [Online]. Available: Analysis,” in Network and System Security, Cham, 2016, pp. 467–
https://data.europa.eu/doi/10.2861/293 482. doi: 10.1007/978-3-319-46298-1_30.
[85] M. Ebers, “Regulating Explainable AI in the European Union. An [105] M. Tavallaee, E. Bagheri, W. Lu, and A. A. Ghorbani, “A detailed
Overview of the Current Legal Framework(s).” Rochester, NY, Aug. analysis of the KDD CUP 99 data set,” in 2009 IEEE Symposium on
09, 2021. doi: 10.2139/ssrn.3901732. Computational Intelligence for Security and Defense Applications,
[86] Z. C. Lipton, “The Mythos of Model Interpretability: In machine Jul. 2009, pp. 1–6. doi: 10.1109/CISDA.2009.5356528.
learning, the concept of interpretability is both important and [106] A. Shiravi, H. Shiravi, M. Tavallaee, and A. A. Ghorbani, “Toward
slippery.,” Queue, vol. 16, no. 3, pp. 31–57, Jun. 2018, doi: developing a systematic approach to generate benchmark datasets for
10.1145/3236386.3241340. intrusion detection,” Computers & Security, vol. 31, no. 3, pp. 357–
[87] “Explainability for artificial intelligence in healthcare: a 374, May 2012, doi: 10.1016/j.cose.2011.12.012.
multidisciplinary perspective | BMC Medical Informatics and [107] C. Kolias, G. Kambourakis, A. Stavrou, and S. Gritzalis, “Intrusion
Decision Making | Full Text.” Detection in 802.11 Networks: Empirical Evaluation of Threats and
https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/ a Public Dataset,” IEEE Communications Surveys & Tutorials, vol.
s12911-020-01332-6 (accessed Jul. 11, 2022). 18, no. 1, pp. 184–208, 2016, doi: 10.1109/COMST.2015.2402161.
[88] A. Holzinger, “Explainable AI and Multi-Modal Causability in [108] “Toward Generating a New Intrusion Detection Dataset and
Medicine,” i-com, vol. 19, no. 3, pp. 171–179, Dec. 2020, doi: Intrusion Traffic Characterization | Request PDF.”
10.1515/icom-2020-0024. https://www.researchgate.net/publication/322870768_Toward_Gene
[89] S. Wachter, B. Mittelstadt, and C. Russell, “Why Fairness Cannot Be rating_a_New_Intrusion_Detection_Dataset_and_Intrusion_Traffic_
Automated: Bridging the Gap Between EU Non-Discrimination Law Characterization (accessed Jul. 13, 2022).
and AI,” SSRN Journal, 2020, doi: 10.2139/ssrn.3547922. [109] I. Sharafaldin, A. H. Lashkari, S. Hakak, and A. A. Ghorbani,
[90] A. Holzinger, G. Langs, H. Denk, K. Zatloukal, and H. Müller, “Developing Realistic Distributed Denial of Service (DDoS) Attack
“Causability and explainability of artificial intelligence in medicine,” Dataset and Taxonomy,” in 2019 International Carnahan
WIREs Data Mining and Knowledge Discovery, vol. 9, no. 4, p. Conference on Security Technology (ICCST), 2019, pp. 1–8. doi:
e1312, 2019, doi: 10.1002/widm.1312. 10.1109/CCST.2019.8888419.
[91] S. Tonekaboni, S. Joshi, M. D. McCradden, and A. Goldenberg, [110] A. Alsaedi, N. Moustafa, Z. Tari, A. Mahmood, and A. Anwar,
“What Clinicians Want: Contextualizing Explainable Machine “TON_IoT Telemetry Dataset: A New Generation Dataset of IoT
Learning for Clinical End Use.” arXiv, Aug. 07, 2019. doi: and IIoT for Data-Driven Intrusion Detection Systems,” IEEE
10.48550/arXiv.1905.05134. Access, vol. 8, pp. 165130–165150, 2020, doi:
[92] O. Yavanoglu and M. Aydos, “A review on cyber security datasets 10.1109/ACCESS.2020.3022862.
for machine learning algorithms,” in 2017 IEEE International [111] R. Damasevicius et al., “LITNET-2020: An Annotated Real-World
Conference on Big Data (Big Data), 2017, pp. 2186–2193. doi: Network Flow Dataset for Network Intrusion Detection,” Electronics,
10.1109/BigData.2017.8258167. vol. 9, no. 5, Art. no. 5, May 2020, doi: 10.3390/electronics9050800.
[93] Y. Xin et al., “Machine Learning and Deep Learning Methods for [112] G. Creech and J. Hu, “Generation of a new IDS test dataset: Time to
Cybersecurity,” IEEE Access, vol. 6, pp. 35365–35381, 2018, doi: retire the KDD collection,” in 2013 IEEE Wireless Communications
10.1109/ACCESS.2018.2836950. and Networking Conference (WCNC), Apr. 2013, pp. 4487–4492.
[94] Y. Meidan et al., “N-BaIoT—Network-Based Detection of IoT doi: 10.1109/WCNC.2013.6555301.
Botnet Attacks Using Deep Autoencoders,” IEEE Pervasive [113] N. Moustafa and J. Slay, “UNSW-NB15: a comprehensive data set
Computing, vol. 17, no. 3, pp. 12–22, 2018, doi: for network intrusion detection systems (UNSW-NB15 network data
10.1109/MPRV.2018.03367731. set),” in 2015 Military Communications and Information Systems
[95] Y. M. P. Pa, S. Suzuki, K. Yoshioka, T. Matsumoto, T. Kasama, and Conference (MilCIS), 2015, pp. 1–6. doi:
C. Rossow, “IoTPOT: A Novel Honeypot for Revealing Current IoT 10.1109/MilCIS.2015.7348942.
Threats,” Journal of Information Processing, vol. 24, no. 3, pp. 522– [114] S. García, M. Grill, J. Stiborek, and A. Zunino, “An empirical
533, 2016, doi: 10.2197/ipsjjip.24.522. comparison of botnet detection methods,” Computers & Security, vol.
[96] S. Garcia, A. Parmisano, and M. J. Erquiaga, “IoT-23: A labeled 45, pp. 100–123, Sep. 2014, doi: 10.1016/j.cose.2014.05.011.
dataset with malicious and benign IoT network traffic.” Zenodo, [115] S. Saad et al., “Detecting P2P botnets through network behavior
2020. doi: 10.5281/zenodo.4743746. analysis and machine learning,” in 2011 Ninth Annual International
[97] H. S. Anderson and P. Roth, “EMBER: An Open Dataset for Conference on Privacy, Security and Trust, Jul. 2011, pp. 174–180.
Training Static PE Malware Machine Learning Models.” arXiv, Apr. doi: 10.1109/PST.2011.5971980.
16, 2018. doi: 10.48550/arXiv.1804.04637. [116] N. Koroniotis, N. Moustafa, E. Sitnikova, and B. Turnbull,
[98] Y. Zhou and X. Jiang, “Dissecting Android Malware: “Towards the development of realistic botnet dataset in the Internet
Characterization and Evolution,” in 2012 IEEE Symposium on of Things for network forensic analytics: Bot-IoT dataset,” Future
Security and Privacy, 2012, pp. 95–109. doi: 10.1109/SP.2012.16. Generation Computer Systems, vol. 100, pp. 779–796, Nov. 2019,
[99] “VirusShare.com.” https://virusshare.com/ (accessed Jul. 13, 2022). doi: 10.1016/j.future.2019.05.041.
[100] A. H. Lashkari, A. F. A. Kadir, L. Taheri, and A. A. Ghorbani, [117] M. Zago, M. Gil Pérez, and G. Martínez Pérez, “UMUDGA: A
“Toward Developing a Systematic Approach to Generate dataset for profiling DGA-based botnet,” Computers & Security, vol.
Benchmark Android Malware Datasets and Classification,” in 2018 92, p. 101719, May 2020, doi: 10.1016/j.cose.2020.101719.
International Carnahan Conference on Security Technology [118]R. Vinayakumar, K. P. Soman, P. Poornachandran, M. Alazab, and S.
(ICCST), 2018, pp. 1–7. doi: 10.1109/CCST.2018.8585560. M. Thampi, “AmritaDGA: A comprehensive data set for domain
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
generation algorithms (DGAs) based domain name detection IEEE Transactions on Intelligent Transportation Systems, vol. 23,
systems and application of deep learning,” in Big Data no. 4, pp. 3302–3312, Apr. 2022, doi: 10.1109/TITS.2020.3035596.
Recommender Systems, vol. 2, O. Khalid, S. U. Khan, and A. Y. [137] J. Yuan, Y. Zheng, X. Xie, and G. Sun, “Driving with knowledge
Zomaya, Eds. Stevenage: Institution of Engineering and Technology, from the physical world,” in Proceedings of the 17th ACM SIGKDD
2019, pp. 455–485. doi: 10.1049/PBPC035G_ch22. international conference on Knowledge discovery and data mining,
[119] H.-K. Shin, W. Lee, J.-H. Yun, and H. Kim, “{HAI} 1.0: {HIL- New York, NY, USA, Aug. 2011, pp. 316–324. doi:
based} Augmented {ICS} Security Dataset,” 2020. Accessed: Jul. 13, 10.1145/2020408.2020462.
2022. [Online]. Available: [138] Y. Zheng, L. Zhang, X. Xie, and W.-Y. Ma, “Mining interesting
https://www.usenix.org/conference/cset20/presentation/shin locations and travel sequences from GPS trajectories,” in
[120] R. C. Borges Hink, J. M. Beaver, M. A. Buckner, T. Morris, U. Proceedings of the 18th international conference on World wide web,
Adhikari, and S. Pan, “Machine learning for power system New York, NY, USA, Apr. 2009, pp. 791–800. doi:
disturbance and cyber-attack discrimination,” in 2014 7th 10.1145/1526709.1526816.
International Symposium on Resilient Control Systems (ISRCS), [139] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun, “Vision meets
2014, pp. 1–8. doi: 10.1109/ISRCS.2014.6900095. robotics: The KITTI dataset,” The International Journal of Robotics
[121] M. S. Elsayed, N.-A. Le-Khac, and A. D. Jurcut, “InSDN: A Novel Research, vol. 32, no. 11, pp. 1231–1237, Sep. 2013, doi:
SDN Intrusion Dataset,” IEEE Access, vol. 8, pp. 165263–165284, 10.1177/0278364913491297.
2020, doi: 10.1109/ACCESS.2020.3022633. [140] D. P. Hughes and M. Salathe, “An open access repository of images
[122] K. Marek et al., “The Parkinson Progression Marker Initiative on plant health to enable the development of mobile disease
(PPMI),” Progress in Neurobiology, vol. 95, no. 4, pp. 629–635, Dec. diagnostics.” arXiv, Apr. 11, 2016. doi: 10.48550/arXiv.1511.08060.
2011, doi: 10.1016/j.pneurobio.2011.09.005. [141] “photometric stereo-based 3D imaging system using computer vision
[123] L. Cui and D. Lee, “CoAID: COVID-19 Healthcare Misinformation and deep learning for tracking plant growth | GigaScience | Oxford
Dataset.” arXiv, Nov. 03, 2020. doi: 10.48550/arXiv.2006.00885. Academic.”
[124] D. Dave, H. Naik, S. Singhal, and P. Patel, “Explainable AI meets https://academic.oup.com/gigascience/article/8/5/giz056/5498634?lo
Healthcare: A Study on Heart Disease Dataset.” arXiv, Nov. 06, gin=true (accessed Jul. 14, 2022).
2020. doi: 10.48550/arXiv.2011.03195. [142] R. Thapa, N. Snavely, S. Belongie, and A. Khan, “The Plant
[125] A. E. W. Johnson et al., “MIMIC-III, a freely accessible critical care Pathology 2020 challenge dataset to classify foliar disease of
database,” Sci Data, vol. 3, no. 1, Art. no. 1, May 2016, doi: apples.” arXiv, Apr. 24, 2020. doi: 10.48550/arXiv.2004.11958.
10.1038/sdata.2016.35. [143] E. Vural, J. Huang, D. Hou, and S. Schuckers, “Shared research
[126] M. Saeed et al., “Multiparameter Intelligent Monitoring in Intensive dataset to support development of keystroke authentication,” in
Care II: a public-access intensive care unit database,” Crit Care Med, IEEE International Joint Conference on Biometrics, Sep. 2014, pp.
vol. 39, no. 5, pp. 952–960, May 2011, doi: 1–8. doi: 10.1109/BTAS.2014.6996259.
10.1097/CCM.0b013e31820a92c6. [144] D. Gunetti and C. Picardi, “Keystroke analysis of free text,” TSEC,
[127] P. Wagner et al., “PTB-XL, a large publicly available 2005, doi: 10.1145/1085126.1085129.
electrocardiography dataset,” Sci Data, vol. 7, no. 1, Art. no. 1, May [145] Y. Sun, H. Ceker, and S. Upadhyaya, “Shared keystroke dataset for
2020, doi: 10.1038/s41597-020-0495-6. continuous authentication,” in 2016 IEEE International Workshop
[128] F. A. Spanhol, L. S. Oliveira, C. Petitjean, and L. Heutte, “Breast on Information Forensics and Security (WIFS), 2016, pp. 1–6. doi:
cancer histopathological image classification using Convolutional 10.1109/WIFS.2016.7823894.
Neural Networks,” in 2016 International Joint Conference on [146] S. Ng, “Opportunities and Challenges: Lessons from Analyzing
Neural Networks (IJCNN), Jul. 2016, pp. 2560–2567. doi: Terabytes of Scanner Data.” National Bureau of Economic Research,
10.1109/IJCNN.2016.7727519. Aug. 2017. doi: 10.3386/w23673.
[129] F. Liu et al., “An Open Access Database for Evaluating the [147] “UCI Machine Learning Repository: Statlog (German Credit Data)
Algorithms of Electrocardiogram Rhythm and Morphology Data Set.”
Abnormality Detection,” Journal of Medical Imaging and Health https://archive.ics.uci.edu/ml/datasets/statlog+(german+credit+data)
Informatics, vol. 8, no. 7, pp. 1368–1373, Sep. 2018, doi: (accessed Jul. 14, 2022).
10.1166/jmihi.2018.2442. [148] T. K. Lengyel, S. Maresca, B. D. Payne, G. D. Webster, S. Vogl, and
[130] Y. Gusev, K. Bhuvaneshwar, L. Song, J.-C. Zenklusen, H. Fine, and A. Kiayias, “Scalability, fidelity and stealth in the DRAKVUF
S. Madhavan, “The REMBRANDT study, a large collection of dynamic malware analysis system,” in Proceedings of the 30th
genomic data from brain cancer patients,” Sci Data, vol. 5, no. 1, Art. Annual Computer Security Applications Conference, New York, NY,
no. 1, Aug. 2018, doi: 10.1038/sdata.2018.158. USA, Dec. 2014, pp. 386–395. doi: 10.1145/2664243.2664252.
[131] R. L. Bowman, Q. Wang, A. Carro, R. G. W. Verhaak, and M. [149] Y. Ye, T. Li, D. Adjeroh, and S. S. Iyengar, “A Survey on Malware
Squatrito, “GlioVis data portal for visualization and analysis of brain Detection Using Data Mining Techniques,” ACM Comput. Surv., vol.
tumor expression datasets,” Neuro-Oncology, vol. 19, no. 1, pp. 50, no. 3, p. 41:1-41:40, Jun. 2017, doi: 10.1145/3073559.
139–141, Jan. 2017, doi: 10.1093/neuonc/now247. [150]R. Vinayakumar, M. Alazab, K. P. Soman, P. Poornachandran, and S.
[132] S. Uppoor, O. Trullols-Cruces, M. Fiore, and J. M. Barcelo-Ordinas, Venkatraman, “Robust Intelligent Malware Detection Using Deep
“Generation and Analysis of a Large-Scale Urban Vehicular Learning,” IEEE Access, vol. 7, pp. 46717–46738, 2019, doi:
Mobility Dataset,” IEEE Transactions on Mobile Computing, vol. 13, 10.1109/ACCESS.2019.2906934.
no. 5, pp. 1061–1075, 2014, doi: 10.1109/TMC.2013.27. [151] A. Yan et al., “Effective detection of mobile malware behavior
[133]P. R. L. de Almeida, L. S. Oliveira, A. S. Britto, E. J. Silva, and A. L. based on explainable deep neural network,” Neurocomputing, vol.
Koerich, “PKLot – A robust dataset for parking lot classification,” 453, pp. 482–492, Sep. 2021, doi: 10.1016/j.neucom.2020.09.082.
Expert Systems with Applications, vol. 42, no. 11, pp. 4937–4949, [152] M. Melis, D. Maiorca, B. Biggio, G. Giacinto, and F. Roli,
Jul. 2015, doi: 10.1016/j.eswa.2015.02.009. “Explaining Black-box Android Malware Detection,” in 2018 26th
[134] “[PDF] Fast Global Alignment Kernels | Semantic Scholar.” European Signal Processing Conference (EUSIPCO), Sep. 2018, pp.
https://www.semanticscholar.org/paper/Fast-Global-Alignment- 524–528. doi: 10.23919/EUSIPCO.2018.8553598.
Kernels-Cuturi/7de1f5079ed7a8a8a5690f72ad2099f52d697393 [153] D. Arp, M. Spreitzenbarth, M. Hübner, H. Gascon, and K. Rieck,
(accessed Jul. 14, 2022). “Drebin: Effective and Explainable Detection of Android Malware
[135] G. Amato, F. Carrara, F. Falchi, C. Gennaro, C. Meghini, and C. in Your Pocket,” San Diego, CA, 2014. doi:
Vairo, “Deep learning for decentralized parking lot occupancy 10.14722/ndss.2014.23247.
detection,” Expert Systems with Applications, vol. 72, pp. 327–334, [154] S. Bose, T. Barao, and X. Liu, “Explaining AI for Malware
Apr. 2017, doi: 10.1016/j.eswa.2016.10.055. Detection: Analysis of Mechanisms of MalConv,” in 2020
[136] G. Oh, D. J. Leblanc, and H. Peng, “Vehicle Energy Dataset (VED), International Joint Conference on Neural Networks (IJCNN), Jul.
A Large-Scale Dataset for Vehicle Energy Consumption Research,” 2020, pp. 1–8. doi: 10.1109/IJCNN48605.2020.9207322.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
[155] E. Raff, J. Barker, J. Sylvester, R. Brandon, B. Catanzaro, and C. K. [174] B. Liu and I. Lane, “Attention-Based Recurrent Neural Network
Nicholas, “Malware Detection by Eating a Whole EXE,” Jun. 2018. Models for Joint Intent Detection and Slot Filling.” arXiv, Sep. 06,
Accessed: Jul. 18, 2022. [Online]. Available: 2016. doi: 10.48550/arXiv.1609.01454.
https://www.aaai.org/ocs/index.php/WS/AAAIW18/paper/view/164 [175] S. S. C. Silva, R. M. P. Silva, R. C. G. Pinto, and R. M. Salles,
22 “Botnets: A survey,” Computer Networks, vol. 57, no. 2, pp. 378–
[156] H. S. Anderson and P. Roth, “EMBER: An Open Dataset for 403, Feb. 2013, doi: 10.1016/j.comnet.2012.07.021.
Training Static PE Malware Machine Learning Models.” arXiv, Apr. [176] “Botnet Detection Market Global Industry Historical Analysis, Size,
16, 2018. doi: 10.48550/arXiv.1804.04637. Growth, Trends, Emerging Factors, Demands, Key Players,
[157] H. Naeem, B. M. Alshammari, and F. Ullah, “Explainable Artificial Emerging Technologies and Potential of Industry till 2027 -
Intelligence-Based IoT Device Malware Detection Mechanism MarketWatch.” https://www.marketwatch.com/press-release/botnet-
Using Image Visualization and Fine-Tuned CNN-Based Transfer detection-market-global-industry-historical-analysis-size-growth-
Learning Model,” Computational Intelligence and Neuroscience, vol. trends-emerging-factors-demands-key-players-emerging-
2022, p. e7671967, Jul. 2022, doi: 10.1155/2022/7671967. technologies-and-potential-of-industry-till-2027-2022-06-29
[158] A. Yan et al., “Effective detection of mobile malware behavior (accessed Jul. 19, 2022).
based on explainable deep neural network,” Neurocomputing, vol. [177] O. Tsemogne, Y. Hayel, C. Kamhoua, and G. Deugoué, “Game-
453, pp. 482–492, Sep. 2021, doi: 10.1016/j.neucom.2020.09.082. Theoretic Modeling of Cyber Deception Against Epidemic Botnets
[159] G. Iadarola, F. Martinelli, F. Mercaldo, and A. Santone, “Towards an in Internet of Things,” IEEE Internet of Things Journal, vol. 9, no. 4,
interpretable deep learning model for mobile malware detection and pp. 2678–2687, 2022, doi: 10.1109/JIOT.2021.3081751.
family identification,” Computers & Security, vol. 105, p. 102198, [178] H. Suryotrisongko, Y. Musashi, A. Tsuneda, and K. Sugitani,
Jun. 2021, doi: 10.1016/j.cose.2021.102198. “Robust Botnet DGA Detection: Blending XAI and OSINT for
[160] S. Wang et al., “TrafficAV: An effective and explainable detection Cyber Threat Intelligence Sharing,” IEEE Access, vol. 10, pp.
of mobile malware behavior using network traffic,” in 2016 34613–34624, 2022, doi: 10.1109/ACCESS.2022.3162588.
IEEE/ACM 24th International Symposium on Quality of Service [179] S. Araki, K. Takahashi, B. Hu, K. Kamiya, and M. Tanikawa,
(IWQoS), Jun. 2016, pp. 1–6. doi: 10.1109/IWQoS.2016.7590446. “Subspace Clustering for Interpretable Botnet Traffic Analysis,” in
[161] M. M. Alani and A. I. Awad, “PAIRED: An Explainable ICC 2019 - 2019 IEEE International Conference on
Lightweight Android Malware Detection System,” IEEE Access, pp. Communications (ICC), 2019, pp. 1–6. doi:
1–1, 2022, doi: 10.1109/ACCESS.2022.3189645. 10.1109/ICC.2019.8761218.
[162] M. Kinkead, S. Millar, N. McLaughlin, and P. O’Kane, “Towards [180] “MAWI Working Group Traffic Archive.”
Explainable CNNs for Android Malware Detection,” Procedia http://mawi.wide.ad.jp/mawi/ (accessed Jul. 19, 2022).
Computer Science, vol. 184, pp. 959–965, Jan. 2021, doi: [181] M. Mazza, S. Cresci, M. Avvenuti, W. Quattrociocchi, and M.
10.1016/j.procs.2021.03.118. Tesconi, “RTbust: Exploiting Temporal Patterns for Botnet
[163] E. G. Dada, J. S. Bassi, H. Chiroma, S. M. Abdulhamid, A. O. Detection on Twitter,” in Proceedings of the 10th ACM Conference
Adetunmbi, and O. E. Ajibuwa, “Machine learning for email spam on Web Science, New York, NY, USA, Jun. 2019, pp. 183–192. doi:
filtering: review, approaches and open research problems,” Heliyon, 10.1145/3292522.3326015.
vol. 5, no. 6, p. e01802, Jun. 2019, doi: [182] H. Bahşi, S. Nõmm, and F. B. La Torre, “Dimensionality Reduction
10.1016/j.heliyon.2019.e01802. for Machine Learning Based IoT Botnet Detection,” in 2018 15th
[164] “Daily number of e-mails worldwide 2025,” Statista. International Conference on Control, Automation, Robotics and
https://www.statista.com/statistics/456500/daily-number-of-e-mails- Vision (ICARCV), 2018, pp. 1857–1862. doi:
worldwide/ (accessed Feb. 21, 2022). 10.1109/ICARCV.2018.8581205.
[165] A. Karim, S. Azam, B. Shanmugam, K. Kannoorpatti, and M. [183] P. P. Kundu, T. Truong-Huu, L. Chen, L. Zhou, and S. G. Teo,
Alazab, “A Comprehensive Survey for Intelligent Spam Email “Detection and Classification of Botnet Traffic using Deep Learning
Detection,” IEEE Access, vol. 7, pp. 168261–168295, 2019, doi: with Model Explanation,” IEEE Transactions on Dependable and
10.1109/ACCESS.2019.2954791. Secure Computing, pp. 1–15, 2022, doi:
[166] R. R. Hoffman, S. T. Mueller, G. Klein, and J. Litman, “Metrics for 10.1109/TDSC.2022.3183361.
Explainable AI: Challenges and Prospects.” arXiv, Feb. 01, 2019. [184] M. M. Alani, “BotStop : Packet-based efficient and explainable IoT
doi: 10.48550/arXiv.1812.04608. botnet detection using machine learning,” Computer
[167] M. Renftle, H. Trittenbach, M. Poznic, and R. Heil, “Explaining Any Communications, vol. 193, pp. 53–62, Sep. 2022, doi:
ML Model? -- On Goals and Capabilities of XAI,” Jun. 2022, doi: 10.1016/j.comcom.2022.06.039.
10.48550/arXiv.2206.13888. [185] D. Buil-Gil, F. Miró-Llinares, A. Moneva, S. Kemp, and N. Díaz-
[168] J. C. S. Reis, A. Correia, F. Murai, A. Veloso, and F. Benevenuto, Castaño, “Cybercrime and shifts in opportunities during COVID-19:
“Explainable Machine Learning for Fake News Detection,” in a preliminary analysis in the UK,” European Societies, vol. 23, no.
Proceedings of the 10th ACM Conference on Web Science, New sup1, pp. S47–S59, Feb. 2021, doi:
York, NY, USA, Jun. 2019, pp. 17–26. doi: 10.1080/14616696.2020.1804973.
10.1145/3292522.3326027. [186] J. Gee and P. M. Button, “The Financial Cost of Fraud 2019,” p. 28.
[169] P. Hacker, R. Krestel, S. Grundmann, and F. Naumann, “Explainable [187] I. Psychoula, A. Gutmann, P. Mainali, S. H. Lee, P. Dunphy, and F.
AI under contract and tort law: legal incentives and technical Petitcolas, “Explainable Machine Learning for Fraud Detection,”
challenges,” Artif Intell Law, vol. 28, no. 4, pp. 415–439, Dec. 2020, Computer, vol. 54, no. 10, pp. 49–59, 2021, doi:
doi: 10.1007/s10506-020-09260-6. 10.1109/MC.2021.3081249.
[170] T. Almeı̇ da, J. M. Hı̇ dalgo, and T. Sı̇ lva, “Towards SMS Spam [188] “IEEE-CIS Fraud Detection.” https://kaggle.com/competitions/ieee-
Filtering: Results under a New Dataset,” International Journal of fraud-detection (accessed Jul. 20, 2022).
Information Security Science, vol. 2, no. 1, Art. no. 1, Mar. 2013. [189] D. Farrugia, C. Zerafa, T. Cini, B. Kuasney, and K. Livori, “A Real-
[171] B. Mathew, P. Saha, S. M. Yimam, C. Biemann, P. Goyal, and A. Time Prescriptive Solution for Explainable Cyber-Fraud Detection
Mukherjee, “HateXplain: A Benchmark Dataset for Explainable Within the iGaming Industry,” SN COMPUT. SCI., vol. 2, no. 3, p.
Hate Speech Detection,” Proceedings of the AAAI Conference on 215, Apr. 2021, doi: 10.1007/s42979-021-00623-7.
Artificial Intelligence, vol. 35, no. 17, Art. no. 17, May 2021. [190] S. X. Rao et al., “xFraud: Explainable Fraud Transaction Detection,”
[172] Z. Zhang, D. Robinson, and J. Tepper, “Detecting Hate Speech on Proc. VLDB Endow., vol. 15, no. 3, pp. 427–436, Nov. 2021, doi:
Twitter Using a Convolution-GRU Based Deep Neural Network,” in 10.14778/3494124.3494128.
The Semantic Web, Cham, 2018, pp. 745–760. doi: 10.1007/978-3- [191] Z. Hu, Y. Dong, K. Wang, and Y. Sun, “Heterogeneous Graph
319-93417-4_48. Transformer,” in Proceedings of The Web Conference 2020, New
[173] M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural York, NY, USA: Association for Computing Machinery, 2020, pp.
networks,” IEEE Transactions on Signal Processing, vol. 45, no. 11, 2704–2710. Accessed: Jul. 20, 2022. [Online]. Available:
pp. 2673–2681, 1997, doi: 10.1109/78.650093. https://doi.org/10.1145/3366423.3380027
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
[192] Z. Liu, C. Chen, X. Yang, J. Zhou, X. Li, and L. Song, Industry 4.0,” Jun. 10, 2022. https://www.researchsquare.com
“Heterogeneous Graph Neural Networks for Malicious Account (accessed Jul. 21, 2022).
Detection,” in Proceedings of the 27th ACM International [209] K. Lee, B. Eoff, and J. Caverlee, “Seven Months with the Devils: A
Conference on Information and Knowledge Management, New York, Long-Term Study of Content Polluters on Twitter,” Proceedings of
NY, USA, Oct. 2018, pp. 2077–2085. doi: the International AAAI Conference on Web and Social Media, vol. 5,
10.1145/3269206.3272010. no. 1, Art. no. 1, 2011.
[193] K. Roshan and A. Zafar, “Utilizing XAI technique to improve [210] H. Liu, C. Zhong, A. Alnusair, and S. R. Islam, “FAIXID: A
autoencoder based model for computer network anomaly detection Framework for Enhancing AI Explainability of Intrusion Detection
with shapley additive explanation(SHAP).” Dec. 14, 2021. doi: Results Using Data Cleaning Techniques,” J Netw Syst Manage, vol.
10.5121/ijcnc.2021.13607. 29, no. 4, p. 40, May 2021, doi: 10.1007/s10922-021-09606-8.
[194] Y. Zhu et al., “Modeling Users’ Behavior Sequences with [211] S. Mane and D. Rao, “Explaining Network Intrusion Detection
Hierarchical Explainable Network for Cross-domain Fraud System Using Explainable AI Framework.” arXiv, Mar. 12, 2021.
Detection,” in Proceedings of The Web Conference 2020, New York, doi: 10.48550/arXiv.2103.07110.
NY, USA: Association for Computing Machinery, 2020, pp. 928– [212] B. Mahbooba, M. Timilsina, R. Sahal, and M. Serrano, “Explainable
938. Accessed: Jul. 20, 2022. [Online]. Available: Artificial Intelligence (XAI) to Enhance Trust Management in
https://doi.org/10.1145/3366423.3380172 Intrusion Detection Systems Using Decision Tree Model,”
[195] K. Yang and W. Xu, FraudMemory: Explainable Memory-Enhanced Complexity, vol. 2021, p. e6634811, Jan. 2021, doi:
Sequential Neural Networks for Financial Fraud Detection. 2019. 10.1155/2021/6634811.
Accessed: Jul. 20, 2022. [Online]. Available: [213] S. Wali and I. Khan, “Explainable AI and Random Forest Based
http://hdl.handle.net/10125/59542 Reliable Intrusion Detection system.” TechRxiv, Dec. 18, 2021. doi:
[196] Z. Xiao and J. Jiao, “Explainable Fraud Detection for Few Labeled 10.36227/techrxiv.17169080.v1.
Time Series Data,” Security and Communication Networks, vol. [214] M. Ghurab, G. Gaphari, F. Alshami, R. Alshamy, and S. Othman, “A
2021, p. e9941464, Jun. 2021, doi: 10.1155/2021/9941464. Detailed Analysis of Benchmark Datasets for Network Intrusion
[197] W. Min, W. Liang, H. Yin, Z. Wang, M. Li, and A. Lal, Detection System.” Rochester, NY, Apr. 14, 2021. Accessed: Jul. 22,
“Explainable Deep Behavioral Sequence Clustering for Transaction 2022. [Online]. Available: https://papers.ssrn.com/abstract=3834787
Fraud Detection.” arXiv, Jan. 11, 2021. doi: [215] T. Zebin, S. Rezvy, and Y. Luo, “An Explainable AI-Based Intrusion
10.48550/arXiv.2101.04285. Detection System for DNS Over HTTPS (DoH) Attacks,” IEEE
[198] A. Das, S. Baki, A. El Aassal, R. Verma, and A. Dunbar, “SoK: A Transactions on Information Forensics and Security, vol. 17, pp.
Comprehensive Reexamination of Phishing Research From the 2339–2349, 2022, doi: 10.1109/TIFS.2022.3183390.
Security Perspective,” IEEE Communications Surveys & Tutorials, [216] M. MontazeriShatoori, L. Davidson, G. Kaur, and A. Habibi
vol. 22, no. 1, pp. 671–708, 2020, doi: Lashkari, “Detection of DoH Tunnels using Time-series
10.1109/COMST.2019.2957750. Classification of Encrypted Traffic,” in 2020 IEEE Intl Conf on
[199] Y. Chai, Y. Zhou, W. Li, and Y. Jiang, “An Explainable Multi- Dependable, Autonomic and Secure Computing, Intl Conf on
Modal Hierarchical Attention Model for Developing Phishing Threat Pervasive Intelligence and Computing, Intl Conf on Cloud and Big
Intelligence,” IEEE Transactions on Dependable and Secure Data Computing, Intl Conf on Cyber Science and Technology
Computing, vol. 19, no. 2, pp. 790–803, Mar. 2022, doi: Congress (DASC/PiCom/CBDCom/CyberSciTech), 2020, pp. 63–70.
10.1109/TDSC.2021.3119323. doi: 10.1109/DASC-PICom-CBDCom-
[200] P. R. Galego Hernandes, C. P. Floret, K. F. Cardozo De Almeida, V. CyberSciTech49142.2020.00026.
C. Da Silva, J. P. Papa, and K. A. Pontara Da Costa, “Phishing [217] Y. Li, K. Xiong, T. Chin, and C. Hu, “A Machine Learning
Detection Using URL-based XAI Techniques,” in 2021 IEEE Framework for Domain Generation Algorithm-Based Malware
Symposium Series on Computational Intelligence (SSCI), 2021, pp. Detection,” IEEE Access, vol. 7, pp. 32765–32782, 2019, doi:
01–06. doi: 10.1109/SSCI50451.2021.9659981. 10.1109/ACCESS.2019.2891588.
[201] O. K. Sahingoz, E. Buber, O. Demir, and B. Diri, “Machine learning [218] F. Becker, A. Drichel, C. Müller, and T. Ertl, “Interpretable
based phishing detection from URLs,” Expert Systems with Visualizations of Deep Neural Networks for Domain Generation
Applications, vol. 117, pp. 345–357, Mar. 2019, doi: Algorithm Detection,” in 2020 IEEE Symposium on Visualization for
10.1016/j.eswa.2018.09.029. Cyber Security (VizSec), 2020, pp. 25–29. doi:
[202] Y. Lin et al., “Phishpedia: A Hybrid Deep Learning Based Approach 10.1109/VizSec51108.2020.00010.
to Visually Identify Phishing Webpages,” 2021, pp. 3793–3810. [219] A. Drichel, N. Faerber, and U. Meyer, “First Step Towards
Accessed: Jul. 21, 2022. [Online]. Available: EXPLAINable DGA Multiclass Classification,” in The 16th
https://www.usenix.org/conference/usenixsecurity21/presentation/lin International Conference on Availability, Reliability and Security,
[203] R. Valecha, P. Mandaokar, and H. R. Rao, “Phishing Email New York, NY, USA, Aug. 2021, pp. 1–13. doi:
Detection Using Persuasion Cues,” IEEE Transactions on 10.1145/3465481.3465749.
Dependable and Secure Computing, vol. 19, no. 2, pp. 747–756, Mar. [220] D. Plohmann, K. Yakdan, M. Klatt, J. Bader, and E. Gerhards-
2022, doi: 10.1109/TDSC.2021.3118931. Padilla, “A Comprehensive Measurement Study of Domain
[204] P. Barnard, N. Marchetti, and L. A. D. Silva, “Robust Network Generating Malware,” 2016, pp. 263–278. Accessed: Jul. 23, 2022.
Intrusion Detection through Explainable Artificial Intelligence [Online]. Available:
(XAI),” IEEE Networking Letters, pp. 1–1, 2022, doi: https://www.usenix.org/conference/usenixsecurity16/technical-
10.1109/LNET.2022.3186589. sessions/presentation/plohmann
[205] G. Andresini, A. Appice, F. P. Caforio, D. Malerba, and G. Vessio, [221] “Home - eduroam.org,” eduroam.org - eduroam global site.
“ROULETTE: A neural attention multi-output model for explainable https://eduroam.org/ (accessed Jul. 23, 2022).
Network Intrusion Detection,” Expert Systems with Applications, vol. [222] A. Drichel, U. Meyer, S. Schüppen, and D. Teubert, “Analyzing the
201, p. 117144, Sep. 2022, doi: 10.1016/j.eswa.2022.117144. real-world applicability of DGA classifiers,” in Proceedings of the
[206] Z. A. E. Houda, B. Brik, and L. Khoukhi, “‘Why Should I Trust 15th International Conference on Availability, Reliability and
Your IDS?’: An Explainable Deep Learning Framework for Security, New York, NY, USA, Aug. 2020, pp. 1–11. doi:
Intrusion Detection Systems in Internet of Things Networks,” IEEE 10.1145/3407023.3407030.
Open Journal of the Communications Society, pp. 1–1, 2022, doi: [223] R. H. Jhaveri, S. J. Patel, and D. C. Jinwala, “DoS Attacks in Mobile
10.1109/OJCOMS.2022.3188750. Ad Hoc Networks: A Survey,” in 2012 Second International
[207] “Network Intrusion Detection Based on Explainable Artificial Conference on Advanced Computing & Communication
Intelligence,” Jun. 16, 2022. https://www.researchsquare.com Technologies, 2012, pp. 535–541. doi: 10.1109/ACCT.2012.48.
(accessed Jul. 21, 2022). [224] S. Aziz et al., “Anomaly Detection in the Internet of Vehicular
[208] “KHO-XAI: Krill herd optimization and Explainable Artificial Networks Using Explainable Neural Networks (xNN),” Mathematics,
Intelligence framework for Network Intrusion Detection Systems in vol. 10, no. 8, Art. no. 8, Jan. 2022, doi: 10.3390/math10081267.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
[225] B. Hsupeng, K.-W. Lee, T.-E. Wei, and S.-H. Wang, “Explainable [243] C. M. Viana, M. Santos, D. Freire, P. Abrantes, and J. Rocha,
Malware Detection Using Predefined Network Flow,” in 2022 24th “Evaluation of the factors explaining the use of agricultural land: A
International Conference on Advanced Communication Technology machine learning and model-agnostic approach,” Ecological
(ICACT), 2022, pp. 27–33. doi: Indicators, vol. 131, p. 108200, Nov. 2021, doi:
10.23919/ICACT53585.2022.9728897. 10.1016/j.ecolind.2021.108200.
[226] R. R. Prasad, R. R. Rejimol Robinson, C. Thomas, and N. [244] J. Daníelsson, R. Macrae, and A. Uthemann, “Artificial intelligence
Balakrishnan, “Evaluation of Strategic Decision taken by and systemic risk,” Journal of Banking & Finance, vol. 140, p.
Autonomous Agent using Explainable AI,” in 2021 4th International 106290, Jul. 2022, doi: 10.1016/j.jbankfin.2021.106290.
Conference on Security and Privacy (ISEA-ISAP), 2021, pp. 1–8. doi: [245] D. V. Kute, B. Pradhan, N. Shukla, and A. Alamri, “Deep Learning
10.1109/ISEA-ISAP54304.2021.9689715. and Explainable Artificial Intelligence Techniques Applied for
[227] “KDD Cup 1999 Data.” Detecting Money Laundering–A Critical Review,” IEEE Access, vol.
http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html (accessed 9, pp. 82300–82317, 2021, doi: 10.1109/ACCESS.2021.3086230.
Jul. 23, 2022). [246] S. Sachan, J.-B. Yang, D.-L. Xu, D. E. Benavides, and Y. Li, “An
[228] K. Amarasinghe, K. Kenney, and M. Manic, “Toward Explainable explainable AI decision-support-system to automate loan
Deep Neural Network Based Anomaly Detection,” in 2018 11th underwriting,” Expert Systems with Applications, vol. 144, p.
International Conference on Human System Interaction (HSI), Jul. 113100, Apr. 2020, doi: 10.1016/j.eswa.2019.113100.
2018, pp. 311–317. doi: 10.1109/HSI.2018.8430788. [247]L. Yang, E. M. Kenny, T. L. J. Ng, Y. Yang, B. Smyth, and R. Dong,
[229]M. Javaid and A. Haleem, “Industry 4.0 applications in medical field: “Generating Plausible Counterfactual Explanations for Deep
A brief review,” Current Medicine Research and Practice, vol. 9, no. Transformers in Financial Text Classification.” arXiv, Oct. 23, 2020.
3, pp. 102–109, May 2019, doi: 10.1016/j.cmrp.2019.04.001. doi: 10.48550/arXiv.2010.12512.
[230] L. Coventry and D. Branley, “Cybersecurity in healthcare: A [248] A. Hanif, “Towards Explainable Artificial Intelligence in Banking
narrative review of trends, threats and ways forward,” Maturitas, vol. and Financial Services.” arXiv, Dec. 14, 2021. doi:
113, pp. 48–52, Jul. 2018, doi: 10.1016/j.maturitas.2018.04.008. 10.48550/arXiv.2112.08441.
[231] D. Dave, H. Naik, S. Singhal, and P. Patel, “Explainable AI meets [249] F. Gurcan, N. E. Cagiltay, and K. Cagiltay, “Mapping Human–
Healthcare: A Study on Heart Disease Dataset.” arXiv, Nov. 06, Computer Interaction Research Themes and Trends from Its
2020. doi: 10.48550/arXiv.2011.03195. Existence to Today: A Topic Modeling-Based Review of past 60
[232] X. Li et al., “BrainGNN: Interpretable Brain Graph Neural Network Years,” International Journal of Human–Computer Interaction, vol.
for fMRI Analysis,” Medical Image Analysis, vol. 74, p. 102233, 37, no. 3, pp. 267–280, Feb. 2021, doi:
Dec. 2021, doi: 10.1016/j.media.2021.102233. 10.1080/10447318.2020.1819668.
[233] “Ada-WHIPS: explaining AdaBoost classification with applications [250] “Toward human-centered AI: a perspective from human-computer
in the health sciences | BMC Medical Informatics and Decision interaction: Interactions: Vol 26, No 4.”
Making | Full Text.” https://dl.acm.org/doi/fullHtml/10.1145/3328485 (accessed Jul. 26,
https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/ 2022).
s12911-020-01201-2 (accessed Jul. 25, 2022). [251] G. Loveleen, B. Mohan, B. S. Shikhar, J. Nz, M. Shorfuzzaman, and
[234] D. R. Chittajallu et al., “XAI-CBIR: Explainable AI System for M. Masud, “Explanation-driven HCI Model to Examine the Mini-
Content based Retrieval of Video Frames from Minimally Invasive Mental State for Alzheimer’s Disease,” ACM Trans. Multimedia
Surgery Videos,” in 2019 IEEE 16th International Symposium on Comput. Commun. Appl., Mar. 2022, doi: 10.1145/3527174.
Biomedical Imaging (ISBI 2019), Apr. 2019, pp. 66–69. doi: [252] V. Dominguez, I. Donoso-Guzmán, P. Messina, and D. Parra,
10.1109/ISBI.2019.8759428. “Algorithmic and HCI Aspects for Explaining Recommendations of
[235] K. Zhang, J. Ni, K. Yang, X. Liang, J. Ren, and X. S. Shen, Artistic Images,” ACM Trans. Interact. Intell. Syst., vol. 10, no. 4, p.
“Security and Privacy in Smart City Applications: Challenges and 30:1-30:31, Nov. 2020, doi: 10.1145/3369396.
Solutions,” IEEE Communications Magazine, vol. 55, no. 1, pp. [253] Q. V. Liao and K. R. Varshney, “Human-Centered Explainable AI
122–129, Jan. 2017, doi: 10.1109/MCOM.2017.1600267CM. (XAI): From Algorithms to User Experiences.” arXiv, Apr. 19, 2022.
[236] M. Zolanvari, Z. Yang, K. Khan, R. Jain, and N. Meskin, “TRUST Accessed: Jul. 26, 2022. [Online]. Available:
XAI: Model-Agnostic Explanations for AI With a Case Study on http://arxiv.org/abs/2110.10790
IIoT Security,” IEEE Internet of Things Journal, pp. 1–1, 2021, doi: [254] K. B. Kelarestaghi, K. Heaslip, V. Fessmann, M. Khalilikhah, and A.
10.1109/JIOT.2021.3122019. Fuentes, “Intelligent Transportation System Security: Hacked
[237] R. Stirnberg et al., “Meteorology-driven variability of air pollution Message Signs,” SAE International Journal of Transportation
(PM1) revealed with explainable machine learning,” Atmospheric Cybersecurity and Privacy, vol. 1, Jun. 2018, doi: 10.4271/11-01-02-
Chemistry and Physics, vol. 21, no. 5, pp. 3919–3948, Mar. 2021, 0004.
doi: 10.5194/acp-21-3919-2021. [255] N. Soni, R. Malekian, and A. Thakur, “Edge Computing in
[238] M. Haeffelin et al., “SIRTA, a ground-based atmospheric Transportation: Security Issues and Challenges.” arXiv, Dec. 21,
observatory for cloud and aerosol research,” Annales Geophysicae, 2020. doi: 10.48550/arXiv.2012.11206.
vol. 23, no. 2, pp. 253–275, Feb. 2005, doi: 10.5194/angeo-23-253- [256] H. Mankodiya, M. S. Obaidat, R. Gupta, and S. Tanwar, “XAI-AV:
2005. Explainable Artificial Intelligence for Trust Management in
[239] L. Monje, R. A. Carrasco, C. Rosado, and M. Sánchez-Montañés, Autonomous Vehicles,” in 2021 International Conference on
“Deep Learning XAI for Bus Passenger Forecasting: A Use Case in Communications, Computing, Cybersecurity, and Informatics
Spain,” Mathematics, vol. 10, no. 9, Art. no. 9, Jan. 2022, doi: (CCCI), 2021, pp. 1–5. doi: 10.1109/CCCI52664.2021.9583190.
10.3390/math10091428. [257] “VeReMi dataset,” VeReMi-dataset.github.io. https://veremi-
[240] G. Kostopoulos, T. Panagiotakopoulos, S. Kotsiantis, C. Pierrakeas, dataset.github.io/ (accessed Jul. 26, 2022).
and A. Kameas, “Interpretable Models for Early Prediction of [258] S. Shams Amiri, S. Mottahedi, E. R. Lee, and S. Hoque, “Peeking
Certification in MOOCs: A Case Study on a MOOC for Smart City inside the black-box: Explainable machine learning applied to
Professionals,” IEEE Access, vol. 9, pp. 165881–165891, 2021, doi: household transportation energy consumption,” Computers,
10.1109/ACCESS.2021.3134787. Environment and Urban Systems, vol. 88, p. 101647, Jul. 2021, doi:
[241] Y. Feng, D. Wang, Y. Yin, Z. Li, and Z. Hu, “An XGBoost-based 10.1016/j.compenvurbsys.2021.101647.
casualty prediction method for terrorist attacks,” Complex Intell. [259] C. Bustos et al., “Explainable, automated urban interventions to
Syst., vol. 6, no. 3, pp. 721–740, Oct. 2020, doi: 10.1007/s40747- improve pedestrian and vehicle safety,” Transportation Research
020-00173-0. Part C: Emerging Technologies, vol. 125, p. 103018, Apr. 2021, doi:
[242] M. C. Garrido, J. M. Cadenas, A. Bueno-Crespo, R. Martínez- 10.1016/j.trc.2021.103018.
España, J. G. Giménez, and J. M. Cecilia, “Evaporation Forecasting [260] A. Kuppa and N.-A. Le-Khac, “Black Box Attacks on Explainable
through Interpretable Data Analysis Techniques,” Electronics, vol. Artificial Intelligence(XAI) methods in Cyber Security,” in 2020
11, no. 4, Art. no. 4, Jan. 2022, doi: 10.3390/electronics11040536.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2022.3204051
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/