ENISA Report - Securing Machine Learning Algorithms
ENISA Report - Securing Machine Learning Algorithms
LEARNING ALGORITHMS
DECEMBER 2021
0
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ABOUT ENISA
The European Union Agency for Cybersecurity, ENISA, is the Union’s agency dedicated to
achieving a high common level of cybersecurity across Europe. Established in 2004 and
strengthened by the EU Cybersecurity Act, the European Union Agency for Cybersecurity
contributes to EU cyber policy, enhances the trustworthiness of ICT products, services and
processes with cybersecurity certification schemes, cooperates with Member States and EU
bodies, and helps Europe prepare for the cyber challenges of tomorrow. Through knowledge
sharing, capacity building and awareness raising, the Agency works together with its key
stakeholders to strengthen trust in the connected economy, to boost resilience of the Union’s
infrastructure, and, ultimately, to keep Europe’s society and citizens digitally secure. More
information about ENISA and its work can be found here: www.enisa.europa.eu.
CONTACT
For contacting the authors please use [email protected]
For media enquiries about this paper, please use [email protected]
EDITORS
Apostolos Malatras, Ioannis Agrafiotis, Monika Adamczyk, ENISA
ACKNOWLEDGEMENTS
We would like to thank the Members and Observers of the ENISA ad hoc Working Group on
Artificial Intelligence for their valuable input and feedback.
LEGAL NOTICE
Notice must be taken that this publication represents the views and interpretations of ENISA,
unless stated otherwise. This publication should not be construed to be a legal action of ENISA
or the ENISA bodies unless adopted pursuant to the Regulation (EU) No 2019/881.
This publication does not necessarily represent state-of the-art and ENISA may update it from
time to time.
Third-party sources are quoted as appropriate. ENISA is not responsible for the content of the
external sources including external websites referenced in this publication.
This publication is intended for information purposes only. It must be accessible free of charge.
Neither ENISA nor any person acting on its behalf is responsible for the use that might be made
of the information contained in this publication.
COPYRIGHT NOTICE
© European Union Agency for Cybersecurity (ENISA), 2021
Reproduction is authorised provided the source is acknowledged.
1
SECURING MACHINE LEARNING ALGORITHMS
December 2021
TABLE OF CONTENTS
EXECUTIVE SUMMARY 3
1. INTRODUCTION 4
1.1 OBJECTIVES 4
1.2 METHODOLOGY 4
1.4 STRUCTURE 6
4. SECURITY CONTROLS 18
4.1 SECURITY CONTROLS RESULTS 18
5. CONCLUSION 26
D ANNEX: REFERENCES 43
2
SECURING MACHINE LEARNING ALGORITHMS
December 2021
EXECUTIVE SUMMARY
The vast developments in digital technology influence every aspect of our daily lives. Emerging
technologies, such as Artificial Intelligence (AI), which are in the epicentre of the digital
evolution, have accelerated the digital transformation contributing in social and economic
prosperity. However, the application of emerging technologies and AI in particular, entails perils
that need to be addressed if we are to ensure a secure and trustworthy environment. In this
report, we focus on the most essential element of an AI system, which are machine learning
algorithms. We review related technological developments and security practices to identify
emerging threats, highlight gaps in security controls and recommend pathways to enhance
cybersecurity posture in machine learning systems.
Based on a systematic review of relevant literature on machine learning, we provide a taxonomy This report
for machine learning algorithms, highlighting core functionalities and critical stages. The provides a
taxonomy sheds light on main data types used by algorithms, the type of training these
taxonomy for
algorithms entail (supervised, unsupervised) and how output is shared with users. Particular
emphasis is given to the explainability and accuracy of these algorithms. Next, the report
machine learning
presents a detailed analysis of threats targeting machine learning systems. Identified threats algorithms, a
include inter alia, data poisoning, adversarial attacks and data exfiltration. All threats are detailed analysis
associated to particular functionalities of the taxonomy that they exploit, through detailed tables. of threats and
Finally, we examine mainstream security controls described in widely adopted standards, such
security controls
as ISO 27001 and NIST Cybersecurity framework, to understand how these controls can
effectively detect, deter and mitigate harms from the identified threats. To perform our analysis,
in widely adopted
we map all the controls to the core functionalities of machine learning systems that they protect standards
and to the vulnerabilities that threats exploit in these systems.
Our analysis indicates that the conventional security controls, albeit very effective for
information systems, need to be complemented by security controls tailored to machine learning
functionalities. To identify these machine-learning controls, we conduct a systematic review of
relevant literature, where academia and research institutes propose ways to avoid and mitigate
threats targeting machine learning algorithms. Our report provides an extensive list of security
controls that are applicable only for machine learning systems, such as “include adversarial
examples to training datasets”. For all controls, we map the core functionality of machine
learning algorithms that they intend to protect to the vulnerabilities that threats exploit.
Our findings indicate that there is no unique strategy in applying a specific set of security
controls to protect machine learning algorithms. The overall cybersecurity posture of
organisations who use machine learning algorithms can be enhanced by carefully choosing
controls designed for these algorithms. As these controls are not validated in depth, nor
standardised in how they should be implemented, further research should focus on creating
benchmarks for their effectiveness. We further identified cases where the deployment of
security controls may lead to trade-offs between security and performance. Therefore, the
context in which controls are applied is crucial and next steps should focus on considering
specific use cases and conducting targeted risk assessments to better understand these trade-
offs. Finally, given the complexity of securing machine learning systems, governments and
related institutions have new responsibilities in raising awareness regarding the impact of
threats on machine learning. It is important to educate data scientists on the perils of threats
and on the design of security controls before machine learning algorithms are used in
organisations’ environments. By engaging experts in machine learning in cybersecurity issues,
we may create the opportunity to design innovative security solutions and mitigate the emerging
threats on machine learning systems.
3
SECURING MACHINE LEARNING ALGORITHMS
December 2021
1. INTRODUCTION
Artificial Intelligence (AI) has grown significantly in recent years and driven by computational
advancements has found wide applicability. By providing new opportunities to solve decision-
making problems intelligently and automatically, AI is being applied to more and more use
cases in a growing number of sectors. The benefits of AI are significant and undeniable.
However, the development of AI is also accompanied by new threats and challenges, which
relevant professionals will have to face.
In 2020, ENISA published a threat landscape report on AI1. This report, published with the
support of the Ad-Hoc Working Group on Artificial Intelligence Cybersecurity2, presents the
Agency’s active mapping of the AI cybersecurity ecosystem and its threat landscape. This threat
landscape not only lays the foundation for upcoming cybersecurity policy initiatives and
technical guidelines, but also stresses relevant challenges.
Machine learning (ML), which can be defined as the ability for machines to learn from data to
solve a task without being explicitly programmed to do so, is currently the most developed and
promising subfield of AI for industrial and government infrastructures. It is also the most
commonly used subfield of AI in our daily lives.
ML algorithms and their specificities, such as the fact that they need large amount of data to
learn, make them the subject of very specific cyber threats that project teams must consider.
The aim of this study is to help project teams identify the specific threats that can target ML
algorithms, associated vulnerabilities, and security controls for addressing these vulnerabilities.
Building on the ENISA AI threat landscape mapping, this study focuses on cybersecurity threats
specific to ML algorithms. Furthermore, vulnerabilities related to the aforementioned threats and
importantly security controls and mitigation measures are proposed.
The adopted description of AI is a deliberate simplification of the state of the art regarding that
vast and complex discipline with the intent of not precisely or comprehensively define it but
rather pragmatically contextualise the specific technique of machine learning.
1.1 OBJECTIVES
The objectives of this publication are:
1.2 METHODOLOGY
To produce this report, the work was divided into three stages. At the core of the methodology
was an extensive literature review (full list of references may be found in Annex D). The aim
1
https://www.enisa.europa.eu/publications/artificial-intelligence-cybersecurity-challenges
2
See https://www.enisa.europa.eu/topics/iot-and-smart-infrastructures/artificial_intelligence/ad-hoc-working-
group/adhoc_wg_calls
4
SECURING MACHINE LEARNING ALGORITHMS
December 2021
was to consult documents that are more specific to ML algorithms in general in order to build the
taxonomy, and to consult documents more specific to security to identify threats, vulnerabilities,
and security controls. At the end of the systematic review, more than 200 different documents
(of which a hundred are related to security) on various algorithms of ML had been collected and
analysed.
The second step was to identify the cybersecurity threats that could target ML algorithms
and potential vulnerabilities. For this task, the threat landscape from ENISA’s report on AI
cybersecurity challenges was the starting point, which was then enriched through desk research
with sources related to the security of ML algorithms. Additionally, the expertise of the ENISA
Ad-Hoc Working Group on Artificial Intelligence Cybersecurity was sought. This work allowed us
to select threats and identify associated vulnerabilities. Subsequently, they were linked to the
previously established ML taxonomy.
The last step of this work was the identification of the security controls addressing the
vulnerabilities. To do this, we utilised the desk research and enriched it with the most relevant
standard security controls from ISO 27001/2 and the NIST 800-53 framework. The output was
reviewed with the experts of the ENISA Ad-Hoc Working Group on Artificial Intelligence
Cybersecurity. This work allowed us to identify security controls that were then linked to the ML
taxonomy.
It is important to note that we opted to enrich the ML-targeted security controls with more
conventional ones to highlight that applications using ML must also comply with more classic
controls in order to be sufficiently protected. Considering measures that are specific to ML
would only give a partial picture of the security work needed on these applications.
3
https://www.enisa.europa.eu/publications/artificial-intelligence-cybersecurity-challenges
5
SECURING MACHINE LEARNING ALGORITHMS
December 2021
1.4 STRUCTURE
The report aims to help the target audience to identify the cyber threats to consider and the
security controls to deploy in order to secure their ML applications. Accordingly, the report is
structure into three sections:
This report focuses on threats that target ML algorithms and on the associated security controls.
It is important to note that this publication examines security controls that are specific to ML
algorithms as well as standard security controls that are also applicable to ML algorithms and
systems making use of them. To use this publication effectively, it is important to note that:
• As is the case for any application, when using ML, one must also consider traditional
security standards (e.g. ISO 27001/2, NIST 800-53), because ML applications are
subject not only to AI/ML specific threats but also to general nature cybersecurity
threats.
• The context of the application (e.g. manipulated data, business case, deployment)
must be considered to correctly assess the risks and prioritise deployment of the
security controls accordingly.
6
SECURING MACHINE LEARNING ALGORITHMS
December 2021
2. MACHINE LEARNING
ALGORITHMS TAXONOMY
One of the objectives of this work was to devise a (non-exhaustive) taxonomy, to support the
process of identifying which specific threats can target ML algorithms, their associated
vulnerabilities, and security controls for addressing these vulnerabilities. An important disclaimer
needs to be made concerning this taxonomy, namely that it is not meant to be complete or
exhaustive when it comes to ML, instead it aims to support the security analysis of ML
algorithms in this report.
Based on the desk research and interviews with experts of the ENISA AI Working group, we
identified 40 of the most commonly used ML algorithms. A taxonomy was built based on the
analysis of these algorithms. In particular, it was noted that ML algorithms were driven mainly by
the learning paradigms and the problem they address (main domain). These aspects were
therefore chosen to form the key taxonomy dimensions, as seen in Figure 1. It should be noted
that Annex A provides a complete listing of the 40 algorithms and their mapping to the features
of the taxonomy, whereas the Figure serves for illustration purposes.
7
SECURING MACHINE LEARNING ALGORITHMS
December 2021
There is a strong correlation between the domain of application (the problem being addressed)
and the data type which is being worked on, as well as between data environments and learning
paradigm. Thus, further dimensions of the taxonomy were introduced accordingly.
The inputs that are given to a ML algorithm are data and therefore, the algorithms can be
categorised based on the types of data that is fed into them. In most cases, specific types of
data are used in certain domains of ML. Indeed, all the algorithms used in computer vision are
fed with images and videos, in the same way that all algorithms used in Natural Language
Processing are fed with text4. In Table 1, the main domains and the type of data used in each of
them are listed.
Main
Data type Definition
domain
Data organised in a predefined model of array with one specific column for each
feature (e.g. textual, numerical data, date). To be more accurate, structured data
refer to organised data that can be found in a relational data base for example
Classic (that may contain textual columns as mentioned).
Data Structured Data
Science Quantitative data can be distinguished from qualitative data. Quantitative data
corresponds to the numerical data that can supports some arithmetic operations
whereas qualitative data is usually used as categorical data to classify data
according to their similarities.
Certain domains such as NLP and Computer Vision have been separated from Classic Data
Science. The purpose of this separation was to make a distinction between algorithms that may
be used specifically or predominantly for each domain.
4
Audio data are also used for speech recognition. For the purposes of this report, we consider only text for the NLP for the
taxonomy. considering that this will not create differences for the work on threats.
5
For the purposes of this report, time series belong to the two main domains: Classic Data Science and Speech
processing. By restraining Time series to Classic Data Science and Speech processing, we aspired to emphasise the
specific approaches that are used for this domain like ARIMA and Hidden Markov Model. Furthermore, we include audio
data under time series and made the choice to separate video from time series.
8
SECURING MACHINE LEARNING ALGORITHMS
December 2021
In addition to the data types fed into the algorithms, we also focused on three learning
paradigms, namely supervised learning, unsupervised learning, and reinforcement learning:
Each of these learning paradigms have different security-related properties which may lead to
attacks and therefore, it is relevant to represent this information in the taxonomy of ML algorithms,
from which security controls will be mapped. For instance, the most common learning paradigm
is classification and thus, it has many more examples of vulnerabilities due to its popularity.
9
SECURING MACHINE LEARNING ALGORITHMS
December 2021
However, some of the widely used and mentioned algorithms are based on common elementary
components, or are extensions of the same principle, and can therefore form families or clusters
of algorithms on this taxonomy grid. Hence, we map those specific algorithms in groups by
using nested boxes, as it allows for the representation of a wide variety of algorithms, while
showing that some have relationships with one another.
To continue with the previous example, a more recent version of RNN is LSTM 7 (Long-Short
Term Memory), which differs from RNN based on its optimisation techniques, making it faster to
learn and more precise. Since LSTM is a specific extension of RNN, the LSTM box was nested
in the RNN box in the taxonomy: this indicates that the two algorithms are part of the same
family.
It is important to note that we focused on the algorithms’ explainability because this work is
important for other parts of the publication. For example, in one identified security control, it is
highlighted that it is necessary to ensure that ML projects comply with regulatory constraints
such as the GDPR, which describes some explainability requirements8.
6
https://apps.dtic.mil/dtic/tr/fulltext/u2/a164453.pdf
7
https://www.bioinf.jku.at/publications/older/2604.pdf
8
GDPR Recital 71 “The data subject should have the right not to be subject to a decision, which may include a measure,
evaluating personal aspects relating to him or her which is based solely on automated processing and which produces legal
effects concerning him or her or similarly significantly affects him or her, such as automatic refusal of an online credit
application or e-recruiting practices without any human intervention. […] In any case, such processing should be subject to
1
Please use footnotes for providing additional or explanatory information and/or relevant links. References should be listed
in a dedicated section. Use only the function References/Insert Footnote
10
SECURING MACHINE LEARNING ALGORITHMS
December 2021
suitable safeguards, which should include specific information to the data subject and the right to obtain human
intervention, to express his or her point of view, to obtain an explanation of the decision reached after such assessment and
to challenge the decision.”
9
https://www.enisa.europa.eu/publications/artificial-intelligence-cybersecurity-challenges
1
Please use footnotes for providing additional or explanatory information and/or relevant links. References should be listed
in a dedicated section. Use only the function References/Insert Footnote
11
SECURING MACHINE LEARNING ALGORITHMS
December 2021
The aim of the ML algorithm taxonomy is to focus not only on the functionalities of the
algorithms but also on the ML models’ workflow represented by the lifecycle. This lifecycle
summarises the principle steps to produce an ML model. It is important to note that several
steps could have been added, such as data creation and data analysis (for instance, to analyse
if there are some personal data or biases). However, to simplify the lifecycle, some steps have
been condensed. Thus, for example, data cleaning has been included. Regarding data creation,
it was considered as being external to the ML lifecycle.
10
Optimisation is also known as model tuning.
11
Data cleaning and data processing have been separated to distinguish the cleaning phase from the adaptation phase of
the dataset for learning (dimension reduction, feature engineering, etc.).
12
SECURING MACHINE LEARNING ALGORITHMS
December 2021
3. ML THREATS AND
VULNERABILITIES
13
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Model Training
Preprocessing
Model Testing
Data Cleaning
Threats | sub-
Model design
Optimisation
Definition
Deployment
threats
Monitoring
Evaluation
Collection
Model
Model
Data
Data
A type of attack in which the attacker works on
the ML algorithm's inputs to find small
perturbations leading to large modification of its
outputs (e.g. decision errors). It is as if the
attacker created an optical illusion for the
Evasion algorithm. Such modified inputs are often called
x
adversarial examples.
Example: the projection of images on a house
could lead the algorithm of an autonomous car to
take the decision to suddenly make it brake.
12
We have chosen to separate the oracle attacks from this threat to address the specifics of both threats and give them both a fair
representation. However, Oracle-type attacks may be considered as a ML specific sub-threat of model or data disclosure.
14
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Model Training
Preprocessing
Model Testing
Data Cleaning
Threats | sub-
Model design
Optimisation
Definition
Deployment
threats
Monitoring
Evaluation
Collection
Model
Model
Data
Data
This threat refers to a leak of data manipulated
by ML algorithms. This data leakage can be
explained by an inadequate access control, a
handling error of the project team or simply
because sometimes the entity that owns the
Data disclosure x x x x x x x x x x
model and the entity that owns the data are
distinct. To train the model, it is often necessary
for the data to be accessed by the model
provider. This involve sharing the data and thus
share sensitive data with a third party.
Cybersecurity
This threat refers to the possibility that a project
incident not
team may not report security incidents to
reported to x x x x x x x x x x
dedicated teams while a policy of mandatory
incident response
incident reporting has been defined.
teams
15
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Use of adversarial examples crafted in Too much information available on the model
white or grey box conditions (e.g.
FGSM…) Too much information about the model given in its outputs
16
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Lack of security process to maintain a good security level of the components of the ML
application
Compromise of ML application Undefined indicators of proper functioning, making complex compromise identification
components
Bad practices due to a lack of cybersecurity awareness
Lack of security process to maintain a good security level of the components of the ML
application
Weak access protection mechanisms for ML model components
Existence of several vulnerabilities because ML specificities are not integrated to existing
policies
Existence of several vulnerabilities because ML application do not comply with security
policies
Contract with a low security third party
17
SECURING MACHINE LEARNING ALGORITHMS
December 2021
4. SECURITY CONTROLS
• “Organisational and Policy” are more traditional security controls, either organisational
or linked to security policies.
• “Technical” are more classic technical security controls.
• “Specific to ML” are security controls that are specific to applications using ML.
In Annex 5.C, a set of operational implementation examples are listed for each of the security
controls. This includes:
• For security controls not specific to ML algorithms: examples from the ISO 27001/2 13
family of standards or NIST 800-53 14 framework that should be considered when
implementing the security control.
• For security controls specific to ML: examples of techniques found in the current
literature. All sources are referenced and may be found in Annex 5.D.
The overall mapping of threats, vulnerabilities and security controls is available in Annex 5.B.
13
https://www.iso.org/isoiec-27001-information-security.html
14
https://csrc.nist.gov/publications/detail/sp/800-53/rev-5/final
18
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
ORGANISATIONAL
19
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
As all applications, those using ML must
Ensure ML applications
comply with protection policies (e.g.
comply with protection
hardening, anti-malware policy) and be x x x x x x x x x x
policies and are integrated to
integrated to security operations processes
security operations processes
(e.g. vulnerability management, backups).
TECHNICAL
15
Please note that ML components with false positives might have adverse effect.
20
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
During the lifecycle of an ML algorithm,
several components (such as software,
programming libraries or even other models)
are used to complete the project. Security
Check the vulnerabilities of checks have to be carried out to ensure that
the components used so that these components offer an adequate level of
x x x x x x x x x x
they have an appropriate security. Moreover, some mechanisms need
security level to be used to prevent tampering with the
components used.
For example: if an open-source library is to
be used, code reviews or check for public
vulnerabilities on it can be done.
A risk analysis of the overall application
should be conducted to take into account the
specificities of its context, including:
- The attacker’s motivations
- The sensitivity of the data handled (e.g.
medical or personal and thus subject to
regulatory constraints, strategic for the
Conduct a risk analysis of the company and should thus be highly
x x x x x x x x x x
ML application protected)
- The application hosting (e.g. through third
parties services, cloud or on premise
environments)
- The model architecture (e.g. its exposition,
learning methods)
- The ML application lifecycle (e.g., model
sharing
21
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
Define dashboards of key indicators
integrating security indicators (peaks of
Define and monitor indicators
change in model behavior etc.) to follow-up
for proper functioning of the x
the proper functioning of the model
model
regarding the business case, in particular to
allow rapid identification of anomalies.
Test environments must also be secured
according to the sensitivity of the information
Ensure appropriate protection they contain. Special care must be paid to
is deployed for test the data used in these environments, to x x x x x x x x x x
environments ensure their protection (e.g., same
protection measures as for production if not
desensitiser).
As all applications, those using ML must
Ensure ML applications
comply with third parties’ security
comply with third parties’ x x x x x x x x x x
requirements if their context involves
security requirements
suppliers.
As any project, ML projects must comply to
process for integrating security into projects,
including the followings:
- Risk analysis on the whole application
- Check of the integration of cybersecurity
best practices regarding architecture, secure
development.
- Check that the application will be
Ensure ML projects follow the integrated in existing operational security
global process for integrating processes: monitoring and response, patch x x x x x x x x x x
security into projects management, access management, cyber-
resilience.
- Check of the production of adequate
documentation to ensure the sustainability of
the application (e.g., technical architecture,
hardening, exploitation, configuration and
installation documents)
- Security checks before going to production
(e.g. security audit, pen tests)
SPECIFIC ML
16
This security control is often referred to as “Robust adversarial training” in the literature.
17
One important thing to keep in mind is that such modifications should not overly impact model performance on benign
inputs.
22
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
The ML models should be explainable, even
if it means simplifying them, to enable a
good understanding of their functioning and
decision factors.
It can also be a regulatory requirement (e.g.
GDPR). However, once again, security
Build explainable models x x
interferes with the explainability property of
the model (easier-to-understand decisions
can be easier-to-build adversarial
examples). It is therefore a trade-off
between the need for explainability and
security.
Some model designs can be more robust
than others against attacks. For instance,
ensemble methods like bagging can mitigate
Choose and define a more
the impact of poisoning (during the training x
resilient model design
phase). Another example is defensive
distillation, which may allow deep neural
networks to better deal with evasion attacks.
Using a set of training data expansion
techniques (e.g. data augmentation)
addresses the lack of data and improves the
robustness of the model to poisoning attacks
by diluting their impact. It is notable,
Enlarge the training dataset however, that this security control more x x
specifically addresses poisoning attacks that
aim to reduce the performance of the model
than those that seek to establish a backdoor.
Moreover, one needs to ensure the reliability
of the sources used to augment the dataset.
The introduction of bias in ML algorithms will
not be detailed because it is not the topic of
the publication.
Ensure that models are However, some techniques can be used to
x x x x x x
unbiased mitigate bias: verify the training dataset is
representative enough regarding the
business case, check the relevance of the
attributes used to make decisions etc.
23
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
ML is a rapidly evolving field, especially
regarding its cybersecurity. Regular
Implement processes to checking of new attacks and defenses must
maintain security levels of ML be integrated into the processes for x x x x x x x x x X
components over time maintaining security level applications. The
security level should thus be regularly
assessed too.
Input-based detection tools can be of
interest to identify whether a given input has
Implement tools to detect if a been modified by an attacker or not.
data point is an adversarial One example, in the case of Deep Neural x x x
example or not Networks (DNNs), is to add a neural
subnetwork to an architecture trained to
detect adversarial examples.
18
See https://arxiv.org/pdf/1902.06531.pdf. It is notable that STRIP (STRong Intentional Perturbatio) may have a huge
runtime overhead and may be infeasible for large dataset.
24
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Data Preprocessing
Model Deployment
Model design and
Model Evaluation
Implementation
Data Collection
Security controls Definition
Model Training
Model Testing
Data Cleaning
Optimisation
Monitoring
Controlling the information (like its verbosity)
provided by the model by applying basic
cybersecurity hygiene rules is a way of
limiting the techniques that an attacker can
use to build adversarial examples.
One of the basic rules of hygiene, for
example, is to reduce the information of the
output determined by the model to the
Reduce the information given
maximum, or by profile making the request. x
by the model19
For example: considering a classification
application, it would consist of
communicating only the predicted class to
the users of solution, not the associated
probability. However, it remains notable that
in many cases, research has shown that
minimal information is sufficient to mount
attacks.
Federated learning is a set of training
techniques that trains a model on several
decentraliser servers containing local data
Use federated learning to samples, without exchanging their data
x x
minimize risk of data breaches samples. This avoids the need to transfer
the data and/or entrust it to an untrusted
third party and thus helps to preserve the
privacy of the data.
The transferability property can be used to
force adversarial examples from a
substitution model to evade another. The
ease of transferring an adversarial example
Use less easily transferable
from a model to another depends on the x
models20
family of algorithms. One possible defense is
thus to choose an algorithm family that is
less sensitive to the transferability of
adversarial examples.
19
It is important to keep in mind that, in case of attacks like evasion or oracle, this security control can help. However, in
some cases, it may possible to bypass the security control by using more queries.
20
Some evasion attacks are based on the following principle: train a model with data like the target model used and
generate adversarial examples from this model. Then, present these adversarial examples to the target model to perform
an evasion attack. Whether or not to transfer an adversarial example generated by one model to another depends on their
respective design as shown in the reference 215.
25
SECURING MACHINE LEARNING ALGORITHMS
December 2021
5. CONCLUSION
Machine Learning algorithms are at the core of modern AI systems and applications. However, There is no silver
they are faced with a series of threats and vulnerabilities. In this report we have identified bullet for ML-
multiple security controls that can be applied to ML applications to address the threats they
face. Some of the security controls are specific to ML algorithms, but others are standard
specific attacks,
technical and organisational cybersecurity controls to mitigate general attacks. It is important to but mitigation
apply both types of controls because AI systems, in addition to ML specific vulnerabilities, there measures can
exist also general type of vulnerabilities, which may also be exploited by adversaries. still raise the bar
for attackers.
Mitigation controls for ML-specific attacks outlined in the report should in general be deployed
during the entire lifecycle of the ML system. This includes measures for assuring the data
Thus, more
quality and protecting its integrity, making the ML algorithms more robust and controlling access attention should
to both the model and the data to ensure their privacy. The report also emphasizes the need for be given to
the explainability of decisions, and the importance of detecting bias that can be present or security controls
injected in a model by an attacker, which can then lead to unethical uses of AI.
to enable
An important point highlighted in the report is that the identified security measures can be
comparability
applied to all algorithms. Nevertheless, their operational implementations (see Annex C) may be and increase
specific to certain types of algorithms. For example, for the security control “Choose and define resilience.
a more resilient model design”, the defensive distillation implementation is specific to neural
networks. It is also notable that with the prevalence of research papers on supervised learning,
there are more examples of operational implementations for this type of algorithms.
This report addresses an emerging subject. Thus, it remains very important to keep an active
watch on threats and security controls in the field of ML in order to understand the latest
innovations both from a technical point of view, or with a view to comply with standards provided
by ISO, IEEE and ETSI21.
When looking ahead and given the complexity of the issue of securing ML, companies and
governments have new responsibilities. For instance, it is increasingly important to raise
cybersecurity awareness within companies, especially regarding the security of ML systems.
For some populations, particularly data science teams, cybersecurity has not been at the
forefront for many years. Moreover, by including data science actors in these actions, they are
also given the opportunity to think of innovative solutions to mitigate the various threats. Thus,
to this end, training and education programs should be organised regularly and the
vulnerabilities of ML should be demonstrated using concrete examples .
Finally, the context in which security controls are applied is crucial and specific use cases
should be considered when conducting targeted risk assessments. All mitigations used should
be proportional to the application-specific threat level and consider specific conditions of the
environment that may either favor or hamper attacks. Moreover, defenders should be aware of
the following points:
1) There is no silver bullet for mitigating ML-specific attacks. Some security controls may
be bypassed by adaptive attackers. However, applied mitigations can still raise the bar
for attackers.
21
https://www.etsi.org/committee/sai
26
SECURING MACHINE LEARNING ALGORITHMS
December 2021
2) ML-specific mitigation controls are not generally evaluated in a standardised way even
if it is a current and important issue to enable comparability. More research should be
devoted to standardised benchmarks for comparing ML-specific mitigations on a level
playing field. These benchmarks should also be enforced to ensure that the methods
used in practice are the ones that perform best.
27
SECURING MACHINE LEARNING ALGORITHMS
December 2021
A ANNEX: TAXONOMY OF
ALGORITHMS
28
SECURING MACHINE LEARNING ALGORITHMS
December 2021
DBSCAN - Density-Based
Spatial Clustering of
Applications with Noise is a
density-based clustering
non-parametric algorithm:
given a set of points in
some space, it groups 26,
Computer Unsupervised
DBSCAN together points that are Image Clustering 129,
Vision Learning
closely packed together 142
(points with many nearby
neighbours), marking as
outliers points that lie alone
in low-density regions
(whose nearest neighbours
are too far away).
A decision tree is a graph
that uses a branching
Classic
method to illustrate every Structured Supervised Classification, 40, 42,
Decision tree Data Fully Explainable
possible output for a specific data learning Regression 120,
Science
input in order to break down
complex problems.
Deep Q-learning works as
Q-learning algorithm at the
difference that it uses a
Classic
Deep Q- neural network to Reinforcement
Data Time series Rewarding Yes 65, 85
learning approximate the Q-value learning
Science
function to manage big
amount of states and
actions.
EfficientNet is a
Convolutional Neural
Network based on depth
wise convolutions, which
Computer Supervised
EfficientNet makes it lighter than other Image Classification Not Explainable Yes 4
Vision learning
CNNs. It also allows to
scale the model with a
unique lever: the compound
coefficient.
The factorial
correspondence analysis
(CFA) is a statistical method
of data analysis which
Factor allows the analysis and
prioritisation of the Classic
analysis of Structured Unsupervised Dimension
information contained in a Data
corresponde data Learning Reduction
rectangular table of data Science
nces and which is today
particularly used to study
the link between two
variables (qualitative or
categorical).
A GAN is a generative
model where two networks
are placed in competition.
The first model is the
generator, it generates a
sample (e.g. an image),
Computer Image, Unsupervised
GAN while its opponent, the 53, 135
Vision Video Learning
discriminator, tries to detect
whether a sample is real or
whether it is the result of the
generator. Both improve on
the performance of the
other.
29
SECURING MACHINE LEARNING ALGORITHMS
December 2021
30
SECURING MACHINE LEARNING ALGORITHMS
December 2021
31
SECURING MACHINE LEARNING ALGORITHMS
December 2021
32
SECURING MACHINE LEARNING ALGORITHMS
December 2021
33
SECURING MACHINE LEARNING ALGORITHMS
December 2021
B ANNEX: MAPPING
SECURITY CONTROLS
TO THREATS
34
SECURING MACHINE LEARNING ALGORITHMS
December 2021
35
SECURING MACHINE LEARNING ALGORITHMS
December 2021
36
SECURING MACHINE LEARNING ALGORITHMS
December 2021
37
SECURING MACHINE LEARNING ALGORITHMS
December 2021
C ANNEX: IMPLEMENTING
SECURITY CONTROLS
The NIST 800-53 and the ISO 27001/2 provides several 148
Apply documentation
points:
requirements to Artificial
- Define change management processes, integrating the ISO 27001/2
Intelligence projects
update of the documentation NIST 800-53
38
SECURING MACHINE LEARNING ALGORITHMS
December 2021
39
SECURING MACHINE LEARNING ALGORITHMS
December 2021
40
SECURING MACHINE LEARNING ALGORITHMS
December 2021
41
SECURING MACHINE LEARNING ALGORITHMS
December 2021
Integrate ML specificities to The NIST 800-53 and the ISO 27001/2 provides several
awareness strategy and ensure points: ISO 27001/2
all ML stakeholders are - Organise training sessions NIST 800-53
receiving it - Perform locally cyber risks reporting
Reduce the information given The literature provides the following technique:
89, 145
by the model - Gradient Masking
42
SECURING MACHINE LEARNING ALGORITHMS
December 2021
D ANNEX: REFERENCES
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Adversarial Machine https://www.morga
• Yevgeniy
Learning, Synthesis nclaypool.com/doi/
Vorobeychik
1 Lectures on Artificial abs/10.2200/S008 2018 X
• Murat
Intelligence and 61ED1V01Y20180
Kantarcioglu
Machine Learning 6AIM039
• Tom B. Brown
• Benjamin Mann
• Nick Ryder
• Melanie Subbiah
• Jared Kaplan
• Prafulla Dhariwal
• Arvind
Neelakantan
• Pranav Shyam
• Girish Sastry
• Amanda Askell
• Sandhini Agarwal
• Ariel Herbert-
Voss
Language Models are https://arxiv.org/pdf • Gretchen Krueger
6 2020 X X
Few-Shot Learners /2005.14165.pdf • Tom Henighan
• Rewon Child
• Aditya Ramesh
• Daniel M. Ziegler
• Jeffrey Wu
• Clemens Winter
• Christopher
Hesse
• Mark Chen
• Eric Sigler
• Mateusz Litwin
• Scott Gray
• Benjamin Chess
• Jack Clark
• Christopher
43
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Berner
• Sam McCandlish
• Alec Radford
• Ilya Sutskeve
• Dario Amodei
• Preetum
Nakkiran
Deep Double Descent: • Gal Kaplun
https://arxiv.org/pdf
7 Where Bigger Models 2019 • Yamini Bansal X X
/1912.02292.pdf
and More Data Hurt • Tristan Yang
• Boaz Barak
• Ilya Sutskever
• Kaiming He
Deep Residual
https://arxiv.org/pdf • Xiangyu Zhang
8 Learning for Image 2015 X X
/1512.03385.pdf • Shaoqing Re
Recognition
• Jian Sun
• Kazuki Osawa
• Siddharth
Swaroop
https://papers.nips.
• Anirudh Jain
Practical Deep cc/paper/2019/file/
• Runa
9 Learning with Bayesian b53477c2821c1bf0 2019 X X
Eschenhagen
Principles da5d40e57b870d3
• Richard E. Turner
5-Paper.pdf
• Rio Yokota
• Mohammad
Emtiyaz Khan
Deep Inside
Convolutional • Karen Simonyan
Networks: Visualising https://arxiv.org/pdf • Andrea Vedaldi
10 2014 X X
Image Classification /1312.6034.pdf • Andrew
Models and Saliency Zisserman
Maps
https://web.stanfor
d.edu/class/psych2
Reinforcement learning: • Richard S. Sutton
11 09/Readings/Sutto 1998 X X X X
An introduction • Andrew G. Barto
nBartoIPRLBook2n
dEd.pdf
• Yuri Burda
Exploration by Random https://arxiv.org/pdf • Harrison Edwards
12 2018 X X
Network Distillation /1810.12894.pdf • Amos Storkey
• Oleg Klimov
DeepFool: a simple
• Seyed-Mohsen
and accurate method to
https://arxiv.org/pdf • Moosavi-Dezfooli
13 fool deep neural 2016 X
/1511.04599.pdf • Alhussein Fawzi
networks, in arXiv, July
• Pascal Frossard
2016
• Yingce Xia
• Di He
• Tao Qin
Dual Learning for https://arxiv.org/pdf
14 2016 • Liwei Wang X X
Machine Translation /1611.00179.pdf
• Nenghai Yu
• Tie-Yan Liu
• Wei-Ying Ma
44
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Jingwei Xu
• Huazhe Xu
Video Prediction via https://arxiv.org/pdf
15 2020 • Bingbing Ni X X
Example Guidance /2007.01738.pdf
• Xiaokang Yang
• Trevor Darrell
MPN: MULTIMODAL
PARALLEL NETWORK • Jiashuo Yu
https://arxiv.org/pdf
18 FOR AUDIO-VISUAL 2021 • Ying Cheng X X X
/2104.02971.pdf
EVENT • Rui Feng
LOCALISATION
Zero-Gradient
Constrained • Kazuki
https://arxiv.org/pdf
19 Optimisation for 2021 Naganuma X X
/2104.02845.pdf
Destriping of 3D • Shunsuke Ono
Imaging Data
• Chen Min
• Jiaolong Xu
Attentional Graph
https://arxiv.org/pdf • Liang Xiao
20 Neural Network for 2021 X X
/2104.02576.pdf • Dawei Zhao
Parking-slot Detection
• Yiming Nie
• Bin Dai
Identity and Posture
• Vandad
Recognition in Smart https://arxiv.org/pdf
21 2019 Davoodnia X X
Beds with Deep /2104.02159.pdf
• Ali Etemad
Multitask Learning
• Abdulaziz M.
A Combined CNN and Alayba
https://arxiv.org/pdf
22 LSTM Model for Arabic 2018 • Vasile Palade X X
/1807.02911.pdf
Sentiment Analysis • Matthew England
• Rahat Iqbal
• Mathias Lechner
• Ramin Hasani
Adversarial Training is
https://arxiv.org/ab • Radu Grosu
23 Not Ready for Robot 2021 X X X
s/2103.08187 • Daniela Rus
Learning
• Thomas A.
Henzinger
http://iwqos2018.ie
Improved Adam ee-
24 Optimizer for iwqos.org/files/201 2018 • Zijun Zhang X X X X X
Deep Neural Networks 8/05/Improved_Ad
am_Optimizer.pdf
https://ieeexplore.i
Vision-Based Fall
eee.org/stamp/sta • Oussema Keskes
25 Detection Using ST- 2021 X X
mp.jsp?tp=&arnum • Rita Noumeir
GCN
ber=9351913
45
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
DBSCAN++: Towards
https://arxiv.org/pdf • Jennifer Jang
26 fast and scalable 2019 X X
/1810.13105.pdf • Heinrich Jiang
density clustering
MeanShift++:
Extremely Fast Mode-
Seeking With https://arxiv.org/pdf • Jennifer Jang
27 2021 X X X
Applications to /2104.00303.pdf • Heinrich Jiang
Segmentation and
Object Tracking
Self-correcting Q- https://arxiv.org/pdf • Rong Zhu
28 2021 X X
Learning /2012.01100.pdf • Mattia Rigotti
• Taha Mansouri
• Mohamadreza
A New Algorithm for https://arxiv.org/ftp/
Sadeghimoghada
29 Hidden Markov Models arxiv/papers/2102/ 2021 X X X X
m
Learning Problem 2102.07112.pdf
• Iman Ghasemian
Sahebi
Deep Reinforcement
Learning Aided Monte • Tz-Wei Mo
https://arxiv.org/pdf
30 Carlo 2021 • Ronald Y. Chang X X X
/2102.00178.pdf
Tree Search for MIMO • Te-Yi Kan
Detection
• Marcel R.
Analysis of Ackermann
https://arxiv.org/pdf
32 Agglomerative 2014 • Johannes Blomer X X
/1012.3697.pdf
Clustering • Daniel Kuntze
• Christian Sohler
DELVING INTO
• Yanpei Liu
TRANSFERABLE
https://arxiv.org/pdf • Xinyun Chen
34 ADVERSARIAL 2017 X X
/1611.02770.pdf • Chang Liu
EXAMPLES AND
• Dawn Song
BLACK-BOX ATTACKS
• Chaowei Xiao
• Bo Li
Generating Adversarial
https://arxiv.org/pdf • Jun-Yan Zhu
35 Examples with 2019 X X
/1801.02610.pdf • Warren He
Adversarial Networks
• Mingyan Liu
• Dawn Song
• Roman
Werpachowski
Detecting Overfitting via https://arxiv.org/pdf
36 2019 • András György X X
Adversarial Examples /1903.02380.pdf
• Csaba
Szepesvári
46
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
uptake-of-artificial- • Apostolos
intelligence-in- Malatras
autonomous- • Ignacio Sanchez
driving
Rapid Object Detection https://www.cs.cm
using a Boosted u.edu/~efros/cours • Paul Viola
38 2001 X X
Cascade of Simple es/LBMV07/Paper • Michael Jones
Features s/viola-cvpr-01.pdf
Detection of Advanced
https://arxiv.org/ftp/ • Sanjay Sharma
Malware
39 arxiv/papers/1903/ 2019 • C. Rama Krishna X X X
by Machine Learning
1903.02966.pdf • Sanjay K. Sahay
Techniques
https://www.theseu
MACHINE LEARNING
s.fi/bitstream/handl
METHODS FOR
e/10024/123412/T • Kateryna
40 MALWARE 2017 X X X
hesis_final.pdf?seq Chumachenko
DETECTION AND
uence=1&isAllowe
CLASSIFICATION
d=y
• Bojan Kolosnjaji
Adversarial Malware • Ambra Demontis
Binaries: Evading Deep • Battista Biggio
https://arxiv.org/pdf
41 Learning for Malware 2018 • Davide Maiorca X X X
/1803.04173.pdf
Detection in • Giorgio Giacinto
Executables • Claudia Eckert
• Fabio Roli
Generating Adversarial
Malware Examples for https://arxiv.org/pdf • Weiwei Hu
42 2017 X X X
Black-Box Attacks /1702.05983.pdf • Ying Tan
Based on GAN
Exploring Adversarial • Octavian Suciu
https://arxiv.org/pdf
43 Examples in Malware 2018 • Scott E. Coull X X X
/1810.08280.pdf
Detection • Jeffrey Johns
• Felix Kreuk
Deceiving End-to-End • Assi Barak
Deep Learning Malware https://arxiv.org/pdf • Shir Aviv-Reuven
44 2019 X X X
Detectors using /1802.04528.pdf • Moran Baruch
Adversarial Examples • Benny Pinkas
• Joseph Keshet
•Edward Raff
•Jon Barker
Malware Detection by https://arxiv.org/pdf •Jared Sylvester
45 2017 X X X
Eating a Whole EXE /1710.09435.pdf •Robert Brandon
•Bryan Catanzaro
•Charles Nicholas
Black-box attacks
against rnn based https://arxiv.org/pdf •Weiwei Hu
46 2017 X X X
malware detection /1705.08131.pdf •Ying Tan
algorithms
47
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Raphael Labaca-
Castro
• Luis Muñoz-
González
Universal Adversarial •Feargus
https://arxiv.org/pdf
48 Perturbations for 2021 Pendlebury X X X
/2102.06747.pdf
Malware •Gabi Dreo
Rodosek
•Fabio Pierazzi
• Lorenzo
Cavallaro
MDEA: Malware
• Xiruo Wang
Detection with https://arxiv.org/pdf
49 2020 • Risto X X X
Evolutionary /2002.03331.pdf
Miikkulainen
Adversarial Learning
Robust Android
• Hemant Rathore
Malware Detection
https://arxiv.org/pdf • Sanjay K. Sahay
51 System against 2021 X X X
/2101.12031.pdf • Piyush Nikam
Adversarial Attacks
• Mohit Sewak
using Q-Learning
48
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Wei Song
Automatic Generation
• Xuezixiang Li
of Adversarial
https://arxiv.org/pdf • Sadia Afroz
57 Examples for 2020 X X X
/2003.03100.pdf • Deepali Garg
Interpreting Malware
• Dmitry Kuznetsov
Classifiers
• Heng Yin
• Aminollah
Khormali
COPYCAT: Practical
• Ahmed
Adversarial Attacks on https://arxiv.org/pdf
58 2019 Abusnaina X X X
Visualisation-Based /1909.09735.pdf
• Songqing Chen
Malware Detection
• DaeHun Nyang
• Aziz Mohaisen
Effectiveness of
• Robert
Adversarial Examples https://arxiv.org/pdf
59 2019 Podschwadt X X X
and Defenses for /1909.04778.pdf
• Hassan Takabi
Malware Classification
• Shashank Srikant
• Sijia Liu
• Tamara
Generating Adversarial
Mitrovska
Computer Programs https://arxiv.org/pdf
60 2021 • Shiyu Chang X X X
using Optimiser /2103.11882.pdf
• Quanfu Fan
Obfuscations
• Gaoyuan Zhang
• Una-May
O'Reilly
• Ecenaz Erdemir
Adversarial Robustness
https://arxiv.org/pdf • Jeffrey Bickford
61 with Non-uniform 2021 X X X
/2102.12002.pdf • Luca Melis
Perturbations
• Sergul Aydore
• Alexandre Araujo
Robust Neural https://hal.archives
• Laurent Meunier
Networks using -ouvertes.fr/hal-
62 2019 • Rafael Pinot X X
Randomiser 02380184v2/docu
• Benjamin
Adversarial Training ment
Negrevergne
•Rafael Pinot
• Laurent Meunier
Theoretical evidence • Alexandre Araujo
for adversarial https://arxiv.org/pdf • Hisashi Kashima
63 2019 X X
robustness through /1902.01148.pdf • Florian Yger
randomisation • Cédric Gouy-
Pailler
• Jamal Atif
• Cihang Xie
Mitigating Adversarial • Jianyu Wang
https://arxiv.org/pdf
64 Effects Through 2017 • Zhishuai Zhang X X
/1711.01991.pdf
Randomisation • Zhou Ren
• Alan Yuille
•Tong Chen
https://cybersecurit
Adversarial attack and • Jiqiang Liu
y.springeropen.co
defense in • Yingxiao Xiang
65 m/track/pdf/10.118 2019 X X X X X
reinforcement learning- • Wenjia Niu
6/s42400-019-
from AI security view • Endong Tong and
0027-x.pdf
Zhen Han
49
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
TextAttack: A Framework • John X. Morris
for Adversarial Attacks, • Eli Lifland
Data https://arxiv.org/pdf • Jin Yong Yoo
66 2020 X X
Augmentation, and /2005.05909.pdf • Jake Grigsby
Adversarial Training in • Di Jin
NLP • Yanjun Qi
• Anirban
Chakraborty
• Manaar Alam
Adversarial Attacks and https://arxiv.org/pdf • Vishal Dey
67 2018 X X
Defences: A Survey /1810.00069.pdf • Anupam
Chattopadhyay
• Debdeep
Mukhopadhyay
DATA • Ferhat Ozgur
AUGMENTATION Catak
BASED MALWARE https://arxiv.org/pdf • Javed Ahmed
68 2021 X X X
DETECTION USING /2010.01862.pdf • Kevser Sahinbas
CONVOLUTIONAL • Zahid Hussain
NEURAL NETWORKS Khand
https://gala.gre.ac. • Nikolaos
uk/id/eprint/25226/ Pitropakisa
7/25226%20LOUK • Emmanouil
A Taxonomy and
AS_Taxonomy_An Panaousisb
Survey of Attacks
69 d_Survey_Of_Atta 2019 • Thanassis X X
Against Machine
cks_Against_Mach Giannetsosc
Learning
ine_Learning_%28 • Eleftherios
AAM%29_2019.pd Anastasiadisd
f • George Loukase
• Andrew Ilyas
• Shibani Santurkar
Adversarial examples • Dimitris Tsipras
https://arxiv.org/pdf
70 are not bugs, they are 2019 • Logan Engstrom X X
/1905.02175.pdf
features. • Brandon Tran
• Aleksander
Madry
• Rémi Bernhard
• Pierre-Alain
Moellic
Impact of Spatial
• Martial Mermillod
Frequency Based https://arxiv.org/pdf
71 2021 • Yannick Bourrier X X
Constraints on /2104.12679.pdf
• Romain
Adversarial Robustness
Cohendet
• Miguel Solinas
• Marina Reyboz
• Vikash Sehwag
• Saeed
Mahloujifar
Improving Adversarial
https://arxiv.org/pdf • Tinashe Handina
72 Robustness Using 2021 X X
/2104.09425.pdf • Sihui Dai
Proxy Distributions
• Chong Xiang
• Mung Chiang
• Prateek Mittal
• Guoqiu Wang
Improving Adversarial
https://arxiv.org/pdf • Huanqian Yan
73 Transferability with 2021 X X
/2105.04834.pdf • Ying Guo
Gradient Refining
• Xingxing Wei
50
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Guiyu Tian
Poisoning MorphNet for
https://arxiv.org/pdf • Wenhao Jiang
74 Clean-Label Backdoor 2021 X X
/2105.04839.pdf • Wei Liu
Attack to Point Clouds
• Yadong Mu
• Siyue Wang
High-Robustness, Low-
• Xiao Wang
Transferability https://arxiv.org/pdf
75 2021 • Pin-Yu Chen X X
Fingerprinting of Neural /2105.07078.pdf
• Pu Zhao
Networks
• Xue Lin
• Mingfu Xue
Robust Backdoor
• Can He
Attacks against Deep https://arxiv.org/pdf
77 2021 • Shichang Sun X X
Neural Networks in /2104.07395.pdf
• Jian Wang
Real Physical World
• Weiqiang Liu
secml-malware: A
Python Library for
https://arxiv.org/pdf • Luca Demetrio
78 Adversarial Robustness 2021 X X X
/2104.12848.pdf • Battista Biggio
Evaluation of Windows
Malware Classifiers
• Nicolas M. Müller
Defending against
• Simon
adversarial denial-of- https://arxiv.org/pdf
79 2021 Roschmann X X
service data poisoning /2104.06744.pdf
• Konstantin
attack
Böttinger
• Weiyi Zhang
• Shuning Zhao
Attack on practical
• Le Liu
speaker verification
https://arxiv.org/pdf • Jianmin Li
80 system using universal 2021 X X
/2105.09022.pdf • Xingliang Cheng
adversarial
• Thomas Fang
perturbations
Zheng
• Xiaolin Hu
Exploiting • Faiq Khalid •
Vulnerabilities in Deep Muhammad
https://arxiv.org/pdf
81 Neural Networks: 2021 Abdullah Hanif X X
/2105.03251.pdf
Adversarial and Fault- • Muhammad
Injection Attacks Shafique
Dynamic Defense
• Ruoxi Qin
Approach for Adversarial
https://arxiv.org/ftp/ • Linyuan Wang
Robustness in Deep
82 arxiv/papers/2105/ 2021 • Xingyuan Chen X X
Neural Networks via
2105.02803.pdf • Xuehui Du
Stochastic Ensemble
• Bin Yan
Smoothed Model
• Bangjie Yin
• Wenxuan Wang
Adv-Makeup: A New • Taiping Yao
Imperceptible and https://arxiv.org/pdf • Junfeng Guo
83 2021 X X
Transferable Attack on /2105.03162.pdf • Zelun Kong
Face Recognition • Shouhong Ding
• Jilin Li
• Cong Liu
51
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Micah Goldblum
Adversarial Attacks on
• Avi
Machine Learning https://arxiv.org/pdf
84 2020 Schwarzschild X X X
Systems for High- /2002.09565.pdf
• Ankit B. Patel
Frequency Trading
• Tom Goldstein
• Hyrum S.
Anderson
Learning Malware • Hyrum S.
https://arxiv.org/pdf
86 Models via 2018 Anderson X X X
/1801.08917.pdf
Reinforcement Learnin • Bobby Filar
• David Evans
• Phil Roth
Wild patterns: ten years
after the rise of https://arxiv.org/pdf • Battista Biggioa
87 2018 X X
adversarial Machine /1712.03141.pdf • Fabio Rolia
Learning
https://www.resear
chgate.net/publicat • Qiang Liu
A survey on security ion/323154427_A_ • Pan Li
threats and defensive Survey_on_Securit • Wentao Zhao
88 techniques of machine y_Threats_and_De 2018 • Wei Cai X X X X X X X X X
learning: a data driven fensive_Technique • Shui Yu
view s_of_Machine_Lea • Victor C. M.
rning_A_Data_Driv Leung
en_View
• Han Xu
• Yao Ma
Adversarial attacks and
• Haochen Liu
defenses in images, https://arxiv.org/pdf
89 2020 • Debayan Deb X X X X x X X
graphs and text: a /1909.08072.pdf
• Hui Liu
review
• Jiliang Tang
• Anil K. Jain
https://www.resear
chgate.net/publicat • Daniel Lowd
90 Adversarial Learning 2005 X X
ion/221654486_Ad • Christopher Meek
versarial_learning
• Ling Huang
https://citeseerx.ist. • Anthony D.
psu.edu/viewdoc/d Joseph
Adversarial Machine-
91 ownload?doi=10.1. 2011 • Blaine Nelson X X
Learning
1.360.168&rep=re • Benjamin I. P.
p1&type=pdf Rubinstein
• J. D. Tygar
https://www.resear
chgate.net/publicat
A Survey of
ion/327074374_A_
Adversarial Machine
92 Survey_of_Advers 2018 • Vasisht Duddu X X X X X X X X X X
Learning in Cyber
arial_Machine_Lea
Warfare
rning_in_Cyber_W
arfare
52
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Intelligence artificielle et https://www.wavest
cybersécurité : protéger one.com/app/uploa • Carole Meziat
93 2019 X X X X X
dès maintenant le ds/2019/09/IA- • Laurent Guille
monde de demain cyber-2019.pdf
• Ian J. Goodfellow
Explaining and
https://arxiv.org/pdf • Jonathon Shlens
94 Harnessing Adversarial 2020 X X
/1412.6572.pdf • Christian
Examples
Szegedy
https://papers.nips.
Feature Cross-
cc/paper/2014/file/ • Yevgeniy
Substitution in
95 8597a6cfa74defcb 2014 Vorobeychik X X
Adversarial
de3047c891d78f90 • Bo Li
Classification
-Paper.pdf
• Wenqi Wei
• Ling Liu
Adversarial Examples • Margaret Loper
in Deep Learning: http://arxiv.org/abs/ • Stacey Truex
96 2018 X X
Characterisation and 1807.00051 • Lei Yu
Divergence • Mehmet Emre
Gursoy
• Yanzhao Wu
Maximal Jacobian-
http://arxiv.org/abs/ • Rey Wiyatno
97 based Saliency Map 2018 X X
1808.07945 • Anqi Xu
Attack
Robustness and • Huan Xu
https://www.jmlr.or
Regularisation of • Constantine
98 g/papers/volume10 2009
Support Vector Caramanis
/xu09b/xu09b.pdf
Machines • Shie Mannor
• Xinyun Chen
Targeted Backdoor
• Chang Liu
Attacks on Deep https://arxiv.org/pdf
99 2017 • Bo Li X X
Learning Systems /1712.05526.pdf
• Kimberly Lu
Using Data Poisoning
• Dawn Song
Towards evaluating the
https://arxiv.org/ab • Nicholas Carlini
100 robustness of neural 2016 X X
s/1608.04644 • David Wagner
networks
• Reuben Feinman
• Ryan R. Curtin
Detecting adversarial https://arxiv.org/ab
101 2017 • Saurabh Shintre X X
samples from artifacts s/1703.00410
• Andrew B.
Gardner
Single Headed
Attention RNN: Stop https://arxiv.org/pdf
102 2019 • Steven Merity X X
Thinking With Your /1911.11423.pdf
Head
• Pauline Luc
Semantic Segmentation
https://arxiv.org/pdf • Camille Couprie
103 using Adversarial 2016 X X
/1611.08408.pdf • Soumith Chintala
Networks
• Jakob Verbeek
• Ming Pang
Unorganiser Malicious https://arxiv.org/pdf • Wei Gao
104 2018
Attacks Detection /1610.04086.pdf • Min Tao
• Zhi-Hua Zhou
53
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• CHRISTOPHE
ANDRIEU
https://link.springer • NANDO DE
An Introduction to
.com/content/pdf/1 FREITAS
105 MCMC for Machine 2003
0.1023/A:1020281 • ARNAUD
Learning
327116.pdf DOUCET
• MICHAEL I.
JORDAN
Scalable Optimisation
of Randomiser http://proceedings. • Bo Li
107 Operational Decisions mlr.press/v38/li15a 2015 • Yevgeniy X x
in Adversarial .pdf Vorobeychik
Classification Settings
54
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
https://towardsdata
science.com/poiso
Poisoning attacks on
114 ning-attacks-on- 2019 • Ilja Moisejevs X
Machine Learning
Machine-Learning-
1ff247c254db
• Mengchen Zhao
https://personal.ntu
Data Poisoning Attacks • Bo An
.edu.sg/boan/pape
116 on Multi-Task 2018 • Yaodong Yu X X X
rs/AAAI18_MTL.pd
Relationship Learning • Shulin Liu
f
• Sinno Jialin Pan
• Chang Liu
Robust High- • Bo Li
https://arxiv.org/pdf
117 Dimensional Linear 2016 • Yevgeniy X X
/1608.02257.pdf
Regression Vorobeychik
• Alina Oprea
• Jacob Steinhardt
Certified Defenses for https://arxiv.org/pdf
118 2017 • Pang Wei Koh X X X
Data Poisoning Attacks /1706.03691.pdf
• Percy Liang
• Ian J. Goodfellow
• Jean Pouget-
Abadie
• Mehdi Mirza
Generative adversarial https://arxiv.org/pdf • Bing Xu
119 2014 X X
nets /1406.2661.pdf • David Warde-
Farley
• Sherjil Ozair
• Aaron Courville
• Yoshua Bengio
• Qiang Liu
A Survey on Security • Pan Li
https://ieeexplore.i
Threats and Defensive • Wentao Zhao
eee.org/stamp/sta
120 Techniques of Machine 2018 • Wei Cai X X X X
mp.jsp?tp=&arnum
Learning: A Data • Shui Yu
ber=8290925
Driven View • Vctor C. M.
Leung
https://www.enisa.
europa.eu/publicati
Artificial Intelligence • Apostolos
ons/artificial-
121 Cybersecurity 2021 Malatras
intelligence-
Challenges • Georgia Dede
cybersecurity-
challenges
https://ieeexplore.i
A Taxonomy of ML for eee.org/stamp/sta
122 2020 • Martin Maas X X X X
Systems Problems mp.jsp?tp=&arnum
ber=9153088
55
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Deep Unsupervised
Learning for • Arjun Kaushik
Generaliser • Mehrazin
https://arxiv.org/pdf
123 Assignment Problems: 2021 Alizadeh X X X
/2103.14548.pdf
A Case-Study of • Omer Waqar
User-Association in • Hina Tabassum
Wireless Networks
https://elie.net/blog
Attacks against
/ai/attacks-against-
124 machine learning — an 2018 • Elie Bursztein X
machine-learning-
overview
an-over
https://towardsdata
science.com/how-
How to attack Machine to-attack-machine-
Learning ( Evasion, learning-evasion-
125 2019 • Alex Polyakov X X
Poisoning, Inference, poisoning-
Trojans, Backdoors) inference-trojans-
backdoors-
a7cb5832595c
Evaluating Input
Perturbation Methods • Lukas Brunke
https://arxiv.org/pdf
127 for Interpreting CNNs 2021 • Prateek Agrawal X X
/2101.10977.pdf
and Saliency Map • Nikhil George
Comparison
56
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Haotong Qina
• Ruihao Gonga
Binary Neural https://arxiv.org/pdf • Xianglong Liu
133 2020 X X
Networks: A Survey /2004.03333.pdf • Xiao Baie
• Jingkuan Songc
• Nicu Sebed
Towards Accurate • Xiaofan Lin
https://arxiv.org/pdf
134 Binary Convolutional 2017 • Cong Zhao X X
/1711.11294.pdf
Neural Network • Wei Pan
• Yunus Saatchi
https://arxiv.org/pdf
135 Bayesian GAN 2017 • Andrew Gordon X X X X X
/1705.09558.pdf
Wilson
• Neda Tavakoli
A Comparison of
https://par.nsf.gov/ • Sima Siami-
ARIMA and LSTM in
136 servlets/purl/10186 2018 Namini X X
Forecasting Time
768 • Akbar Siami
Series
Namin
• John Schulman
•Filip Wolski
Proximal Policy https://arxiv.org/pdf
137 2017 •Prafulla Dhariwal, X
Optimisation Algorithms /1707.06347.pdf
•Alec Radford,
•Oleg Klimov
https://op.europa.e
u/en/publication-
detail/-
Ethics Guidelines for
138 /publication/d3988 2019
Trustworthy AI
569-0434-11ea-
8c1f-
01aa75ed71a1
• Battista
BiggioIgino
https://link.springer • Igino Corona
Evasion Attacks against .com/content/pdf/1 • Davide Maiorca
139 Machine Learning at 0.1007%2F978-3- 2017 • Blaine Nelson X X
Test time 642-40994- • Nedim Srndie
3_25.pdf • Pavel Laskov
• Giorgio Giacinto
• Fabio Roli
Robustness
Evaluations of
https://www.mdpi.c • Corey Dunn
Sustainable Machine
om/2071- • Nour Moustafa
140 Learning Models 2020 X X
1050/12/16/6434/h • Benjamin
against Data Poisoning
tm Turnbull
Attacks in the Internet
of Things
Defending network
https://www.scienc
intrusion detection • Marek Pawlicki
edirect.com/scienc
141 systems against 2020 • Michał Choraś
e/article/abs/pii/S0
adversarial evasion • Rafał Kozik
167739X20303368
attacks
Defending SVMs
against Poisoning • Hu Ding
https://arxiv.org/pdf
142 Attacks: the Hardness 2021 • Fan Yang X X
/2006.07757.pdf
and DBSCAN • Jiawei Huang
Approach
57
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Battista Biggio
Bagging classifiers for http://pralab.diee.u
• Igino Corona
fighting poisoning nica.it/sites/default/
143 2011 • Giorgio Fumera X X X
attacks in adversarial files/biggio11-
• Giorgio Giacinto
classification tasks mcs.pdf
• Fabio Roli
• Tom B. Brown
• Dandelion Mané
https://arxiv.org/pdf
144 Adversarial Patch 2018 • Aurko Roy X X
/1712.09665.pdf
• Martín Abadi
• Justin Gilmer
STRIDE-AI: An
Approach to Identifying https://github.com/
• Lara Mauri
147 Vulnerabilities of LaraMauri/STRIDE 2021
• Ernesto Damiani
Machine Learning -AI
Assets
https://www.ai4eu.
eu/news/meaningf
ul-artificial-
For a meaningful
148 intelligencetowards 2018 • Cedric Villani
Artificial Intelligence
-french-artificial-
and-european-
strategy
https://www.ai4eu.
eu/news/strategic-
Strategic Action Plan action-plan-
149 2019
for Artificial Intelligence artificial-
intelligence-
netherlands
• Hamid Eghdal-
zadeh
• Khaled Koutini
• Paul Primus
• Verena
On Data Augmentation
https://arxiv.org/pdf Haunschmid
150 and Adversarial risk: An 2020 X X
/2007.02650.pdf • Michal
empirical Analysis
Lewandowski
• Werner Zellinger
• Bernhard
A.Moser
• Gerhard Widmer
https://ieeexplore.i
Review of Deep
eee.org/stamp/sta • Ajay Shrestha
151 Learning Algorithms 2019
mp.jsp?tp=&arnum • Ausif Mahmood
and Architectures
ber=8694781
58
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Learning in a Large • Benjamin I. P.
Function Space: Rubinstein
https://arxiv.org/pdf
152 Privacy-Preserving 2009 • Peter L. Bartlett X
/0911.5708.pdf
Mechanisms for SVM •Ling Huang
Learning • Nina Taft
• Liang Tong
Improving Robustness • Bo Li
of ML Classifiers • Chen Hajaj
https://arxiv.org/pdf
153 against Realizable 2019 • Chaowei Xiao X X
/1708.08327.pdf
Evasion Attacks Using • Ning Zhang
Conserved Features • Yevgeniy
Vorobeychik
• Chaowei Xiao
• Jun-Yan Zhu
Spatially Transformed https://arxiv.org/pdf • Bo Li
154 2018 X X
Adversarial Examples /1801.02612.pdf • Warren He
• Mingyan Liu
• Dawn Song
• Arjun Nitin
Exploring the Space of Bhagoji
https://arxiv.org/pdf
155 Black-box Attacks on 2015 • Warren He X X
/1712.09491.pdf
Deep Neural Networks • Bo Li
• Dawn Song
• Kevin Eykholt
• Ivan Evtimov
• Earlence
Robust Physical-World Fernandes
Attacks on Deep https://arxiv.org/pdf • Bo Li
156 2018 X X
Learning Visual /1707.08945.pdf • Amir Rahmati
Classification • Chaowei Xiao
• Atul Prakash
• Tadayoshi Kohno
• Dawn Song
• PIERRE-
FRANÇCOIS
Hybrid Isolation Forest - MARTEAU
https://arxiv.org/pdf
157 Application to Intrusion 2017 • SAEID SOHEILY- X X X X X X
/1705.03800.pdf
Detection KHAH
• NICOLAS
BÉCHET
• Neda Tavakoli
The Performance of
https://par.nsf.gov/ • Sima Siami-
LSTM and BiLSTM in
158 servlets/purl/10186 2019 Namini
Forecasting Time
554 • Akbar Siami
Series
Namin
• Maurras Togbe
• Mariam Barry
Anomaly Detection for
https://hal.archives • Aliou Boly
Data Streams Based on
-ouvertes.fr/hal- • Yousra
159 Isolation 2020 X X X
02874869/docume Chabchoub
Forest using Scikit-
nt • Raja Chiky
multiflow
• Jacob Montiel
• Vinh-Thuy Tran
59
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Wei Wang
Robust Differentiable https://arxiv.org/pdf • Zheng Dang
160 2021
SVD /2104.03821.pdf • Yinlin Hu
• Pascal Fua
Accurate Stock Price
https://arxiv.org/ftp/
Forecasting Using • Jaydip Sen
161 arxiv/papers/2103/ 2021 X X
Robust and Optimiser • Sidra Mehtab
2103.15096.pdf
Deep Learning Models
• Christian
Vulnerabilities of Berghoff
https://arxiv.org/pdf
ConnectioNIST 800-53
162 /2003.08837.pdf 2020 • Matthias Neu X X X X X X
AI Applications:
Evaluation and Defence • Arndt Von
Twickel
• Zhijie Deng
LiBRe: A Practical • Xiao Yang
https://arxiv.org/pdf
163 Bayesian Approach to 2021 • Shizhen Xu X X
/2103.14835.pdf
Adversarial Detection • Hang Su
• Jun Zhu
• Christian
Berghoff,
• Battista Biggio
• Elisa Brummel
• Vasilios Danos
• Thomas Doms
• Heiko Ehrich
• Thorsten
Gantevoort
• Barbara Hammer
• Joachim Iden
https://www.bsi.bu
nd.de/SharedDocs/ • Sven Jacob
Downloads/EN/BSI • Heidy Khlaaf
Towards Auditable AI /KI/Towards_Audit • Lars Komrowski
164 able_AI_Systems. 2021
Systems • Robert Kröwing
pdf?__blob=public
ationFile&v=4 • Jan Hendrik
Metzen
• Matthias Neu
• Fabian Petsch
• Maximilian
Poretschkin
• Wojciech Samek
• Hendrik Schäbe
• Arndt von
Twickel
• Martin Vechev
• Thomas
Wiegand
• Ian J. Goodfellow
• David Warde-
https://arxiv.org/pdf Farley
165 Maxout network 2013 X X
/1302.4389.pdf • Mehdi Mirza
• Aaron Courville
• Yoshua Bengio
60
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Victor
Shepardson
A Taxonomy of ML https://berryvilleiml.
166 2019 • Gary McGraw
Attacks com/taxonomy/
• Harold Figueroa
• Richie Bonett
https://towardsdata
AI Security and
science.com/ai-
168 Adversarial Machine 2019 • Alex Polyakov
and-ml-security-
Learning 101
101-6af8026675ff
https://www.resear
• Indira Kalyan
chgate.net/publicat
Generative Adversarial Dutta
ion/344519514_Ge
169 Networks in Security: A 2020 • Bhaskar Ghosh
nerative_Adversari
Survey • Michael Totaro
al_Networks_in_S
• Albert H. Carlson
ecurity_A_Survey
https://www.bsi.bu
nd.de/SharedDocs/
Secure, Robust and Downloads/EN/BSI • Bundesamt für
170 transparent application /KI/Secure_robust_ 2021 Sicherheit in der
of AI and_transparent_a Informationstechnik
pplication_of_AI.pd
f
• Bryant Chen
• Wilka Carvalho
• Nathalie
Detecting Backdoor Baracaldo
http://ceur-
Attacks on Deep Neural • Heiko Ludwig
171 ws.org/Vol- 2018 X X X
Networks by Activation • Benjamin
2301/paper_18.pdf
Clustering Edwards
• Taesung Lee
• Ian Molloy
• Biplav Srivastava
https://blogs.scienti
ficamerican.com/o
When AI Misjudgement
173 bservations/when- 2018 • Douglas Yeung
is not an accident
ai-misjudgment-is-
not-an-accident/
https://www.media. • Samuel
mit.edu/publication G.Finlayson
Adversarial attacks on
s/adversarial- • Jonathan Zittrain
174 medical machine 2029
attacks •Joi Ito
learning
-on-medical- • Andrew L.Beam
machine-learning/ • Isaac S.Kohane
61
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
https://thenewstack
Camouflaged grafiti on
.io/camouflaged-
road signs can fool
175 graffiti-road-signs- 2017 • Kimberley Mok X X
machine learning
can-fool-machine-
models
learning-models/
https://www.idgcon
nect.com/article/35
Deepfakes and deep 83356/deepfakes-
fraud: the new security and-deep-fraud-
176 challenge the-new-security- 2020 • Sadia Sajiad
of misinformation and challenge-of-
impersonation misinformation-
and-
impersonation.html
Stealing
https://arxiv.org/pdf • B.Wang
177 Hyperparameters in 2019 X X X X
/1802.05351.pdf • N.Z.Gong
Machine Learning
Explanation-Guided • Giorgio Severi
https://www.usenix
Backdoor Poisoning • Jim Meyer
178 .org/system/files/se 2020 X X
Attacks Against • Scott Coull
c21-severi.pdf
Malware Classifiers • Alina Oprea
• Gamaleldin F.
Elsayed
Adversarial Examples • Shreya Shankar
that Fool both • Brian Cheung
//arxiv.org/pdf/1802
179 Computer 2018 • Nicolas Papernot X X
.08195.pd
Vision and Time- • Alex Kurakin
Limited Humans • Ian Goodfellow
• Jascha Sohl-
Dickstein
With Great training
https://www.usenix • Bolun Wang
comes great
.org/system/files/co • Bimal Viswanath
vulnerability:
180 nference/usenixse 2018 • Yuanshun Yao
Practical Attacks
curity18/sec18- • Haitao Zheng
against transfer
wang.pdf • Ben Y.Zhao
learning
• Nicolas Papernot
• Patrick McDaniel
Practical black-box • Ian Goodfellow
https://arxiv.org/pdf
181 attacks against 2017 • Somesh Jha X X
/1602.02697.pdf
machine learning • Z. Berkay Celik
• Ananthram
Swami
• Seyed-Mohsen
Moosavi-Dezfooli
Universal adversarial https://arxiv.org/pdf
182 2017 • Alhussein Fawzi X X
perturbations /1610.08401.pdf
• Omar Fawzi
• Pascal Frossard
Benchmarking neural
• Dan Hendrycks
network robustness to https://arxiv.org/pdf
183 2019 • Thomas X X
common corruptions /1903.12261.pdf
Dietterich
and perturbations
62
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
https://www.resear
chgate.net/publicat
ion/341792988_Se
Securing Artificial curing_Artificial_Int
Intelligence, Part 1 The elligence_Part_1_
184 attack surface of The_attack_surfac 2019 • Sven Herping
machine learning and e_of_machine_lear
its implications ning_and_its_impli
cations/link/5ed506
4a299bf1c67d3238
f4/download
Support vector http://proceedings. • Battista Biggio
185 machines under mlr.press/v20/biggi 2011 • Blaine Nelson X X X
adversarial label noise o11/biggio11.pdf • Pavel Laskov
• Avi
Just How Toxic is Data Schwarzschild
Poisoning? A Unified https://arxiv.org/pdf
• Micah Golddblum
186 Benchmark for /2006.12557.pdf 2020 X X
• Arjun Gupta
Backdoor and
• John P.Dickerson
Data Poisoning Attacks
• Tom Goldstein
• Huang Xiao
• Battista Biggio
Is feature selection
https://arxiv.org/pdf • Gavin Brown
187 secure against training 2018
/1804.07933.pdf • Giorgio Fumera
data poisoning?
• Claudia Eckert
• Fabio Roli
• Yujie Ji
• Xinyang Zhang
Model-reuse attacks on https://arxiv.org/pdf
189 2018 • Shouling Ji X X X X X
deep learning systems /1812.00483.pdf
• Xiapu Luo
• Ting Wang
https://ieeexplore.i
A survey on transfer • Sinno Jialin Pan
190 eee.org/document/ 2009
learning • Qiang Yang
5288526
• Ilia Shumailov
• Yiren Zhao
Sponge examples: https://arxiv.org/pdf
• Daniel Bates
191 Energy-Latency Attacks /2006.03463.pdf 2021 X X
• Nicolas Papernot
on Neural Network
• Robert Mullins
• Ross Anderson
• Jiongcong Chen
. Impact analysis of • Gaoqi Liang
https://link.springer
false data injection • Zexiang CAI
.com/content/pdf/1
192 attacks 2016 • Chunchao Hu
0.1007%2Fs40565
on power system static • Yan Xu
-016-0223-6.pdf
security assessment • Fengji Luo
• Junhua Zhao
63
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
https://www.resear
chgate.net/publicat
The ND2DB attack: ion/250195790_Th
Database content e_ND2DB_attack_ • Ariel Futoransky
193 extraction using timing Database_content 2007 • Damian Saura X
attacks on the indexing _extraction_using_ • Ariel Waissbein
algorithms timing_attacks_on
_the_indexing_alg
orithms
La nouvelle technologie
194 de protection des 2020 • Théo Ryffel
données
Reliable evaluation of
adversarial robustness https://arxiv.org/pdf
• Francesco Croce
195 with an ensemble of /2003.01690.pdf 2020 X X
diverse parameter-free • Matthias Hein
attacks
• M.Jagielski
Manipulating Machine • Alina Oprea
Learning: Poisoning • Battista Biggio
https://arxiv.org/pdf
196 attacks and 2021 • Chang Liu X X
/1804.00308.pdf
countermeasures for • Cristina Nita-
regression learning Rotaru
• Bo Li
• Yansong Gao
• Chang Xu
STRIP: A defence
• Derui Wang
against trojan attacks https://arxiv.org/pdf
198 2020 • Shiping Chen X X
on deep neural /1902.06531.pdf
• Damith C.
networks
Ranasinghe
•Surya Nepal
• Blaine Nelson
• Marco Barreno
• Fuching Jack Chi
https://people.eecs • Anthony
Misleading learners: .berkeley.edu/~tyg D.Joseph
199 Co-opting your spam ar/papers/SML/mis 2009 • Benjamin I.P X X
filter leading.learners.pd Rubinstein
f •Udam Saini
• Charles Sutton
• J.D Tygar
•Kai Xia
64
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Tong Chen
Gradient band-based • Wenjia Niu
adversarial training for • Yingxiao Xiang
https://arxiv.org/pdf
201 generaliser attack 2018 • Xiaoxuan Bai X X
/1807.06752.pdf
immunity of A3C path • Jiqiang Liu
finding. •Zhen Han
•Gang Li
https://www.jmlr.or • Kamalika
Differentially Private
g/papers/volume12 Chaudhuri
203 Empircal Risk 2011 X X
/chaudhuri11a/cha • Claire Monteleoni
Minimisation
udhuri11a.pdf • Anand D.Sarwate
•Elham Tabassi
•Kevin J. Burns
A Taxonomy and
https://csrc.nist.gov •Michael
Terminology of
204 /publications/detail/ 2019 Hadjimichael
Adversarial Machine
nistir/8269/draft •Andres D.
Learning
Molina-Markham
•Julian T. Sexton
Deep Defense: Training • Ziang Yan
https://arxiv.org/pdf
205 DNNs with improved 2018 • Yiwen Guo X X
/1803.00404.pdf
adversarial Robustness • Changshui Zhang
Improving the
Adversarial Robustness • Andrew Slavin
and Interpretability of https://arxiv.org/pdf Ross
206 2017 X X
Deep Neural Network /1711.09404.pdf • Finale Doshi-
by Regularizing their Velez
input Gradients
• Jan Hendrik
Metzen
On detecting https://arxiv.org/pdf
207 2017 • Tim Genewein X X
Adversarial perturbation /1702.04267.pdf
• Volker Fischer
• Bastian Bischoff
• Nicolas Papernot
https://ieeexplore.i
SoK: Security and • Patrick McDaniel
eee.org/stamp/sta
208 privacy in machine 2017 • Arunesh Sinha
mp.jsp?tp=&arnum
learning • Michael P.
ber=8406613
Wellman
• Qinglong Wang
Random Feature • Wenbo Guo
Nullification for https://arxiv.org/pdf • Kaixuan Zhang
209 2016 X X
Adversary Resistant /1610.01239v1.pdf • Xinyu Xing
Deep Architecture • C. Lee Giles
• Xue Liu
https://people.eecs
• Marco Barreno
.berkeley.edu/~adj/
The security of machine • Blaine Nelson
210 publications/paper- 2010 X X
learning • Anthony D.
files/SecML-
Joseph
MLJ2010.pdf
65
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
• Florian Tramer
• Alexey Kurakin
Ensemble adversarial
https://arxiv.org/pdf • Nicolas Papernot
211 training: attacks and 2018 X X
/1705.07204.pdf • Ian Goodfellow
defenses
• Dan Boneh
• Patrick McDaniel
• Nicolas Papernot
Distillation as a defense • Patrick McDaniel
to adversarial https://arxiv.org/pdf • Xi Wu
213 2016 X X
perturbations against /1511.04508.pdf • Somesh Jha
deep neural networks • Ananthram
Swami
• Nicolas Papernot
• Shuang Song
• Ilya Mironov
Scalable Private https://arxiv.org/pdf
214 2016 • Ananth X X
Learning with Pate /1802.08908.pdf
Raghunathan
• Kunal Talwar
• Ulfar Erlingsson
https://seclab.stanf
Gradient masking in ord.edu/AdvML201
215 2017 • Nicolas Papernot
machine learning 7/slides/17-09-aro-
aml.pdf
• Kamalika
Near-Optimal
Chaudhuri
Algorithms for https://arxiv.org/
221 2013 • Anand D. X X
Differentially-Private pdf/1207.2812.pdf
Sarwate
principal components
• Kaushik Sinha
66
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
Differentially private
model selection via http://proceeding • Adam Smith
222 Stability argument and s.mlr.press/v30/Gu 2013 • Abhradeep X
the robustness of the ha13.pdf Thakurta
Lasso
• Shehzeen
Hussain
WaveGuard:
https://www.usenix • Paarth Neekhara
Understanding and
223 .org/system/files/se 2021 • Shlomo Dubnov X X
Mitigating Audio
c21-hussain.pdf • Julian McAuley,
Adversarial Examples
• Farinaz
Koushanfar
http://researchers.li
lle.inria.fr/abellet/te
Privacy Preserving
224 aching/private_ma 2020 •Aurélien Bellet
Machine Learning
chine_learning_co
urse.html
Interpretability of https://www.wavest
Machine Learning one.com/app/uploa
What are the ds/2019/09/Wavest •Alexandre Vérine
225 2019
challenges in the era of one_Interpretability •Stephan Mir
automated decision- _machine_learning
Making Progresses? .pdf
https://www.analyti
6 Python Libraries to csvidhya.com/blog/
interpret Machine 2020/03/6-python-
226 2020 •Purva Huilgol
Learning Models and libraries-interpret-
Build Trust machine-learning-
models/
• Cristoph Molnar
• Gunnar König
• Julia Herbinger
• Timo Freiesleben
• Susanne Dandl
Pitfalls to avoid when https://arxiv.org/pdf
• Christian
227 Interpreting Machine / 2020
A.Scholbeck
Learning Models 2007.04131.pdf
• Giuseppe
Casalicchio
• Moritz Grosse-
Wentrup
• Bernd Bischl
https://towardsdata
science.com/7-
7 steps to ensure and steps-to-ensure-
228 2019 • Stéphanie Shen
sustain Data quality and-sustain-data-
quality-
3c0040591366
67
SECURING MACHINE LEARNING ALGORITHMS
December 2021
ed Learning
ent learning
Reinforcem
Unsupervis
Supervised
Learning
Type of data
ingested
Publication date
Index
Structured Data
Classification
Time series
Regression
Dimension
Rewarding
Clustering
Reduction
Image
Video
Text
https://techcrunch.
com/2018/11/06/3-
ways-to-avoid-
bias-in-machine-
learning/?guccount
er=1&guce_referre
r=aHR0cHM6Ly93
d3cuZ29vZ2xlLmN
vbS8&guce_referr
er_sig=AQAAAIfXh
Three ways to avoid VIDfTYv80Vxw4JV
229 bias in machine aKFZyt_3_2DTapB 2018 • Vince Lynch
learning QQjW8C1vzjPTQq
ViKdAE5O-
BV1Q5J5waGMcY
o4yu2R4QBOr9H1
7RpApdX9vlXDUlo
_MS28Q4GD8qCX
qhogX534lJcR7DP
rzwANTY8WPnJX
5GXVmytlHxM0ZK
9Ym2ANzKGSnae
6QgH
https://www.kdnug
5 Ways to deal with the gets.com/2019/06/
•Alexandre
230 lack of data in Machine 5-ways-lack-d 2019
Gonfalonieri
Learning ata-machine-
learning.html
https://towardsdata
science.com/artifici
Artificial Intelligence is al-intellige
Crucial TO the Success nce-is-crucial-to-
231 2019 • Amit Makhija
of Your Business and the-success-of-
here is why your-business-
learn-why-
d5b96fa3564d
https://www.kaggle
7 Simple Techniques to •Devendra Kumar
232 .com/learn- 2020
Prevent Overfitting Yadav
forum/157623
68
TP-06-21-153-EN-N
ABOUT ENIS A
The European Union Agency for Cybersecurity, ENISA, is the Union’s agency dedicated to
achieving a high common level of cybersecurity across Europe. Established in 2004 and
strengthened by the EU Cybersecurity Act, the European Union Agency for Cybersecurity
contributes to EU cyber policy, enhances the trustworthiness of ICT products, services and
processes with cybersecurity certification schemes, cooperates with Member States and EU
bodies, and helps Europe prepare for the cyber challenges of tomorrow. Through knowledge
sharing, capacity building and awareness raising, the Agency works together with its key
stakeholders to strengthen trust in the connected economy, to boost resilience of the Union’s
infrastructure, and, ultimately, to keep Europe’s society and citizens digitally secure. More
information about ENISA and its work can be found here: www.enisa.europa.eu.
ISBN: 978-92-9204-543-2
DOI: 10.2824/874249