Guidelines On Securing AI Systems
Guidelines On Securing AI Systems
ON
SECURING
AI SYSTEMS
CYBER SECURITY AGENCY OF SINGAPORE
OCTOBER 2024
TABLE OF
CONTENTS
1. INTRODUCTION ....................................................................................................... 3
1.1. PURPOSE AND SCOPE OF THIS DOCUMENT ..................................................... 4
2. UNDERSTANDING AI THREATS ................................................................................. 5
3. SECURING AI .......................................................................................................... 7
3.1. TAKE A LIFECYCLE APPROACH ......................................................................... 7
3.2. START WITH A RISK ASSESSMENT ..................................................................... 8
3.3. GUIDELINES FOR SECURING AI SYSTEMS ....................................................... 10
GLOSSARY .................................................................................................................... 14
ANNEX A........................................................................................................................ 18
GUIDELINES ON SECURING AI SYSTEMS
2
1. INTRODUCTION
Artificial Intelligence (AI) poses benefits for economy, society, and
national security. It has the potential to drive efficiency and innovation
in almost every sector – from commerce and healthcare to
transportation and cybersecurity.
To reap the benefits of AI, users must have confidence that the AI will behave as designed,
and outcomes are safe and secure. However, in addition to safety risks, AI systems can
be vulnerable to adversarial attacks, where malicious actors intentionally manipulate or
deceive the AI system. The adoption of AI can introduce or exacerbate existing
cybersecurity risks to enterprise systems. These can lead to risks such as data
leakage or data breaches, or result in harmful or otherwise undesired model
outcomes.
The Cyber Security Agency of Singapore (CSA) has developed Guidelines on Securing AI
Systems for system owners to secure the use of AI throughout its lifecycle. As AI is
increasingly integrated into enterprise systems, security should be considered holistically
at the system level. As such, these guidelines should be used alongside existing security
best practices and requirements for IT environments. While these guidelines are not
mandatory, we strongly encourage system owners to consider these key principles, so
that they can make informed decisions on their adoption of AI vis-à-vis the potential risks.
GUIDELINES ON SECURING AI SYSTEMS
3
1.1. PURPOSE AND SCOPE OF
THIS DOCUMENT
Purpose
These guidelines are designed to support systems owners that are adopting, or considering
the adoption of AI systems. It identifies potential security risks associated with the use of AI
and sets out guidelines for mitigating security risks at each stage of the AI lifecycle.
This document can be read together with the Companion Guide on Securing AI Systems,
which provides an informative compilation of practical security control measures, that
system owners may consider in implementing these guidelines.
Scope
These guidelines address the cybersecurity risks to AI systems. It does not seek to address
AI safety, or other common attendant considerations for AI such as fairness, transparency
or inclusion, or cybersecurity risks introduced by AI systems, although some of the
recommended actions may overlap. It also does not address the misuse of AI in
cyberattacks (AI-enabled malware), mis/disinformation, and scams (deepfakes).
GUIDELINES ON SECURING AI SYSTEMS
4
2. UNDERSTANDING
AI THREATS
AI is a type of software system, and is itself vulnerable to cyber threats,
while also posing a new attack surface for the broader enterprise
system that it is integrated to, or interfaces with. As such, securing AI
is in addition to practising good ‘classical’ cybersecurity hygiene.
Figure 1. Classical and AI-specific risks of AI systems– diagram adapted from OWASP1
GUIDELINES ON SECURING AI SYSTEMS
1
Threats overview - https://owaspai.org/docs/ai_security_overview/
5
CLASSICAL CYBERSECURITY RISKS TO AI SYSTEMS
AI systems require vast amounts of data for training; some also require importing
external models and libraries. If inadequately secured, AI systems can be
undermined by supply chain attacks, or may be susceptible to intrusion or
unauthorised access, through vulnerabilities in the AI model or the underlying IT
infrastructure. In addition, organisations and users risk losing the ability to
access and use AI tools if there are disruptions to cloud services, data centre
operations, or other digital infrastructure (e.g. through Denial of Service attacks),
this could in turn disable systems that depend on AI tools to function.
Malicious actors may use novel Adversarial ML techniques to attack AI models and
data, influencing machine learning models to produce inaccurate, biased, or
harmful output; and/or reveal confidential information. Adversarial ML 2 attacks
include: data poisoning (injecting malicious or corrupted data into training data
sets) or evasion attacks (on trained models) to distort outcomes, inference attacks
or extraction attacks (probing the model) to expose sensitive or restricted data, or
to steal the model.
GUIDELINES ON SECURING AI SYSTEMS
2
A Taxonomy and Terminology of Attacks and Mitigations https://csrc.nist.gov/pubs/ai/100/2/e2023/final. The MITRE ATLAS is a
useful reference to understand and situate classical cybersecurity risks from Adversarial ML.
6
3. SECURING AI
The security of AI is a widely cited concern, but this field of work is still
relatively nascent. While practitioners continue to grow the body of
research and resources on the security threats to AI, these guidelines
lay out key considerations that system owners should take to support
secure adoption of AI. Given the rapid speed of AI development,
system owners should continue to apprise themselves on the latest
developments in AI security, and refresh their risk management
strategies accordingly.
As with good cybersecurity practice, CSA recommends that system owners take a lifecycle
approach to consider security risks. Hardening only the AI model is insufficient to ensure
a holistic defence against AI related threats. All stakeholders involved across the lifecycle
of an AI system should seek to better understand the security threats and their potential
impact on the desired outcomes of the AI system, and what decisions or trade-offs will
need to be made.
7
Figure 2: AI System Development Lifecycle (AI SDLC)
Some organisations may have implemented the Machine Learning Operations (ML Ops)
pipeline, which may not map exactly to the AI SDLC. Nonetheless, ML Ops teams that run
a dev ops pipeline comprising ML Design, Development and Operation stages (similar to
Figure 3), will find the guidelines across the AI SDLC's stages of Planning & Design,
Development, Deployment and Operations relevant.
instructions for each task. As such, organisations should consider conducting risk
assessments more frequently than for conventional systems, even if they generally base
their risk assessment approach on existing governance and policies. These assessments
may also be supplemented by continuous monitoring and a strong feedback loop.
We recommend these four steps to tailor a systematic defence plan that best addresses
your organisation’s highest priority risks – protecting the things you care about the most.
8
STEP 1
Conduct risk assessment, focusing on security risks to AI systems
Conduct a risk assessment, focusing on security risks related to AI systems, either based
on best practices or your organisation’s existing Enterprise Risk Assessment/Management
Framework.
Risk assessment can be done with reference to CSA published guides, if applicable:
▪ Guide To Cyber Threat Modelling
▪ Guide To Conducting Cybersecurity Risk Assessment for Critical Information
Infrastructure
STEP 2
Prioritise areas to address based on risk/impact/resources
Prioritise which risks to address, based on risk level, impact, and available resources.
STEP 3
Identify and implement the relevant actions to secure the AI
system
Identify relevant actions and control measures to secure the AI system, such as by
referencing those outlined in the Companion Guide on Securing AI Systems and
implement these across the AI life cycle.
STEP 4
GUIDELINES ON SECURING AI SYSTEMS
Evaluate the residual risk after implementing security measures for the AI system to inform
decisions about accepting or addressing residual risks.
9
3.3. GUIDELINES FOR
SECURING AI SYSTEMS
These guidelines apply across the various lifecycle stages of the AI
system. System owners should read these as key issues to consider
in securing their adoption of AI. In view of the diversity of use cases
and developments in AI security, these guidelines do not provide
prescriptive controls or requirements.
System owners should apply these to their specific context, and can
reference the Companion Guide to Securing AI systems for potential
controls.
Organisations should understand the potential security risks posed by AI, in order to
make informed decisions about adoption. Provide adequate training and guidance on
the security risks of AI to all personnel, including developers, system owners and senior
leaders.
10
2. DEVELOPMENT
The AI supply chain includes (but is not limited to) the training data, models, APIs, and
software libraries. Each of these components may introduce new vulnerabilities (e.g,
models may carry malware encoded as model parameters that could enable attackers
to extract and inject malicious software onto user machines). Assess and monitor
potential security risks of the AI system’s supply chain across its life cycle. Ensure that
suppliers adhere to security policies and internationally recognised standards, or that
risks are otherwise appropriately managed. Consider evaluating supply chain
components (e.g. through Software Bills of Material [SBOM], code checking, or against
vulnerability databases).
Different AI models (e.g. machine learning, deep learning, generative) pose unique
characteristics and risks (e.g. LLMs can be vulnerable to input manipulation attacks)
and as such require different security measures. When developing or selecting an
appropriate AI model for your system, consider factors which may affect its security
(such as complexity, explainability, interpretability, and sensitivity of training data).
11
3. DEPLOYMENT
AI systems are complex and adaptive, and this can sometimes result in unpredictable
behaviour. Given the diversity in AI use cases, incidents can range from minor issues
such as malfunctioning chat bots to critical outcomes such as disruption in the
operation of critical infrastructure. System owners should put in place appropriate
incident response, escalation and remediation plans.
AI systems can be vulnerable to the risks described above, including misuse, data
breaches, and model manipulation. These have impact on the trust and confidence of
users, and may have reputational implications for organisations. A good practice is to
release models, applications or systems only after subjecting them to appropriate and
effective security checks and evaluation.
AI systems are dynamic and adaptive to input. There have already been real-life
incidents, in which users/ attackers have deliberately crafted input to trick AI systems
into making incorrect or unintended decisions. AI system owners may wish to monitor
GUIDELINES ON SECURING AI SYSTEMS
and log inputs to the AI system, such as queries, prompts and requests, as third-party
providers may not do so due to privacy reasons. Proper logging allows for compliance,
audit, investigation and remediation.
12
4.2. Monitor AI system outputs and behaviour
Changes to the data and model can lead to changes in behaviour. System owners
should ensure that risks associated to model updates have been considered and
appropriately managed.
Even with monitoring mechanisms in place, the adaptive nature of AI can make it
challenging to detect attacks and unintended behaviour. There should be a feedback
process for users to share any findings of concern, which might uncover potential
vulnerabilities to the system.
5. END OF LIFE
As models are trained on large amounts of training data (incl. potentially confidential
information), improper disposal can lead to incidents such as data breaches. There
should be proper and secure disposal/destruction of data and models in accordance
with relevant industry standards or regulations.
GUIDELINES ON SECURING AI SYSTEMS
13
GLOSSARY
Term Brief description
Adversarial The process of extracting information about the behaviour and characteristics
Machine of an ML system and/or learning how to manipulate the inputs into an ML
Learning system in order to obtain a preferred outcome.
Anomaly The identification of observations, events or data points that deviate from what
Detection is usual, standard, or expected, making them inconsistent with the rest of
data.
Backdoor A backdoor attack is when an attacker subtly alters AI models during training,
attack causing unintended behaviour under certain triggers.
Data Breach Data Breach occurs when a threat actor gains unauthorised access to
sensitive/confidential data.
Data Integrity The property that data has not been altered in an unauthorised manner. Data
integrity covers data in storage, during processing, and while in transit.
14
Data Leakage Unintentional exposure of sensitive, protected, or confidential information
outside its intended environment.
Data Loss A system’s ability to identify, monitor, and protect data in use (e.g., endpoint
Prevention actions), data in motion (e.g., network actions), and data at rest (e.g., data
storage) through deep packet content inspection, and contextual security
analysis of transaction (e.g., attributes of originator, data object, medium,
timing, recipient/destination, etc.) within a centralised management
framework.
Data Science An interdisciplinary field of technology that uses algorithms and processes to
gather and analyse large amounts of data to uncover patterns and insights that
inform business decisions.
Deep Learning A function of AI that imitates the human brain by learning from how it
structures and processes information to make decisions. Instead of relying on
an algorithm that can only perform one specific task, this subset of machine
learning can learn from unstructured data without supervision.
Evasion attack Crafting input to AI in order to mislead it into performing its task incorrectly.
Extraction Copy or steal an AI model by appropriately sampling the input space and
attack observing outputs to build a surrogate model that behaves similarly.
Generative AI A type of machine learning that focuses on creating new data, including text,
video, code and images. A generative AI system is trained using large amounts
GUIDELINES ON SECURING AI SYSTEMS
Guardrails Restrictions and rules placed on AI systems to make sure that they handle
data appropriately and don't generate unethical content.
15
Hallucination An incorrect response from an AI system, or false information in an output that
is presented as factual information.
Image Image recognition is the process of identifying an object, person, place, or text
Recognition in an image or video.
ML Machine Learning.
A subset of AI that incorporates aspects of computer science, mathematics,
and coding. Machine learning focuses on developing algorithms and models
that can learn from data, and make predictions and decisions about new data.
Membership Data privacy attacks to determine if a data sample was part of the
Inference training set of a machine learning model.
attack
Neural Network A deep learning technique designed to resemble the human brain’s structure.
Neural networks require large data sets to perform calculations and create
outputs, which enables features like speech and vision recognition.
Overfitting Occurs in machine learning training when the algorithm can only work on
specific examples within the training data. A typical functioning AI model
should be able to generalise patterns in the data to tackle new tasks.
GUIDELINES ON SECURING AI SYSTEMS
Prompt A prompt is a natural language input that a user feeds to an AI system in order
to get a result or output.
Reinforcement A type of machine learning in which an algorithm learns by interacting with its
Learning environment and then is either rewarded or penalised based on its actions.
16
SDLC Software Development Life Cycle
Training data Training data is the information or examples given to an AI system to enable it
to learn, find patterns, and create new content.
GUIDELINES ON SECURING AI SYSTEMS
17
ANNEX A
UNDERSTANDING AI THREATS
Adversarial threats are caused by threat actors with deliberate intention to cause harm.
Typically, these threat actors are referred to as attackers or adversaries.
To understand these threats, system owners can refer to resources such as the OWASP
Top 10 for Large Language Model Applications, or OWASP Machine Learning Security Top
10, or the MITRE ATLAS™ (Adversarial Threat Landscape for Artificial-Intelligence Systems).
The MITRE ATLAS in particular provides a structured knowledge base for AI and
cybersecurity professionals to understand and defend against AI cyber threats. It
compiles adversary tactics, techniques, and case studies for AI systems based on real-
world observations, demonstrations from ML red teams and security groups, as well as
state-of-the-possible from academic research.
Any attempt to secure an AI system should be on top of the ‘traditional’ good cybersecurity
hygiene, such as implementing the principle of least privileges, multi-factor
authentication, continuous security monitoring and auditing.
The ATLAS3 Matrix (see Table A1) covers 2 types of adversarial ‘techniques’.
• Techniques specific to AI/ML systems (indicated in orange boxes), and
• Techniques that are conventional cybersecurity offensive techniques, but
applicable to both AI and non-AI systems and come directly from the MITRE
Enterprise ATT&CK Matrix (indicated in white boxes).
System owners should continue to build their awareness of security threats using these
resources, to better understand emerging risks that may have implications on their
adoption of AI. As this space continues to evolve, such resources will aid both AI and cyber
teams in their security risk assessment and management activities.
GUIDELINES ON SECURING AI SYSTEMS
3
MITRE ALTAS Framework: https://atlas.mitre.org/. It leverages the same core principles and structure of the well-known
MITRE ATT&CK (Adversarial Tactics, Techniques, and Common Knowledge) framework, which is widely used by cyber
defenders to map the terminologies of cybersecurity attacks. The ATLAS adapts these to the unique context of AI systems and
potential adversarial attacks.
18
Table A1: MITRE ATLAS Matrix
19