AI Cyber Benchmark
AI Cyber Benchmark
© WAVESTONE | 2
Wavestone AI Cyber Benchmark - 2025
LEARNING Non-deterministic
Non completely
explainable
On learning…
PROCESSING
INPUTS
On the outputs…
On processing…
© WAVESTONE | 3
Wavestone AI Cyber Benchmark - 2025
An approach focused on The EU positioned itself as the China focuses on pushing for best
positioning the US as an AI leader world’s police officer and push practices in AI management and
for citizen protection data management
• Risk-based approach.
• Rescind former Executive Order
14110 that provided • Every organization must comply by • China is focusing on the
guidelines. May 2027. cybersecurity of its system with a
• Aims to remove any potential • Already some consequences: new risk-based approach and on
barriers to AI development. iPhone with GenAI & ChatGPT voice regulating the processing of data,
chat functionality pospotned … especially labeling.
© WAVESTONE | 5
Wavestone AI Cyber Benchmark - 2025
• Govern
We benched these clients on their AI • Identify
Worked with +20 maturity, based on the 5 NIST’s pillars, • Protect
clients already and consolidated those results to • Detect
working on the topic. produce this first AI Cyber Benchmark. • Respond
© WAVESTONE | 6
Wavestone AI Cyber Benchmark - 2025
Govern Identify
Maturity 39% 39%
Our recommendation: compensate with an integrated governance that will help people to augment their skills
~60% of ~10% of
our
clients
Integrated model clients Decentralized model
AI HUB
~30% of Cybersecurity Legal CSR
Cyber for AI Legal for AI CSR/Ethics for clients in AI specific activities AI specific activities AI specific activities
Privacy for AI hybrid
(security, (data, third AI (fairness,
(privacy)
resilience) parties) sustainable) mode
© WAVESTONE | 9
Wavestone AI Cyber Benchmark - 2025
Of companies assessed 1
have an AI security policy
• Frame use of AI large public Assess them against a limited set of questions
drawn from the 4 pillars of:
64% •
application
Indicates the process to secure • Intended use • Data and Input • Task and Output
AI project
2 • AI Models
• Integrate Third Party stance
against AI
Classify them into four risk categories based on the AI
Act:
3
Unacceptable Strong Moderate Minimal
© WAVESTONE | 10
Wavestone AI Cyber Benchmark - 2025
But most of AI use case we assessed are typically used for non-critical processes that don't demand high
availability or strict integrity, often relying on human oversight
© WAVESTONE | 11
Protect: there is no “one-size-
fits-all” approach
Protect
Maturity 40%
Example of an AI architecture
Build infrastructure
End users
Monitoring infrastructure
© WAVESTONE | 13
Wavestone AI Cyber Benchmark - 2025
Build infrastructure
End users
Of our clients adapted their
Monitoring infrastructure
40% Third Party assessment
methodology for AI vendors
Plug-in & Interconnections
Component to protect
© WAVESTONE | 14
Wavestone AI Cyber Benchmark - 2025
Example of an AI architecture
• Set up criteria to choose the right model: whitelist
suppliers, code review, operational testing…
Training dataset Model Applicative
or RAG frontend • Build inputs and output controls
In-house documentation- API / SaaS • Ensure proper security of the front end
driven enrichment In house User GUI
• Make AI project “secure by design” with MLSecOps
Build infrastructure
End users
Of our clients have a model
Monitoring infrastructure
43% selection process to identify
trusted sources
Plug-in & Interconnections
Component to protect
© WAVESTONE | 15
Wavestone AI Cyber Benchmark - 2025
Component to protect
© WAVESTONE | 16
Wavestone AI Cyber Benchmark - 2025
Must-have controls
Standards controls
High-end controls
88 nd
a s
ic
Eth
Scan me for the
M ne r
vu
Providers
od ab
l
full publication
el ilit
identified in
ro y a
bu ss
September 2024 se
stn ess
n
d R ng
po
ess me
Detection an earni
es
and nt
eL
Machin
Sy
An nth
on e A
c I
tic iza
re plia
ym
om
da tion
gu nc
ta
lat e
AI risk manageme
/
ion
ection
s
AI & Privacy
ta Prot
nt
Da
AI security radar • 2024 • 18
Some companies have offers covering more than one category: our
ke
© WAVESTONE | 18
Detection: two pillars to combine
Detect
Maturity 29%
… that we tested and adapted to land our AI redteam framework on the market
7%
Assessing AI capabilities and biases Assessing AI systems flaws Of our clients use
Hallucination, Misinformation, Robustness, Pre-prompt access, Input/Output filtering, Illegitimate advanced model
Harmfulness Prompt Injection… internal data retrieval , API limitations, Detection &
monitoring robustness evaluation
New approach and tooling required, often using LLM to attack LLM !
© WAVESTONE | 20
Wavestone AI Cyber Benchmark - 2025
Standard ML Specific
© WAVESTONE | 21
Wavestone AI Cyber Benchmark - 2025
72% 13%
Respond
Maturity 9%
90%
80%
70%
60%
50%
40%
30%
20%
10%
0%
Align your effort with your stance: start steady… but start now!
AI Advanced Creators
© WAVESTONE | 27
Wavestone AI Cyber Benchmark - 2025
High
3 Use of deepfake for phishing / crisis exercises
activities 4 Document creation and modification assistance 12
6 5
5 Third Party Security questionnaires analysis 14 15
ADDED
9
VALUE
11 7
6 Automated labelisation for DLP 13
Accelerate cyber 1
7 Live data anonymization (text/voice)
processes 8 2 3
8 Augmented redteam / attack path discovery 10
4
9 Source code security analysis
Low
Difficult Easy
10 GenAI SOC Copilots FEASABILITY
Reinvent detection 11 SOC playbook update via ML Use case highlighted are offered by a large number software vendor
and reaction 12 AI-based automated reaction / attack blocking Use case underlined and bold are the most implemented by our clients
13 User behavior analysis for nudging
Contact
Gérôme BILLOIS
Partner
[email protected]
Wavestone.com
Wavestone.com