Computer Security 2020
Computer Security 2020
Computer Security
ESORICS 2019 International Workshops,
CyberICPS, SECPRE, SPOSE, and ADIoT
Luxembourg City, Luxembourg, September 26–27, 2019
Revised Selected Papers
Lecture Notes in Computer Science 11980
Founding Editors
Gerhard Goos
Karlsruhe Institute of Technology, Karlsruhe, Germany
Juris Hartmanis
Cornell University, Ithaca, NY, USA
Computer Security
ESORICS 2019 International Workshops,
CyberICPS, SECPRE, SPOSE, and ADIoT
Luxembourg City, Luxembourg, September 26–27, 2019
Revised Selected Papers
123
Editors Frédéric Cuppens
Sokratis Katsikas IMT Atlantique
Open University of Cyprus Brest, France
Latsia, Cyprus
Costas Lambrinoudakis
and University of Piraeus
Piraeus, Greece
Norwegian University of Science
and Technology John Mylopoulos
Gjøvik, Norway University of Toronto
Toronto, ON, Canada
Nora Cuppens
IMT Atlantique Stefanos Gritzalis
Brest, France University of Piraeus
Piraeus, Greece
Christos Kalloniatis
University of the Aegean Jörg Pohle
Mytilene, Greece Alexander von Humboldt Institute
for Internet and Society
Annie Antón
Berlin, Germany
Georgia Institute of Technology
Atlanta, GA, USA Weizhi Meng
Technical University of Denmark
Frank Pallas Kongens Lyngby, Denmark
Technical University of Berlin
Berlin, Germany Joaquin Garcia-Alfaro
Télécom SudParis
Angela Sasse Evry, France
Ruhr University Bochum
Bochum, Germany
Steven Furnell
University of Plymouth
Plymouth, UK
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
CyberICPS 2019 Preface
This book contains revised versions of the papers presented at the 5th Workshop on
Security of Industrial Control Systems and Cyber-Physical Systems (CyberICPS 2019).
The workshop was co-located with the 24th European Symposium on Research in
Computer Security (ESORICS 2019) and was held in Luxembourg during September
26–27, 2019.
Cyber-physical systems (CPS) are physical and engineered systems that interact
with the physical environment, whose operations are monitored, coordinated, con-
trolled, and integrated by information and communication technologies. These systems
exist everywhere around us, and range in size, complexity, and criticality, from
embedded systems used in smart vehicles, to SCADA systems in smart grids, to control
systems in water distribution systems, to smart transportation systems, to plant control
systems, engineering workstations, substation equipment, programmable logic con-
trollers (PLCs), and other Industrial Control Systems (ICS). These systems also include
the emerging trend of Industrial Internet of Things (IIoT) that will be the central part
of the fourth industrial revolution. As ICS and CPS proliferate, and increasingly
interact with us and affect our lives, their security becomes of paramount importance.
CyberICPS 2019 brought together researchers, engineers, and governmental actors
with an interest in the security of ICS and CPS in the context of their increasing
exposure to cyber-space, by offering a forum for discussion on all issues related to their
cyber security.
CyberICPS 2019 attracted 13 high-quality submissions, each of which was assigned
to 3 referees for review; the review process resulted in 5 full and 2 short papers being
accepted to be presented and included in the proceedings. These cover topics related to
threats, vulnerabilities, and risks that cyber-physical systems and industrial control
systems face; cyber attacks that may be launched against such systems; and ways of
detecting and responding to such attacks.
We would like to express our thanks to all those who assisted us in organizing the
events and putting together the programs. We are very grateful to the members of the
Program Committee for their timely and rigorous reviews. Thanks are also due to the
Organizing Committee for the events. Last, but by no means least, we would like to
thank all the authors who submitted their work to the workshop and contributed to an
interesting set of proceedings.
General Chairs
Nora Cuppens IMT Atlantique, France
Costas Lambrinoudakis University of Piraeus, Greece
Program Chairs
Sokratis Katsikas Norwegian University of Science and Technology,
Norway and Open University of Cyprus, Cyprus
Frédéric Cuppens IMT Atlantique, France
Publicity Chair
Anis Bkakria IMT Atlantique, France
Program Committee
Marios Anagnostopoulos Norwegian University of Science and Technology,
Norway
Youssef Laarouchi EDF R&D, France
Chris Mitchell Royal Holloway, University of London, UK
Vasileios Gkioulos Norwegian University of Science and Technology,
Norway
Stefano Zanero Politecnico di Milano, Italy
Samiha Ayed IMT Atlantique, France
Mauro Conti University of Padua, Italy
Joaquin Garcia-Alfaro Télécom SudParis, France
Pankaj Pandey Center for Cyber and Information Security (CCIS),
Norwegian University of Science and Technology,
Norway
Houbing Song Embry-Riddle Aeronautical University, USA
Georgios Spathoulas University of Thessaly, Greece
Andrea Saracino CNR, Italy
Weizhi Meng University of Denmark, Denmark
Sjouke Mauw University of Luxembourg, Luxembourg
Cristina Alcaraz University of Malaga, Spain
Khan Ferdous Wahid Airbus Group Innovations, Germany
David Espes University of Brest, France
Dieter Gollmann Hamburg University of Technology, Germany
Masahiro Mambo Kanazawa University, Japan
SECPRE 2019 Preface
This volume contains revised versions of the papers presented at the Third International
Workshop on SECurity and Privacy Requirements Engineering (SECPRE 2019) which
was co-located with the 24th European Symposium on Research in Computer Security
(ESORICS 2019) held in Luxembourg during September 26–27, 2019.
For many years, software engineers have focused on the development of new
software, thus considering security and privacy mainly during the development stage as
an ad-hoc process rather than an integrated one initiated during the system design stage.
However, the data protection regulations, the complexity of modern environments
(such as IoT, IoE, Cloud Computing, Big Data, Cyber-Physical Systems, etc.) and the
increased level of users awareness in IT have forced software engineers to identify
security and privacy as fundamental design aspects leading to the implementation of
more trusted software systems and services. Researchers have addressed the necessity
and importance of implementing design methods for security and privacy requirements
elicitation, modeling, and implementation in the last decades in various innovative
research domains. Today Security by Design (SbD) and Privacy by Design (PbD) are
established research areas that focus on these directions. The new GDPR regulation sets
even stricter requirements for organizations regarding its applicability. SbD and PbD
play a very critical and important role in assisting stakeholders in understanding their
needs, complying with the new legal, organizational, and technical requirements, and
finally selecting the appropriate measures for fulfilling these requirements. SECPRE
aimed to provide researchers and professionals with the opportunity to present novel
and cutting-edge research on these topics.
SECPRE 2019 attracted 14 high-quality submissions, each of which was assigned to
3 referees for review; the review process resulted in 9 papers being selected for
presentation and inclusion in these proceedings. The topics covered include: security
and privacy requirements, assurance and evaluation, security requirements elicitation
and modeling, and GDPR compliance.
We would like to express our thanks to all those who assisted us in organizing the
events and putting together the programs. We are very grateful to the members of the
Program Committee for their timely and rigorous reviews. Thanks are also due to the
Organizing Committee of the events. Last, but by no means least, we would like to
thank all the authors who submitted their work to the workshop and contributed to an
interesting set of proceedings.
General Chairs
Annie Antón Georgia Institute of Technology, USA
Stefanos Gritzalis University of Piraeus, Greece
Program Chairs
John Mylopoulos University of Toronto, Canada
Christos Kalloniatis University of the Aegean, Greece
Program Committee
Frédéric Cuppens IMT Atlantique, France
Sabrina De Capitani Università degli Studi di Milano, Italy
di Vimercati
Vasiliki Diamantopoulou University of the Aegean, Greece
Eric Dubois Luxembourg Institute of Science and Technology,
Luxembourg
Carmen Fernandez-Gago University of Malaga, Spain
Eduardo Fernandez-Medina University of Castilla-La Mancha, Spain
Mohamad Gharib University of Florence, Italy
Maritta Heisel University of Duisburg-Essen, Germany
Jan Juerjens University of Koblenz-Landau, Germany
Costas Lambrinoudakis University of Piraeus, Greece
Tong Li Beijing University of Technology, China
Javier Lopez University of Malaga, Spain
Fabio Martinelli CNR, Italy
Aaron Massey University of Maryland, USA
Haralambos Mouratidis University of Brighton, UK
Liliana Pasquale University College Dublin, Ireland
Michalis Pavlidis University of Brighton, UK
William Robinson Georgia State University, USA
David Garcia Rosado University of Castilla-La Mancha, Spain
Mattia Salnitri University of Trento, Italy
Pierangela Samarati Università degli Studi di Milano, Italy
Jessica Staddon North Carolina State University, USA
Nicola Zannone Eindhoven University of Technology, The Netherlands
Jianying Zhou Singapore University of Technology and Design,
Singapore
SPOSE 2019 Preface
Over the past decades, a multitude of security and privacy enhancing technologies have
been developed and brought to considerable maturity. However, the design and
engineering of such technologies often ignores the organizational context that
respective technologies are to be applied in. A large and hierarchical organization,
for example, calls for significantly different security and privacy practices and
respective technologies than a small and agile startup. Similarly, whenever employees
behavior plays a significant role for the ultimate level of security and privacy provided,
their individual interests and incentives as well as typical behavioral patterns must be
taken into account and materialized in concrete technical solutions and practices. Even
though research on security- and privacy-related technologies increasingly takes into
account questions of practical applicability in realistic scenarios, respective approaches
are typically still rooted in the technical domain alone, motivated by technical givens
and constraints from the practice.
On the other hand, a substantial body of organization-related security and privacy
research already exists, incorporating aspects like decision-making and governance
structures, individual interests and incentives of employees, organizational roles and
procedures, organizational as well as national culture, or business models and
organizational goals. Nonetheless, these research activities are only rarely translated
into concrete technical mechanisms, frameworks, and systems.
This disconnection between rather technical and rather organization-related security
and privacy research leaves substantial room for improving the fit between concrete
technologies on the one hand and organizational practices on the other hand. Achieving
a better fit between these two sides through security and privacy technologies that
soundly incorporate organizational and behavioral theories and practices promises
substantial benefits for organizations and data subjects, engineers, policy makers, and
society as a whole.
The aim of the First Workshop on Security, Privacy, Organizations, and Systems
Engineering (SPOSE 2019) therefore was to discuss, exchange, and develop ideas and
questions regarding the design and engineering of technical security and privacy
mechanisms with particular reference to organizational contexts. We invited
researchers and practitioners working in security- and privacy-related systems
engineering as well as in the field of organizational science to submit their
contributions. Besides regular and short papers, we also invited practical demonstra-
tions, intermediate reports, and mini-tutorials on respective technologies currently
under development to stimulate forward-looking discussions.
The papers included on the following pages demonstrate the possible spectrum for
fruitful research at the intersection of security, privacy, organizational science, and
systems engineering. Yang Liu and Andrew Simpson present an empirical study on
privacy attitudes, utility preferences, and respective trade-offs of mobile device users
across cultural boundaries. In their demo-paper, Hugh Lester and Martin Miller provide
xiv SPOSE 2019 Preface
insights into the practical application of discrete event simulation for fostering
organizational change in the domain of jail operations, which may (and, in fact,
actually did during the workshop) serve as an inspirational mind-opener for thinking
about structurally similar challenges in the context of information security. Jan
Zibuschka, in turn, presents a first analysis of existing automation potentials for the
execution of privacy impact assessments as demanded for several use-cases by the EU
General Data Protection Regulation. Finally, Sebastian Pape and Jelena Stankovic
present revealing results from an empirical study on the role of security aspects in
companies’ selection of cloud providers.
Altogether, these papers, complemented by an open-minded, keen-to-debate, and
constructively thinking audience as well as a stunning keynote on “A User-Centric
Approach to Secure the Internet Ecosystem” given by Katharina Krombholz from the
Saarbrücken-based CISPA Helmholtz Center for Information Security, made the first
iteration of the workshop a raving success that definitely calls for many recurrences in
the years to come. We would like to thank everybody who contributed to this success –
authors, presenters, participants, reviewers, and, of course, the organizing team of
ESORICS 2019 - and are looking forward to the next iteration of SPOSE.
Organizers
Frank Pallas TU Berlin, Germany
Jörg Pohle Humboldt Institute for Internet and Society, Germany
Angela Sasse Ruhr-University Bochum, Germany
Program Committee
Jatinder Singh Cambridge University, UK
Seda Gürses KU Leuven, Belgium
Ronald Leenes Tilburg University, The Netherlands
Burkhard Schäfer Edinburgh University, UK
Heleen Janssen Cambridge University, UK
Daniel Le Métayer Inria, France
Marit Hansen ULD Schleswig-Holstein, Germany
Melanie Volkamer KIT, Germany
Ingolf Becker UCL, UK
Max-R. Ulbricht TU Berlin, Germany
Additional Reviewers
Peter Mayer KIT, Germany
Alireza Zarei KIT, Germany
ADIoT 2019 Preface
This volume contains the papers that were selected for presentation and publication at
the Second International Workshop on Attacks and Defenses for Internet-of-Things
(ADIoT 2019), which was held in Luxembourg on September 26, 2019. Internet of
Things (IoT) technology is widely adopted by the vast majority of businesses and is
impacting every aspect of the world. However, the nature of the Internet, communi-
cation, embedded OS, and backend recourses make IoT objects vulnerable to cyber
attacks. In addition, most standard security solutions designed for enterprise systems
are not applicable to IoT devices. As a result, we are facing a big IoT security and
protection challenge, and it is urgent to analyze IoT-specific cyber attacks to design
novel and efficient security mechanisms. This workshop focused on IoT attacks and
defenses, and sought original submissions that discussed either practical or theoretical
solutions to identify IoT vulnerabilities and IoT security mechanisms.
This year, 5 full papers out of 16 submissions were selected with an acceptance rate
of 31.3%. We also accepted two short papers, and all papers were reviewed by at least
three members of the Program Committee. We would like to extend our thanks to the
Program Committee members as well as the additional reviewers who contributed their
precious time and expertise to provide professional reviews and feedback to authors in
a timely manner. We would also like to express our thanks to all the authors who
submitted papers to ADIoT 2019.
General Chairs
Anthony T. S. Ho University of Surrey, UK
Shouhuai Xu University of Texas at San Antonio, USA
Program Chairs
Weizhi Meng Technical University of Denmark, Denmark
Steven Furnell University of Plymouth, UK
Qian Chen University of Texas at San Antonio, USA
Program Committee
Alessandro Bruni IT University of Copenhagen, Denmark
Kai Chen Chinese Academy of Sciences, China
Sebastien Faye Luxembourg Institute of Science and Technology,
Luxembourg
Jinguang Han Queen’s University Belfast, UK
Georgios Kambourakis University of the Aegean, Greece
Sokratis Katsikas Norwegian University of Science and Technology,
Norway
Wenjia Li New York Institute of Technology, USA
Rongxing Lu University of New Brunswick, Canada
Xiaobo Ma Xi’an Jiaotong University, China
Javier Parra-Arnau Universitat Rovira i Virgili, Spain
Kewei Sha University of Houston–Clear Lake, USA
Jun Shao Zhejiang Gongshang University, China
Hao Wang Shandong Normal University, China
Ding Wang Peking University, China
Lam Kwok Yan Nanyang Technological University, Singapore
Guomin Yang University of Wollongong, Australia
Kehuan Zhang The Chinese University of Hong Kong, Hong Kong
Peng Zhou Shanghai University, China
xx ADIoT 2019 Organization
Additional Reviewers
Yue Zhao
Mitra Sananda
Florian Gondesen
Yuxuan Chen
Marios Anagnostopoulos
Zhixiu Guo
Steering Committee
Steven Furnell University of Plymouth, UK
Anthony T. S. Ho University of Surrey, UK
Sokratis Katsikas Norwegian University of Science and Technology,
Norway
Weizhi Meng (Chair) Technical University of Denmark, Denmark
Shouhuai Xu University of Texas at San Antonio, USA
Contents
CyberICPS Workshop
SECPRE Workshop
SPOSE Workshop
ADIoT Workshop
1 Introduction
Since Stuxnet struck a nuclear facility in 2010, threats towards industrial con-
trol systems (ICSs) have increased. Unfortunately, as most of ICS attackers
are state-sponsored and use zero-day vulnerabilities, signature-based detection
(maintaining the blacklist and updating it) is inappropriate.
The most common and safe approaches that do not harm the availability
of ICSs monitor the network traffic of these systems. ICSs present more peri-
odic behavior than information technology systems. Several studies [5,13,20]
applied the statistical characteristics of traffic for ICS-specific security mech-
anisms. Although this approach is suitable for ICS traffic characteristics, it
presents limitations in detecting attacks at the ICS operation level.
Other researches focused on the detection of anomalies with physical prop-
erties [14]. By using the specification or control logic, the monitoring system
rarely emits false alarms [3,15]. However, it is relatively expensive to obtain
and specify the specification or control logic. An ICS recognizes the environ-
ment with sensors, makes decisions for its purpose, and delivers the right action
with actuators. To detect anomalies at the ICS operation level, its normal state
must be defined and the control logic that decides actuators’ behaviors must be
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 3–18, 2020.
https://doi.org/10.1007/978-3-030-42048-2_1
4 J. Kim et al.
understood. However, understanding the entire set of the control logic is com-
plicated. In fact, the volume of the control logic is enormous, and acquiring it
from vendors is not allowed in most cases.
Herein, the aim is to monitor the ICS operational data. A feasible alterna-
tive is the data-driven approach. Machine-learning-based approaches have been
highly studied and especially deep-learning-based anomaly detection methods
which have been reported recently using fully-connected networks (FCN) [18],
convolutional neural networks (CNN) [10], recurrent neural networks (RNN) [7],
and generative adversarial networks (GAN) [12].
We propose a deep learning-based anomaly detection method using a
sequence-to-sequence model (seq2seq) [19]. Seq2seq is designed initially for nat-
ural language translation. It encodes the words of a given sentence with RNN
into a latent vector, then decodes from it to a set of words in the target language.
Seq2seq’s encoding-decoding approach presents a significant advantage as it can
understand the context of the entire sentence, while vanilla RNN gives the out-
put immediately for every input. Seq2seq is expected to be an effective method
for learning the context or semantics of time-series operational data, and obtain
a better prediction based on the given data.
To date, no abnormal samples have been reported to train machine learning
models robustly. Therefore, the reported model is trained with the only normal
dataset (training dataset), and it is considered that the training data are clean. In
the detection phase, the developed model investigates unseen data with trained
neural networks. Using the model after the learning phase, the detection method
proceeds in three steps: Step 1, the model predicts the future values of the sensors
and actuators, Step 2, the difference between the prediction and actual data is
determined, and Step 3, alerts are sent for significant differences.
The rest of this paper is organized as follows. Section 2 introduces the
anomaly detection method using the seq2seq neural network. Section 3 presents
the experimental results after applying the proposed method to the secure water
treatment (SWaT) dataset [6]. Section 4 analyzes the experimental results in
detail and Sect. 5 concludes this study.
2 Proposed Method
2.1 SWaT Dataset
Several studies have recently been reported on dataset generation for ICS
research [4,6,11,17]. The most frequently used dataset is the SWaT dataset [6]
by Singapore University of Technology and Design (SUTD), which has opera-
tional data and attack labels. Herein, the method is developed and evaluated
with the SWaT dataset.
The SWaT dataset was collected from a testbed water treatment system.
Fifty-one tags (25 sensors and 26 actuators) are sampled every second. Some
are digital, and others are analog. Tag names define their roles. For example,
MV denotes motorized valve, P for pump, FIT for flow meter, and LIT for level
transmitter.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 5
SWaT consists of six processes. Water flows from process 1 to the process 6.
The numbers following the tag names indicate the process ID and the gadget
ID. For example, MV-101 is the first motorized valve in process 1.
In the SWaT dataset, normal and attack datasets are separated. The normal
part has 480,800 samples, and the attack part includes 41 attacks during 449,919
samples.
Fig. 2. Prediction errors for processes 1, 2, 3, 4, and 5 of SWaT (red: attack, purple:
prediction error, and blue: anomaly score) (Color figure online)
operational data with the attention mechanism. The decoder part is optional.
The reason why we added a decoder with attention is that it gives us more
accurate results. The decoder predicts the last second with a 9-second hint. The
values of 9th second of the hint is almost the same with those of the last record
of the window (we wanted to expect), which helps the model give almost-zero
prediction error. The shape of the decoder’s final output is (batch, # of tags).
An independent model was applied for each process of SWaT. The model
n learns the process n. Figure 2 shows the prediction error of each process. At
the early stage of this study, a holistic model1 was tested for six processes. The
result was not accurate because each process shows a different prediction error
pattern, especially the process 2, as observed in Fig. 2.
1
[18] used this approach: one model for the whole processes.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 7
where n is the number of tags, Ii is the i-th tag values in the dataset, and Oi is
the i-th output of the model.
The proposed method considers that the system is under attack if the model
has never seen the current state. The developed model was trained to perform
a precise prediction. When the model detects a never-seen window, it cannot
perform an accurate prediction, which leads to more notable 4-norm value.
Multiple approaches can be used to determine anomalies such as cumula-
tive sum (CUSUM) and anomaly likelihood [1]. The custom rating method was
applied by considering the prediction errors – due to the following factors:
Fig. 3. Asymmetric ratings at the start and end of attack 41 (red: attack, purple:
prediction error, and blue: anomaly score) (Color figure online)
3 Experiment
All source codes, pre-processed datasets, trained network parameters, and results
are available at https://github.com/jukworks/swat-seq2seq.
3.1 Training
Occasionally, a neural network goes bad local minima during training process. A
general approach to solve this issue is to train the neural network multiple times
independently and choose the best result among the multiple training results.
As a neural network is trained with a stochastic gradient descent and mostly
initialized with random numbers, different results are obtained every run. Two
independent training sessions were run for each model (each process) and the
network presenting the lowest training loss was chosen.
The model was optimized with Adam [9], amsgrad [16], and without weight
decay. Each model trained 150 epochs with a 4,096 batch size. Early stopping
was not applied.
The hardware consisted of Intel Xeon CPU E5-2960 v4 2.60 GHz, 6 NVIDIA
Tesla V100, and 512 GB RAM. Table 1 shows that the training time was approx-
imately 2 h on average. Six models have a different size of input and output, but
their internal LSTM architectures were entirely the same. Therefore, the number
of trainable parameters is similar, which leads to similar training time.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 9
Fig. 4. LIT-101 was not stable after attack 30 happens (from [10])
The results are compared with Kaspersky’s research [18] because this is the only
study providing the list of found attack points.
If an alert is received within 15 min after the attack range of SWaT, it is
considered as a detection. 15 min is chosen as attacks attacks on cyber-physical
systems tend to have a long impact. The shortest attack in SWaT, attack 29,
lasts 2 min2 . Figure 4 shows LIT-101 on attack 30 [10]. The two black vertical
lines represent the start and end of the attack. After the end of the attack, LIT-
101 was unstable. As the reported model learns the normal-labeled data only, it
may perceive the stabilizing region as an anomaly.
Tables 2 and 3 compare the detection results with those of [18]. Attacks 5,
9, 12, 15, and 18 have no physical impact. These attacks were ignored as they
cannot be detected with operational data. The first column shows the attack
numbers. The second column presents the answer to attack points labeled in
the SWaT dataset. The third column represents the detection of the attack: Yes
means 100% sure, not sure means 30%–100% means sure, and No means less
than 30% sure. As mentioned in Sect. 2.5, 30% represents the heuristic threshold.
The fourth and fifth column represents the attack points determined by the
model and [18], respectively. Bold text indicates correct answers. The parentheses
indicate the second-longest distance. N/A indicates that the model failed to
detect.
[18] reported 25 attacks and 11 attack points (nine with the first predictions
and 2 with the second predictions). Herein, the model found 29 attacks and 25
attack points (22 with the first predictions and 3 with the second predictions).
21 attacks were detected by both the developed model and that in [18]. Four
2
The longest attack in SWaT is attack 28, which lasted 9.5 h.
10 J. Kim et al.
Table 2. Anomaly detection results compared with those of [18] (attacks 1 to 30)
Attack Answer Detection (our Attack point (our work) Attack point ([18])
# (by SWaT) work)
1 MV-101 Yes MV-101 N/A
2 P-102 Yes MV-101 (P-102) MV-301 (P-102)
3 LIT-101 Not sure (65%) MV-101 (LIT-101) N/A
4 MV-504 No N/A N/A
6 AIT-202 Yes AIT-202 AIT-202 (P-203)
7 LIT-301 Yes LIT-301 LIT-301 (PIT-502)
8 DPIT-301 Yes DPIT-301 DPIT-301 (MV-302)
10 FIT-401 Yes FIT-401 FIT-401 (PIT-502)
11 FIT-401 Yes FIT-401 (FIT-504) MV-304 (MV-302)
13 MV-304 No MV-304 N/A
14 MV-303 No N/A N/A
16 LIT-301 Yes LIT-301 MV-301 (MV-303)
17 MV-303 Yes MV-301 (MV-303) N/A
19 AIT-504 No (15%) AIT-504 AIT-504 (P-501)
20 AIT-504 Yes AIT-504 N/A
21 MV-101, Not sure (35%) LIT-101 UV-401 (P-501)
LIT-101
22 UV-401, Yes FIT-401, FIT-504 DPIT-301 (MV-302)
AIT-502,
P-501
23 P-602, Yes DPIT-301 P-302, P-203
DPIT-301,
MV-302
24 P-203, No N/A LIT-401
P-205
25 LIT-401, No (20%) LIT-401 P-602, MV-303
P-401
26 P-101, Yes (25% at P-102, LIT-301 LIT-401 (AIT-402)
LIT-301 P3)
27 P-302, Yes LIT-401 N/A
LIT-401
28 P-302 Yes FIT-401, AIT-504 MV-201, LIT-101
29 P-201, No N/A LIT-401, AIT-503
P-203,
P-205
30 LIT-101, Yes LIT-101 LIT-301 (FIT-301)
P-101,
MV-201
attacks were detected only by the model in [18]: attacks 19, 24, 25, and 29. Eight
attacks were detected only by the developed model: attacks 1, 3, 17, 20, 27, 32,
33, and 41, which also detected attacks 21 and 31 with 35% and 50% rates,
respectively. The SWaT dataset indicates that the attack points of attack 35 is
process 1, but it was detected by the developed model for process 3 (20% rate
at the model for process 1). Both methods failed to detect three attacks: attacks
4, 13, and 14.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 11
Table 3. Anomaly detection results compared with those of [18] (attacks 31 to 41)
Attack Answer (by SWaT) Detection (our Attack point Attack point ([18])
# work) (our work)
31 LIT-401 Not sure (50%) LIT-101 P-602, MV-303
32 LIT-301 Yes LIT-301 N/A
33 LIT-101 Yes LIT-101 N/A
34 P-101 Yes P-101 MV-201 (P-203)
35 P-101, P-102 Not sure (20% at P-101 MV-201, MV-303
P1, 45% at P3)
36 LIT-101 Yes LIT-101 LIT-101, AIT-503
37 P-501, FIT-502 Yes FIT-401, FIT-504 (FIT-503)
FIT-504
38 AIT-402, AIT-502 Yes (15% at P5) MV-101, AIT-502, AIT-402
AIT-402,
AIT-502
39 FIT-401, AIT-502 Yes FIT-401 FIT-401, P-201
40 FIT-401 Yes FIT-401, UV-401 (FIT-401)
FIT-504
41 LIT-301 Yes LIT-301 N/A
Fig. 5. List of detected attacks and comparison with [18]. Red (4, 14, 29) represents
attacks that are impossible to detect. Yellow (24) represents attacks that are difficult
to detect. (Color figure online)
The model was not used with process 6 as it has only two tags. The SWaT
dataset presents only one attack3 and has an impact on process 6.
3
Attack 23 also has an impact on process 3. The developed model for process 3
detected this attack.
12 J. Kim et al.
Attack 4. This attack opens MV-504 that does not exist in the dataset. The
description [6] in SWaT’s list of attacks indicates that this attack has no impact.
Attack 14. According to SUTD, the attack 14 failed because tank 301 was
already full [6].
Attack 29. SUTD said that P-201, P-203, and P-205 did not start because of
mechanical interlocks. In the dataset, nothing was changed around 2015/12/31
at 3:32:00 PM.
Attack 13. This attack attempted to close MV-304 but MV-304 was closed later
than when this attack occurred. In the SWaT dataset, MV-304 did not change.
Attack 19. This attack attempted to set value of AIT-504 to 16 µs/cm. AIT-
504 is below 15 µs/cm in the normal state. In Fig. 6, the attack appeared to be
detected, but the high rate was derived from attack 20 which set a value of AIT-
504 to 255 µs/cm. The distances provided by the developed model are too short
to detect for attack 19. The rate S was of approximately 20% for attack 19.
Attack 24. This attack attempted to turn off P-203 and P-205 (both are pumps
are used for injecting chemicals). However, there was only a small impact due
to the closure of P-101.
Attack 25. This attack attempted to set the value of LIT-401 to 1,000 and open
P-402 while P-402 was still operating. LIT-401 presented notable distances but
the rate was of approximately 20% (Fig. 7). According to [18], attack 25 was
detected but wrong attack points were determined.
4
Their intervals are 0, 42, and 1 s respectively, while the developed model uses a
100-second sliding windows.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 13
Fig. 6. Attack 19 in process 5 (red: attack, purple: prediction error, and blue: anomaly
score) (Color figure online)
Fig. 7. Attack 25 in process 4 (red: attack, purple: prediction error, and blue: anomaly
score) (Color figure online)
Attack 35. According to [6], attack 35 occurred at process 1, but herein, the
attack was detected at process 3. Before the attack, P-101 was open, and P-202
was close. The attack opened P-101 and kept P-102 close. Figure 8 shows the
attacks 34 and 35. Processes 1, 2, and 3 presented remarkable distances. The
distance of process 2 appears to be large, but the scale is small. There are two
high peaks of rate in process 1, but they come from the attack 34.
Attack 34 closed P-101 (2 → 1) and opened P-102 (1 → 2). Later, P-101
was opened (1 → 2) at 17:14:59 and P-102 was closed (2 → 1). Attack 35 was
different from attack 34 as it kept P-102 closed. In the training (normal) dataset,
P-102 was closed (of value 1) at all time, which is why the developed model
always indicated 1. Because P-102 is a backup pump for P-101, the developed
model must understand their connection. However, the model could not learn
the connection as the training dataset did not give enough information.
14 J. Kim et al.
Fig. 8. Attacks 34 and 35 in processes 1, 2, and 3 (red: attack, purple: prediction error,
and blue: anomaly score) (Color figure online)
Table 4. Number of false positives. OP indicates the number of false positives for
attacks on other processes, LT indicates the long-tailed detection (over 15 min), and
TFP indicates the true false positives.
as a true-positive, but sometimes the tail was too long and the attack lasted over
15 min. The long-tail positives were not counted as true-positives. In summary,
there was one true false positive (TFP, the fifth column) in process 1 and five
in process 2.
Table 5. The analysis of false positives in process 1. OP means the number of false
positives for attacks on other processes; LT for long-tailed detection (over 15 min); TFP
for true false positives.
related attacks. Therefore, there were five independent false positives, and four
of them were true.
Most false positives occurred at P-201, P-205, and MV-201. P-201 is the NaCl
injection jump; P-205 is the NaOCl injection pump. In the training dataset, P-
201 never changed. We guess that the model regarded any change of P-201 as
an attack. AIT-201, a sensor for NaCl, caused a false positive (No. 8 in Table 6)
because P-201 had changed the level of NaCl.
After the last false positive, the prediction errors of AIT-201 and AIT-203
went high. AIT-201 and AIT-203 are sensors for NaCl and NaOCl, respectively.
We guess that the unexpected behaviors of P-201 and P-205 led to new but
normal dataset.
False Positives in Process 4. In Table 7, there was one false positive in process
4, which came from attack 37 hitting process 5.
Anomaly Detection for ICSs Using Seq2Seq Neural Networks 17
5 Conclusion
It is difficult to get internal specification and control logic of ICSs. If routine
ICS operational data is the only information available, a data-driven approach
is a proper way to develop security products.
We proposed an anomaly detection method for industrial control systems
using sequence-to-sequence neural networks with attention. Due to the difficulty
of defining the abnormal state, the model learns the normal dataset in an unsu-
pervised way. In the detection phase, the model predicts future values based on
the previously observed ones. The difference between the model’s prediction and
the measured value is the key criterion to detect anomalies.
The alarm decision depends on the threshold, and heuristic hyper-parameters
were necessary for this experiment. Our proposed method is not dedicated to the
dataset. It can be generalized to train any ICS datasets and extract the decision
grounds because the specification of operational data and control logic inside
was not required. It is also able to detect anomalies with only the dataset from
normal operations.
References
1. Ahmad, S., Lavin, A., Purdy, S., Agha, Z.: Unsupervised real-time anomaly detec-
tion for streaming data. Neurocomputing 262, 134–147 (2017)
2. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning
to align and translate. In: 3rd International Conference on Learning Representa-
tions, ICLR 2015, San Diego, CA, USA, 7–9 May 2015 (2015). Conference Track
Proceedings
3. Chen, Y., Poskitt, C.M., Sun, J.: Learning from mutants: using code mutation to
learn and monitor invariants of a cyber-physical system. In: 2018 IEEE Symposium
on Security and Privacy (SP), pp. 648–660. IEEE (2018)
4. Choi, S., Yun, J.-H., Kim, S.-K.: A comparison of ICS datasets for security research
based on attack paths. In: Luiijf, E., Žutautaitė, I., Hämmerli, B.M. (eds.) CRITIS
2018. LNCS, vol. 11260, pp. 154–166. Springer, Cham (2019). https://doi.org/10.
1007/978-3-030-05849-4 12
5. Formby, D., Srinivasan, P., Leonard, A., Rogers, J., Beyah, R.A.: Who’s in con-
trol of your control system? device fingerprinting for cyber-physical systems. In:
Network and Distributed Systems Security (NDSS) (2016)
18 J. Kim et al.
6. Goh, J., Adepu, S., Junejo, K.N., Mathur, A.: A dataset to support research in
the design of secure water treatment systems. In: Havarneanu, G., Setola, R., Nas-
sopoulos, H., Wolthusen, S. (eds.) CRITIS 2016. LNCS, vol. 10242, pp. 88–99.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-71368-7 8
7. Goh, J., Adepu, S., Tan, M., Lee, Z.S.: Anomaly detection in cyber physical systems
using recurrent neural networks. In: 2017 IEEE 18th International Symposium on
High Assurance Systems Engineering (HASE), pp. 140–145. IEEE (2017)
8. Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8),
1735–1780 (1997)
9. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. In: 3rd Interna-
tional Conference on Learning Representations, ICLR 2015, San Diego, CA, USA,
7–9 May 2015 (2015). http://arxiv.org/abs/1412.6980. Conference Track Proceed-
ings
10. Kravchik, M., Shabtai, A.: Detecting cyber attacks in industrial control systems
using convolutional neural networks. In: Proceedings of the 2018 Workshop on
Cyber-Physical Systems Security and Privacy CPS-SPC 2018, pp. 72–83 (2018)
11. Lemay, A., Fernandez, J.M.: Providing SCADA network data sets for intrusion
detection research. In: Proceedings of the 9th USENIX Conference on Cyber Secu-
rity Experimentation and Test, CSET 2016, Berkeley, CA, USA, p. 6 (2016)
12. Li, D., Chen, D., Shi, L., Jin, B., Goh, J., Ng, S.: MAD-GAN: multivariate
anomaly detection for time series data with generative adversarial networks. CoRR
abs/1901.04997 (2019)
13. Lin, C.Y., Nadjm-Tehrani, S., Asplund, M.: Timing-based anomaly detection
in scada networks. In: D’Agostino, G., Scala, A. (eds.) CRITIS 2017. LNCS,
vol. 10707, pp. 48–59. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-
99843-5 5
14. Mitchell, R., Chen, I.R.: A survey of intrusion detection techniques for cyber-
physical systems. ACM Comput. Surv. (CSUR) 46(4), 55 (2014)
15. Mitchell, R., Chen, R.: Behavior-rule based intrusion detection systems for safety
critical smart grid applications. IEEE Trans. Smart Grid 4(3), 1254–1263 (2013)
16. Reddi, S.J., Kale, S., Kumar, S.: On the convergence of adam and beyond. In: 6th
International Conference on Learning Representations, ICLR 2018, Vancouver, BC,
Canada, April 30–May 3 2018 (2018). Conference Track Proceedings
17. Rodofile, N.R., Schmidt, T., Sherry, S.T., Djamaludin, C., Radke, K., Foo, E.: Pro-
cess control cyber-attacks and labelled datasets on S7Comm critical infrastructure.
In: Pieprzyk, J., Suriadi, S. (eds.) ACISP 2017. LNCS, vol. 10343, pp. 452–459.
Springer, Cham (2017). https://doi.org/10.1007/978-3-319-59870-3 30
18. Shalyga, D., Filonov, P., Lavrentyev, A.: Anomaly detection for water treatment
system based on neural network with automatic architecture optimization. In:
DISE1 Workshop, International Conference on Machine Learning (ICML) (2018)
19. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural
networks. In: Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Wein-
berger, K.Q. (eds.) Advances in Neural Information Processing Systems, vol. 27,
pp. 3104–3112 (2014)
20. Yun, J.-H., Hwang, Y., Lee, W., Ahn, H.-K., Kim, S.-K.: Statistical similarity
of critical infrastructure network traffic based on nearest neighbor distances. In:
Bailey, M., Holz, T., Stamatogiannakis, M., Ioannidis, S. (eds.) RAID 2018. LNCS,
vol. 11050, pp. 577–599. Springer, Cham (2018). https://doi.org/10.1007/978-3-
030-00470-5 27
Reflective Attenuation
of Cyber-Physical Attacks
1 Introduction
Cyber-physical systems (CPS) are modern control systems used to manage and
control critical infrastructures [7]. Physical properties of such infrastructures are
modeled via control-theoretic tools, e.g., control-loops and feedback controllers
[16]. Feedback controllers have to be able to manage the behavior of the CPS, by
confirming that the commands are executed correctly and the information com-
ing from the physical states is consistent with the predicted behavior [26]. Feed-
back controllers are also used to compute corrective actions, e.g., by minimizing
the deviation between a reference signal and the system output measurements.
A CPS is composed of three main layers: (1) the physical layer, which involves
the physical process (monitored and controlled by physical sensors and physical
actuators); (2) the control layer, which is in charge of regulating the operation
of the physical process via control commands; and (3) the cyber layer, which
is responsible for monitoring operation and supervision tasks. These three lay-
ers are interconnected using a communication network. In other words, the CPS
can be modeled as a networked-control system [21]. The interconnection between
information and operational systems leads to new security threats [20,21]. Tra-
ditional cyber attacks are well know and countermeasures have been studied.
However, launching a cyber-physical attack requires a different knowledge from
the one used in traditional cyber security and different protection techniques are
also required.
A cyber-physical attack causes tangible damage to physical components, for
instance, adding disturbances to a physical process via exploitation of vulnerabil-
ities in computing and networking resources of the systems (i.e., the components
at the cyber layer). However, to achieve just a cyber attack, the adversary may
be able to inject any input in the system but this does not necessarily mean to
be able to influence the processes in the physical world. The processes and their
dynamics have to be properly understood to cause a real damage [12].
A physical process has automatic safety measures and operational con-
straints, e.g., to disable a physical process when certain dangerous conditions
are met. For instance, to properly react when a physical component fails. For
this reason, an adversary who aims at damaging the physical process needs to
understand how the dynamics of the physical plant works. This means that com-
promising and disrupting a device or communication channel used to sense or
control a physical system is a necessary requirement to perform cyber-physical
attacks. However, the damage can be limited if the adversary succeeds at affect-
ing the cyber layer, but remains unable to manipulating the control system (i.e.,
fails at perturbing the physical process). To achieve the desired impact and
achieve a cyber-physical attack, the adversary needs to assess how the attack
will perform at the control level. Therefore, to achieve a cyber-physical attack,
the first step is to hack the cyber layer, to obtain a remote access within the
target system. Then, the second step is to learn about the physical process and
how the control layer works in order to manipulate the physical layer and cause
a damage to physical components. Adversaries need to know how the physical
process is controlled, failure conditions of the equipment, process behavior and
signal processing [20,21].
In this paper, we propose a technique to attenuate cyber physical attacks
that uses programmable reflection and programmable networks to sanitize the
malicious actions introduced by some cyber-physical injection attack such as
false data injection, bias injection, replay attack, command injection and cover
attack [27]. The adversary uses the network to manipulate the process through
the modification of specific payloads. Then, the proposed technique uses the
network to neutralize the attack effects. This concept relies on the use of pro-
grammable reflection, which is a meta programming technique that has the
potential to allow a programmable system manipulate itself at runtime and the
use of programmable networks to sanitize the traffic.
The main contributions of the paper are summarized as follows: (1) we pro-
pose a technique to handle cyber-physical injection attacks; (2) we revisit the
Reflective Techniques to Attenuate Cyber-Physical Attacks 21
2 Related Work
We survey next some related work, structured in terms of attack tolerance, pro-
grammable networking and programmable reflection.
3 Preliminaries
We provide in this section some initial preliminaries about our assumptions in
terms of system and adversarial models.
uk yk uk yk uk yk
Plant Plant Plant
−uk + uk
monitoring
uk Reflective agent
uk Adversary Adversary
uk
−yk + yk
uk yk uk yk uk yk
Feedback controller Feedback controller Feedback controller
Fig. 1. Feedback control view. (a) Normal operation mode. (b) System under attack.
(c) Attack attenuation.
We also assume that the adversary performs its malicious actions in the
cyber-physical system, i.e., at the data layer of the network domain. This means
that the adversary is not attacking the programmable network itself, e.g, the con-
trol layer. We focus on adversaries that use the network to damage the system.
Adversaries that may compromise the physical nodes themselves, to damage the
system, are out of the scope of this paper due to this kind of systems usually
have good physical protection mechanism implemented.
4 Our Approach
decreasing the chances of the adversary to correlate the proper input and
output values.
– Phase 2 – Reflection. The feedback controller creates a reflective agent,
which gets executed within the domain of the network controller. The reflec-
tive agent has the control capabilities associated to the victims of the attack.
It uses programmable reflection to create, at runtime, a component that exe-
cutes the same program and equivalent interfaces as the feedback controller.
By programmable networking reflection, we refer to the system capability
of modifying its networking behavior, i.e., changing accordingly to what is
required. For this reason, an on-demand process for loading and unloading
components as services could be performed.
– Phase 3 – Traffic sanitization. The forwarding elements using network
programming capabilities allow to perform a dynamic network traffic san-
itizing by modifying the packet containing malicious payloads. The packet
affected by the adversary gets sanitized by the reflective agent, which deter-
mines what is the correct payload the packet should have. All the network
actions required to sanitize the traffic are coordinated by the network con-
troller.
5 Experimental Results
Programming Interfaces) over the INET [30] and SCADASim [18] libraries, to
enforce the use of the Modbus protocol over TCP and UDP traffic. All the com-
ponents (both in the Lego SCADA testbed and the OMNeT++ co-simulation)
are synchronized by feedback controllers. Every motion device has a distance
sensor in the frontal part, to measure its relative distance to the boundaries of
a unit square area. The distance is transmitted to the feedback controllers via
Modbus SCADA messages. The feedback controller computes the relative veloc-
ity of each motion device, and the Euclidean distance between the two motion
devices, in order to guarantee spatial collision-free operations.
The goal of the adversary is to launch an attack at the control level to move
the physical process to an undesirable state resulting in the physical collision
of the two motion devices. Figure 3(a–b) show the kinetic dynamics of the sys-
tem during the nominal case (i.e., absence of attacks, left-side); and during the
attack (i.e., the moment at which the adversary takes control over the system,
right-side). Time is normalized between 0.0 and 1.0, representing the temporal
percentage of multiple experimental runs. We can appreciate how the system
moves to unstable states, disrupted by the adversary. Some live demonstration
videos showing the spatial collision that cause the disruption represented in Fig. 3
are available at http://j.mp/legoscada.
(a) Lego Mindstorms testbed for the genera- (b) OMNeT++ CPS co-simulation, using the
tion of SCADA data. generated data.
Fig. 2. (a) Lego testbed for the generation of SCADA-driven CPS data (cf.
http://j.mp/legoscada for live demonstration videos and source code). (b) CPS co-
simulation implemented over OMNeT++ (cf. http://j.mp/omnetcps for live demon-
stration videocaptures and source code).
1 1
0.5 0.5
Value of states
Value of states
0 0
0.1 0.2 0.3 0.4 0.5 0.6 Time 0.1 0.2 0.3 0.4 0.5 0.6 Time
-0.5 -0.5
-1 -1
(a) Temporal representation of the CPS kinetic dynamics, associated to the two
motion devices (left-side, nominal mode dynamics; right-side, dynamics during
the attack).
90 1 90 1
120 60 120 60
0.8 0.8
0.6 0.6
150 30 150 30
0.4 0.4
0.2 0.2
180 0 180 0
Fig. 3. Lego testbed results (systems dynamics during nominal and attack modes).
(a) Temporal representation (the dotted line represents the moment when an attack
starts). (b) Winding graph representation of the nominal and attack modes.
that perform the commands (accelerators that increase or decrease the relative
velocity of the two motion devices).
The adversary starts the cyber-physical attack by either tampering the con-
troller with fake sensor readings or modifying the control commands sent from
the controller. With the OMNeT++ co-simulation, we evaluate the attenuation
of the bias injection attack, i.e., by forging tampered control commands from
the controller to the plant. For simplicity reasons, we focus only on the physical
part of the cyber-physical attack using the network to damage the system. In
other words, we assume an adversary that already found a way to hack the cyber
layer and gain remote access to the system.
Each co-simulation evaluates fifty Monte Carlo different runs. In addition,
according to the sensor specification, the simulation considers a possible error of
up to 1 cm w.r.t. the measured distance value. We also model the network delays
using the probability distribution in [5]. Figure 4(a) shows the results obtained
for the nominal case (i.e., absence of attack), considering the aforementioned
30 M. Segovia et al.
possible variation. The plots depict the average Euclidean distance, with 95%
confidence intervals, between the motion devices in function of time. The hori-
zontal axis of the plots in Fig. 4(a–d) provides a normalized time between 0.0 and
1.0, representing the temporal percentage prior concluding the simulation runs.
The vertical axis of the plots in Fig. 4(a–d) provides the Euclidean distance
between the two motion devices, from 0 to 1400 cm. Some further evaluation
details are discussed below.
Discussion—During the perpetration of the attacks, the adversary performs a
bias injection of cyber-physical data. The adversary uses the network to modify
the exchanged packets between the feedback controller and the plant. We assume
an adversary recording and learning the system dynamics from commands and
sensor outputs. The adversary performs an initial learning phase, in order to
eavesdrop data and infer the system dynamics, i.e., the same one used by the
feedback controller to guarantee the stability of the system, shown as nominal
case in Fig. 4(a).
Let uk be a feedback controller command sent to the actuator of a motion
device at time k. Let uact
k be the command received by the actuator at time k,
where 0 ≤ k ≤ Ts and Ts be the full duration of each simulation run. The attack
interval Ta is limited to the simulation time Ts , as summarized next:
uk if k ∈/ Ta
uact =
k uk if k ∈ Ta
For our evaluation, we compare two type of adversaries according to the bias
injected in the payload of the packets, i.e, according to the difference between
the value uk injected by the attacker and the real value uk sent by the controller.
This way, we define two adversary models: an aggressive adversary and a non
aggressive adversary. The aggressive adversary injects in uk a bigger difference
with respect to the correct command uk sent by the feedback controller compared
to the non aggressive adversary. In consequence, an aggressive adversary will
make the system move faster from its nominal state. Figure 4(b) shows the results
obtained for the two attack scenarios. The feedback controller loses its control
over the system, while the adversary forces the spatial collision of the two motion
devices.
During the attenuation process, the system reacts using reflective pro-
grammable networking. The reflective agent takes control of the situation, after
a hangover of the feedback controller functionality (which moves to the pro-
grammable controller domain). This reflective agent takes control over the adver-
sary communications and neutralizes the attack. For each of the defined adver-
saries, we simulate two scenarios using different values for the time the solution
starts working. This is a parameter of the simulation that depends mainly on
the time required for the detection mechanism to detect the attack plus the time
required to set up and coordinate all the components working in the approach.
Figure 4(c)–(d) show how the approach guarantees the controllability property.
The first vertical dotted line shows the moment when the attack starts and the
second vertical dotted line shows the moment when the technique starts. It is
Reflective Techniques to Attenuate Cyber-Physical Attacks 31
1400
1400
Nominal Case Aggressive Adversary
Non aggressive Advesary
1200
1200
1000 1000
Euclidean distance
Euclidean distance
800 800
600 600
400 400
200 200
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Time Time
1000 1000
Euclidean distance
Euclidean distance
800 800
600 600
400 400
200 200
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Time Time
0.6
150 0.6 30 150 30
0.4
0.4
0.2
0.2
180 0 0 180 0 0
270 270
(e) Winding graph of dynamics cycles (f) Winding graph of dynamics cycles
(attenuation 1) (attenuation 2)
Fig. 4. OMNeT++ results. (a–b) Euclidean distance (with 95% confidence intervals),
nominal and attack simulations. (c–d) Euclidean distance, attenuation of two different
remediation starting time models. (e–f) Winding graphs, same attenuation models.
32 M. Segovia et al.
6 Conclusion
Acknowledgements. The authors acknowledge support from the Cyber CNI chair
of the Institut Mines-Télécom. The chair is supported by Airbus Defence and Space,
Amossys, EDF, Nokia, BNP Paribas and the Regional Council of Brittany. The chair
has been acknowledged by the Center of excellence in Cybersecurity. Authors acknowl-
edge as well support from the European Commission (H2020 SPARTA project), under
grant agreement 830892.
References
1. Albert, R., Jeong, H., Barabási, A.-L.: Error and attack tolerance of complex net-
works. Nature 406(6794), 378 (2000)
2. Campbell, A.T., Katzela, I., Miki, K., Vicente, J.: Open signaling for ATM, internet
and mobile networks (OPENSIG’98). SIGCOMM Comput. Commun. Rev. 29(1),
97–108 (1999)
3. Cavalli, A.R., Ortiz, A.M., Ouffoué, G., Sanchez, C.A., Zaı̈di, F.: Design of a secure
shield for internet and web-based services using software reflection. In: Jin, H.,
Wang, Q., Zhang, L.-J. (eds.) ICWS 2018. LNCS, vol. 10966, pp. 472–486. Springer,
Cham (2018). https://doi.org/10.1007/978-3-319-94289-6 30
4. Coulson, G., et al.: Reflective middleware-based programmable networking. In: The
2nd International Workshop on Reflective and Adaptive Middleware, pp. 115–119
(2003)
5. Elteto, T., Molnar, S.: On the distribution of round-trip delays in TCP/IP net-
works, pp. 172–181, November 1999
6. Enns, R., Bjorklund, M., Schoenwaelder, J., Bierman, A.: Network configuration
protocol (NETCONF) - internet engineering task force, RFC 6241, June 2011.
http://www.ietf.org/rfc/rfc6241.txt
7. European Union Agency for Network and Information Security Agency (ENISA).
Methodologies for the identification of Critical Information Infrastructure assets
and services (2015). https://fullreportatwww.enisa.europa.eu/
8. Hachem, N., Debar, H., Garcia-Alfaro, J.: HADEGA: a novel MPLS-based mitiga-
tion solution to handle network attacks. In: 31st IEEE International Performance
Computing and Communications Conference, IPCCC 2012, Austin, TX, USA, 1–3
December 2012, pp. 171–180 (2012)
9. Jajodia, S., Ghosh, A.K., Swarup, V., Wang, C., Sean Wang, X.: Moving Target
Defense: Creating Asymmetric Uncertainty For Cyber Threats, vol. 54. Springer,
New York (2011)
10. Joolia, A., Coulson, G., Blair, G., Gomes, A.T., Lee, K., Ueyama, J.: Flexible
programmable networking: a reflective, component-based approach (2003)
11. Kreutz, D., Ramos, F.M.V., Verissimo, P.E., Rothenberg, C.E., Azodolmolky,
S., Uhlig, S.: Software-defined networking: a comprehensive survey. Proc. IEEE
103(1), 14–76 (2015)
12. Krotofil, M., Larsen, J.: Rocking the pocket book: hacking chemical plants for
competition and extortion. DEF CON, 23 (2015)
34 M. Segovia et al.
13. Kuipers, D., Fabro, M.: Control systems cyber security: defense in depth strategies.
Technical report, Idaho National Laboratory (INL) (2006)
14. Lagu, S.S., Deshmukh, S.B.: Raspberry Pi for automation of water treatment
plant. In: 2015 International Conference on Computing Communication Control
and Automation (ICCUBEA), pp. 532–536, February 2015
15. Modbus Organization. Official Modbus Specifications (2016). http://www.modbus.
org/specs.php. Accessed Apr 2019
16. Ogata, K., Yang, Y.: Modern Control Engineering, vol. 4. Prentice-Hall, Upper
Saddle River (2002)
17. Piedrahita, A.F.M., Gaur, V., Giraldo, J., Cardenas, A.A., Rueda, S.J.: Virtual
incident response functions in control systems. Comput. Netw. 135, 147–159 (2018)
18. Queiroz, C., Mahmood, A., Tari, Z.: SCADAsim–a framework for building SCADA
simulations. IEEE Trans. Smart Grid 2(4), 589–597 (2011)
19. Rollins, M.: Beginning LEGO MINDSTORMS EV3. Apress, New York (2014)
20. Rubio-Hernan, J., De Cicco, L., Garcia-Alfaro, J.: Event-triggered watermarking
control to handle cyber-physical integrity attacks. In: Brumley, B.B., Röning, J.
(eds.) NordSec 2016. LNCS, vol. 10014, pp. 3–19. Springer, Cham (2016). https://
doi.org/10.1007/978-3-319-47560-8 1
21. Rubio-Hernan, J., De Cicco, L., Garcia-Alfaro, J.: Revisiting a watermark-based
detection scheme to handle cyber-physical attacks. In: 2016 11th International
Conference on Availability, Reliability and Security (ARES), pp. 21–28. IEEE,
August 2016
22. Rubio-Hernan, J., Rodolfo-Mejias, J., Garcia-Alfaro, J.: Security of cyber-physical
systems. In: Cuppens-Boulahia, N., Lambrinoudakis, C., Cuppens, F., Katsikas,
S. (eds.) CyberICPS 2016. LNCS, vol. 10166, pp. 3–18. Springer, Cham (2017).
https://doi.org/10.1007/978-3-319-61437-3 1
23. Rubio-Hernan, J., Sahay, R., De Cicco, L., Garcia-Alfaro, J.: Cyber-physical archi-
tecture assisted by programmable networking. Internet Technol. Lett. 1, e44 (2018)
24. Sahay, R., Blanc, G., Zhang, Z., Debar, H.: Towards autonomic DDoS mitigation
using software defined networking. In: SENT 2015: NDSS Workshop on Security of
Emerging Networking Technologies, San Diego, CA, USA. Internet society, Febru-
ary 2015
25. Segovia, M., Cavalli, A.R., Cuppens, N., Garcia-Alfaro, J.: A study on mitigation
techniques for SCADA-driven cyber-physical systems (position paper). In: Zincir-
Heywood, N., Bonfante, G., Debbabi, M., Garcia-Alfaro, J. (eds.) FPS 2018. LNCS,
vol. 11358, pp. 257–264. Springer, Cham (2019). https://doi.org/10.1007/978-3-
030-18419-3 17
26. Soupionis, Y., Ntalampiras, S., Giannopoulos, G.: Faults and cyber attacks detec-
tion in critical infrastructures. In: Panayiotou, C.G.G., Ellinas, G., Kyriakides, E.,
Polycarpou, M.M.M. (eds.) CRITIS 2014. LNCS, vol. 8985, pp. 283–289. Springer,
Cham (2016). https://doi.org/10.1007/978-3-319-31664-2 29
27. Teixeira, A., Shames, I., Sandberg, H., Johansson, K.H.: A secure control frame-
work for resource-limited adversaries. Automatica 51, 135–148 (2015)
28. Tennenhouse, D.L., Smith, J.M., Sincoskie, W.D., Wetherall, D.J., Minden, G.J.:
A survey of active network research. Comm. Mag. 35(1), 80–86 (1997)
29. The OMNeT++ network simulation framework. http://www.omnetpp.org/.
Accessed Apr 2019
30. The OMNeT++/INET framework. http://inet.omnetpp.org/. Accessed Apr 2019
31. Varga, A., Hornig, R.: An overview of the OMNeT++ simulation environment. In:
1st International Conference on Simulation Tools and Techniques for Communica-
tions, Networks and Systems & Workshops (Simutools) (2008)
Distributed UCON in CoAP and MQTT
Protocols
1 Introduction
This work has been partially funded by EU Funded projects H2020 NeCS, GA #675320,
H2020 C3ISP, GA #700294 and EIT Digital HC&IoT.
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 35–52, 2020.
https://doi.org/10.1007/978-3-030-42048-2_3
36 A. Rizos et al.
expected to reach 50 billion [6]. The IoT is an inherently complex ecosystem, due
to the many different architectures and protocols used, and the different types of
hardware and software present in IoT devices. Besides, there are a lot of imma-
ture products and cheap devices being released in the market. As a result, sev-
eral security issues have been identified, such as interception of communications,
data compromise by unauthorized parties to collect Personal Identifiable Infor-
mation (PII), authentication can be brute-forced, credentials can be extracted
from device firmware, mobile apps or intercepted at login, and new firmware can
be uploaded with malware [2]. Thus, it is of paramount importance to provide
techniques that allow secure information exchange between IoT devices.
When it comes to application layer protocols, two of the most widely used are
the Constrained Application Protocol (CoAP) and the Message Queue Telemetry
Transport (MQTT) [1]. Both protocols were designed for IoT and Machine-to-
Machine (M2M) applications. On one hand, CoAP was introduced as a standard
by the Internet Engineering Task Force (IETF) in 2014 [25] and its key features
are simplicity for constrained environments, very low overhead, and easy inter-
operability with Hypertext Transfer Protocol (HTTP). Hence, CoAP uses a
Client/Server communication pattern, in which Servers make resources avail-
able and Clients interact with resources using REpresentational State Transfer
(REST) methods. When a Client requests access to a certain piece of informa-
tion, the access granting decisions are based on the optional use of the Datagram
Transport Layer Security (DTLS) library, which provides the security-oriented
features of CoAP. On the other hand, MQTT was recently standardized by
OASIS1 and follows the Publish/Subscribe communication pattern. In this pat-
tern, a central entity called Broker handles the communication by collecting
data, organized by topics, from Publishers and distributing it to the set of valid
Subscribers. According to [1], the MQTT standard provides support for Trans-
port Layer Security (TLS)/Secure Socket Layer (SSL) protocols but offers only a
basic authentication feature by default, applied to Publishers/Subscribers during
registration. Most of the security solutions developed for MQTT are application-
specific or try to leverage TLS/SSL in some way.
In this paper, we propose a security enhancement of both CoAP and MQTT
protocols by adding strong authorization capabilities provided by the Dis-
tributed Usage Control (UCON). UCON enhances Attribute-Based Access Con-
trol (ABAC) models by providing two novelties: Continuity of control and muta-
bility of attributes [20]. Hence, it provides the ability to evaluate policies that
have attributes which may change their values during a session. For securing
CoAP and MQTT simultaneously we used the distributed version of UCON in
[19]. With this solution, we aim to provide dynamic policy enforcement towards
more secure data distribution on both protocols at the same time. Moreover,
we enhance the interoperability between the protocols, by sharing attribute val-
ues via UCON. This work presents the implementation of UCON on top of
CoAP and MQTT. We highlight that the presence of UCON does not impact
1
https://www.oasis-open.org/.
Distributed UCON in CoAP and MQTT Protocols 37
how the protocols work, which means any CoAP and/or MQTT application
can support it. To further our approach, we measure the performance of a real
implementation.
The rest of the paper is organized as follows: In Sect. 2 we report some
background information on IoT protocols, detailing also UCON and its dis-
tributed version. Section 3 details the proposed architecture and operative work-
flow. Section 4 details our implementation and discusses the results of the per-
formance analysis. In Sect. 5 we present a set of related works about security in
CoAP and MQTT, and applications of UCON in IoT. Finally, Sect. 6 we offer
some conclusions and hint at future directions.
2 Background
In this section we present the main protocols for IoT, motivating the choice to
focus on CoAP and MQTT. We also introduce the most important concepts of
UCON.
updates. This operation is triggered by sending a request to the Server with the
Observe flag off or by replying to a resource update with a reset message to the
Server. Figure 1 presents the CoAP architecture, showing a Server interacting
with multiple clients and, on the right side, a Client interacting with other
Server.
Conversely to HTTP, CoAP uses the User Datagram Protocol (UDP) instead
of the Transmission Control Protocol (TCP) for message exchanging. In essence,
this means that reliability for message exchange is not guaranteed as UDP
doesn’t support mechanisms for reliable communications, focusing instead on
fast and simple message exchanging. As a result, CoAP implements two request
types: Confirmable (CON) and Non-Confirmable (NON). On one hand, CON
requests achieve reliable communications by expecting an acknowledgement
(ack) message from the Server in response to each request. On the other hand,
NON requests are “fire and forget” messages that don’t expect any confirmation
that the request was indeed received by the Server. As a result to the previous,
CoAP provides two Quality of Service (QoS) levels: at least once (using CON
requests), and At most once (using NON requests). Because CoAP uses UDP
it supports multicast requests, where one Client can send the same request to
multiple Servers at the same time. However, a CoAP Server always replies in
unicast to a multicast request. The CoAP protocol by itself does not provide
any security features (e.g. authentication), so security is not designed in the
context of the application layer but instead, it is optionally supported at the
transport layer [11]. CoAP supports the DTLS protocol [22]. DTLS provides
three modes for secure message exchanging: PreSharedKey, RawPublicKey and
Certificate [9]. These modes allow for strong authentication and data encryption
and integrity in transit. However, they don’t provide refined authorization capa-
bilities. This work focuses on the CoAP protocol since it is the most constrained
one and widely used in the area of IoT. Existing security features of CoAP pro-
tocol rely only on the use of DTLS, which does not deal with what happens
after the information is shared to the Client and does not provide a continuous
Distributed UCON in CoAP and MQTT Protocols 39
(C, Java, PHP, Python, Ruby, JavaScript) and for the two major mobile plat-
forms (iOS and Android) [6]. The only security feature MQTT provides is Pub-
lisher /Subscriber authentication with the Broker. They can provide a Username
and Password during registration, however this information is optional and it is
also transmitted in plain text so in reality no security is provided [18]. A well
known effort to add more security to MQTT is proposed in [26], but this solu-
tion makes changes to the protocol itself, breaking compatibility as a result, and
also does not address changes in Publisher /Subscriber which might violate any
existing policies. Our proposal addresses this problem by providing continuous
control of Publishers/Subscribers during both authentication and access.
action is received by the UCS, it deletes the session details from the SM and
communicates to the PIPs to unsubscribe the attributes related to that session,
unless other sessions are using them.
UCON in MQTT. For the UCON in MQTT part, our work is based on the
existing work that the authors of [14] did. In that setup, an instance of the
UCS is installed on the Broker of every MQTT system controlling the access
of Subscribers to the data coming from the Publishers over specific Topics. The
difference between that work and the work presented in this paper is that due
to the use of the distributed model of UCON, in our setup the UCS instances
that are installed in MQTT systems can share their attributes or retrieve the
values of remote attributes from other systems.
3.1 Architecture
For example, we can have the systems (Ci, i ∈ N>0 ) that use the CoAP protocol
and the systems (Mj, j ∈ N>0 ) that use the MQTT protocol. For the sake of
simplicity, in Fig. 4 we present only a couple of subsystems for each protocol.
Because of the different architecture of each IoT protocol, the PEP component
of UCON has to be specifically adjusted to each protocol, whereas the UCS
does not change but can be executed in different components according to the
protocol used. Every system has their own local attributes. In the meantime, each
system (Ci/Mj) for their local evaluation of access requests may need access to
the values of remote attributes that belong to another subsystem (Cx,Mx). This
communication between different instances of UCS for sharing attribute values
gives us the ability to control different protocols without the necessity of the
sensors that provide the attributes to use the same protocol for communication.
At this point, it is worth re-iterating that the use and functionality of both
protocols works as standard despite the addition of UCON.
CoAP protocol follows the Client/Server model. Each Client can either request
and/or provide data to the CoAP system. Clients could be very constrained
devices whereas the Server has to provide enough computational power to sup-
port all the communications and actions of the CoAP system. The registration
and management of the Clients is performed by the Server that assigns unique
tokens to each Client. Yet, this model is very simple and does not provide any
mechanisms for checking the access during time. All the Clients remain con-
nected as soon as they firstly register correctly. In reality, there might occur
cases in which sensitive information should not be delivered to specific Clients.
To achieve such continuous monitoring of the access rights of the Clients, the
CoAP architecture has been enhanced by the addition of UCON.
44 A. Rizos et al.
and return it to the CH L (Task 9). Then all the above are sent to the local PDP
(PDP L) to be evaluated (Task 10) and the PDP L replies with the result to the
CH L (Task 11). Considering the result is either Permit or Deny, the request is
approved or not accordingly and the CLIENT R is informed about this (Tasks
12–14).
Then, in the case of a Permit in the previous request, the CLIENT R per-
forms another request to obtain data, which in our case will be provided by the
CLIENT P (Task 15), after the Server communicates with PEP L (Task 16)
and gets a Permit on that request. For this to happen, PEP L must perform the
StartAccess to the CH L which again passes through evaluation by the PDP L as
previously (Tasks 4–11). Supposing that CLIENT P starts sending data to the
Server (Task 20) that must be delivered to CLIENT R, the Server distributes
them without any interference by UCON (Task 21). In the meantime, UCS L is
performing a continuous re-evaluation of the attributes (Task 22). In the case
there is a policy violation, the access of CLIENT R should be revoked (Task
23). The CH L informs the glspep L (Task 24) and the latter informs the Server
(Task 25). The Server deletes CLIENT R from its authorized clients and informs
it about this fact (Task 26).
installed (UCSm and UCSc respectively). Figure 7 describes the topology of our
use case scenario.
In this scenario a smart vacuum cleaner wants to operate. The policy states
that it is allowed to operate only if the power consumption and noise levels are
within a certain threshold. Some appliances like a smart TV, a smart meter (that
measures power consumption) and a smart speaker together with the vacuum
cleaner belong to the same system, communicating via MQTT protocol (System
M). Furthermore, we consider a smart washing machine, a smart thermostat
and a smart noise sensor that are also inside the same home but communicate
via CoAP protocol and they belong to a CoAP System (System C). In order
for the vacuum cleaner (Subscriber ) to start cleaning, it must first connect to
the Broker of system M to subscribe to the power consumption values which
are gathered and stored by the smart meter (Publisher ). The Broker receives
this request and forwards it to the UCSm. UCSm realizes that in the policy
regarding authorization of subscriptions contains a noise level attribute that
cannot be retrieved locally by a PIP. Since this value is not local to the system
M but belongs to the UCSc of System C, the UCSm contacts the UCSc and asks
for access to the value of the noise level from the noise sensor. After receiving the
value from the UCSc, the UCSm evaluates the request and forwards the result to
the Broker, which, in case of “Permit”, allows the vacuum cleaner to retrieve the
data coming from the smart meter. In parallel, the UCSm continuously checks
all the attributes and evaluates the policy. In the event that the value of the
noise level or the power consumption rises above policy-defined threshold, the
policy is violated and the access of the vacuum cleaner to the data of the smart
meter is revoked, which also leads to the cancellation of the subscription from
the MQTT system. Thus, without this information the vacuum cleaner will halt
its task.
Distributed UCON in CoAP and MQTT Protocols 47
4.2 Implementation
The UCS framework was implemented as a Java application. The software used
to implement the Broker of the MQTT protocol was Moquette 2 . This Broker is
based on the same programming language as the UCS framework which helped
making sure that they could be integrated in the same device. Regarding MQTT
Subscribers and Publishers we used off the shelf Python-based implementations.
These ran as standard, without any modifications. Only the Moquette Broker
was partially modified so as to host the PEP and call the UCS when invoking
the TryAccess, StartAccess, EndAccess actions and wait the response from UCS.
The software used for the CoAP system was Californium 3 which is also
developed in Java like Moquette. Californium is a well-known implementation of
CoAP and provides implementation of both Clients and Servers. Following the
same pattern of the MQTT implementation, the Clients of CoAP ran without
any modification whereas the Server was partially modified in order to host the
PEP and the UCS. When a Client wants to observe some resource on a Server,
the PEP calls the UCS invoking the same actions as previously in the case of
the Moquette Broker. In the case of a negative response, the Server returns a
message to the Client that it was not allowed to observe that resource. If, in
the case of both protocols, there is a policy violation, RevokeAccess is invoked
by the UCS to the PEP and on both cases the session is terminated and the
Subscriber or the Client, depending on the protocol, is removed.
5 Related Work
Over the past few years a number of research works have presented new ways
to improve the security of CoAP. Most of these works are focused on reducing
the overhead of using DTLS on top of CoAP. In [21], the authors introduced
Lithe which proposes improvements in the integration between CoAP and DTLS
to allow increased performance and more efficient packet sizes and energy con-
sumption. In [28], the authors propose a lightweight security scheme in CoAP
using Advanced Encryption Standard (AES) 128 symmetric key algorithm, intro-
ducing an object security (payload embedded)-based authentication mechanism
with integrated key management. Finally, in [3] the authors introduced REST-
ful DTLS connections as CoAP resources using Elliptic Curve Cryptography
(ECC)-based cryptography, achieving gains in Read Only Memory (ROM) and
Random Access Memory (RAM) occupancy. The IETF supports efforts to secure
coap, like DTLS for CoAP [9], or the Object Security for Constrained REST-
ful Environments (OSCORE) group which is developing a mode for protect-
ing group communication over CoAP [27]. The new DTLS standard (1.3) adds
improvements in performance and security [23] which CoAP can take advantage.
However our solution, provides a continuous control monitoring mechanism, that
according to the values of critical attributes, can revoke the access when pol-
icy is violated. Moreover, for MQTT protocol, the most significant effort about
security, is SMQTT [26]. It adds a security feature based on Key/Ciphertext
Policy-Attribute Based Encryption (KP/CP-ABE) that uses lightweight Ellip-
tic Curve Cryptography. This solution produces significant overhead and requires
specific Publishers and Subscribers for encryption and decryption. However, our
solution does not require specific Publisher s or Subscriber s and the overhead is
only on the side of the Broker hosting the UCS.
The area of IoT includes various protocols and applications that target to a
wide range of constrained environments. Thus, the complexity to create security
enforcement tools that can handle policies in such environments increases. To
this effort, there are other works based on UCON in the area of IoT. In [19],
the authors have presented an effort to create a variation of UCON that tries to
adapt UCON to the area of IoT, but they do not provide any application to an
application layer protocol where the information is shared. There is also another
work that tries to address this problem that focuses on how UCON can be inte-
grated in MQTT protocol [15], [14]. This effort focuses on the MQTT protocol
highlighting that it is the most generic, since, as they claim, the architecture
of MQTT protocol fits nicely with the architecture of UCS. However, this work
tries to prove that UCON can fit not only in one protocol, but also in other pro-
tocols with different architectures in the same time providing interoperability
and attribute sharing among different protocols.
6 Conclusion
To the best of our knowledge, up to now the efforts for policy enforcement and
continuous monitoring of the access in IoT application layer protocols are still
50 A. Rizos et al.
References
1. Al-Fuqaha, A., Guizani, M., Mohammadi, M., Aledhari, M., Ayyash, M.: Internet
of Things: a survey on enabling technologies, protocols, and applications. IEEE
Commun. Surv. Tutor. 17(4), 2347–2376 (2015). https://doi.org/10.1109/COMST.
2015.2444095. (Fourthquarter)
2. Bastos, D., Shackleton, M., El-Moussa, F.: Internet of Things: a survey of tech-
nologies and security risks in smart home and city environments. In: Living in the
Internet of Things: Cybersecurity of the IoT - 2018, pp. 1–7, March 2018. https://
doi.org/10.1049/cp.2018.0030
3. Capossele, A., Cervo, V., De Cicco, G., Petrioli, C.: Security as a CoAP resource:
an optimized DTLS implementation for the IoT. In: 2015 IEEE International Con-
ference on Communications (ICC), pp. 549–554. IEEE (2015)
4. Carniani, E., D’Arenzo, D., Lazouski, A., Martinelli, F., Mori, P.: Usage control
on cloud systems. Future Gener. Comput. Syst. 63(C), 37–55 (2016). https://doi.
org/10.1016/j.future.2016.04.010
5. Chen, D., Varshney, P.K.: QoS support in wireless sensor networks: a survey. In:
International Conference on Wireless Networks, vol. 233, pp. 1–7 (2004)
6. Collina, M., Corazza, G.E., Vanelli-Coralli, A.: Introducing the QEST broker: scal-
ing the IoT by bridging MQTT and REST. In: 2012 IEEE 23rd International
Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC),
pp. 36–41, September 2012. https://doi.org/10.1109/PIMRC.2012.6362813
7. Costantino, G., La Marra, A., Martinelli, F., Mori, P., Saracino, A.: Privacy pre-
serving distributed attribute computation for usage control in the Internet of
Things. In: 2018 17th IEEE International Conference on Trust, Security and Pri-
vacy in Computing and Communications/12th IEEE International Conference on
Big Data Science and Engineering (TrustCom/BigDataSE), pp. 1844–1851, August
2018. https://doi.org/10.1109/TrustCom/BigDataSE.2018.00279
8. Faiella, M., Martinelli, F., Mori, P., Saracino, A., Sheikhalishahi, M.: Collaborative
attribute retrieval in environment with faulty attribute managers. In: 2016 11th
International Conference on Availability, Reliability and Security (ARES), pp. 296–
303, August 2016. https://doi.org/10.1109/ARES.2016.51
9. Gerdes, S., Bergmann, O., Bormann, C., Selander, G., Seitz, L.: Datagram Trans-
port Layer Security (DTLS) Profile for Authentication and Authorization for
Constrained Environments (ACE). Internet-Draft draft-ietf-ace-dtls-authorize-07,
Internet Engineering Task Force, March 2019. https://datatracker.ietf.org/doc/
html/draft-ietf-ace-dtls-authorize-07. (work in Progress)
Distributed UCON in CoAP and MQTT Protocols 51
10. Giusto, D., Iera, A., Morabito, G., Atzori, L.: The Internet of Things. Springer,
New York (2010). https://doi.org/10.1007/978-1-4419-1674-7
11. Granjal, J., Monteiro, E., Silva, J.S.: Security for the Internet of Things: a survey
of existing protocols and open research issues. IEEE Commun. Surv. Tutor. 17(3),
1294–1312 (2015)
12. Hartke, K.: Observing Resources in the Constrained Application Protocol (CoAP).
RFC 7641, September 2015. https://doi.org/10.17487/RFC7641. https://rfc-
editor.org/rfc/rfc7641.txt
13. Karopoulos, G., Mori, P., Martinelli, F.: Usage control in SIP-based multimedia
delivery. Comput. Secur. 39, 406–418 (2013). https://doi.org/10.1016/j.cose.2013.
09.005
14. La Marra, A., Martinelli, F., Mori, P., Rizos, A., Saracino, A.: Improving MQTT
by inclusion of usage control. In: Wang, G., Atiquzzaman, M., Yan, Z., Choo, K.-
K.R. (eds.) SpaCCS 2017. LNCS, vol. 10656, pp. 545–560. Springer, Cham (2017).
https://doi.org/10.1007/978-3-319-72389-1 43
15. La Marra, A., Martinelli, F., Mori, P., Rizos, A., Saracino, A.: Introducing usage
control in MQTT. In: Katsikas, S.K., et al. (eds.) CyberICPS/SECPRE 2017.
LNCS, vol. 10683, pp. 35–43. Springer, Cham (2018). https://doi.org/10.1007/
978-3-319-72817-9 3
16. Lazouski, A., Martinelli, F., Mori, P.: Usage control in computer security: a survey.
Comput. Sci. Rev. 4(2), 81–99 (2010). https://doi.org/10.1016/j.cosrev.2010.02.002
17. Lazouski, A., Martinelli, F., Mori, P., Saracino, A.: Stateful data usage control
for android mobile devices. Int. J. Inf. Secur., pp. 1–25 (2016). https://doi.org/10.
1007/s10207-016-0336-y
18. Locke, D.: MQ telemetry transport (MQTT) v3. 1 protocol specification. IBM
developerWorks Technical Library (2010)
19. Marra, A.L., Martinelli, F., Mori, P., Saracino, A.: Implementing usage control
in Internet of Things: a smart home use case. In: 2017 IEEE Trustcom/Big-
DataSE/ICESS, pp. 1056–1063, August 2017. https://doi.org/10.1109/Trustcom/
BigDataSE/ICESS.2017.352
20. Park, J., Sandhu, R.: The UCONABC usage control model. ACM Trans. Inf. Syst.
Secur. 7(1), 128–174 (2004). https://doi.org/10.1145/984334.984339
21. Raza, S., Shafagh, H., Hewage, K., Hummen, R., Voigt, T.: Lithe: lightweight
secure CoAP for the Internet of Things. IEEE Sens. J. 13(10), 3711–3720 (2013).
https://doi.org/10.1109/JSEN.2013.2277656
22. Rescorla, E., Modadugu, N.: Datagram Transport Layer Security Version 1.2.
RFC 6347, January 2012. https://doi.org/10.17487/RFC6347. https://rfc-editor.
org/rfc/rfc6347.txt
23. Rescorla, E., Tschofenig, H., Modadugu, N.: The Datagram Transport Layer Secu-
rity (DTLS) Protocol Version 1.3. Internet-Draft draft-ietf-tls-dtls13-31, Internet
Engineering Task Force, March 2019. https://datatracker.ietf.org/doc/html/draft-
ietf-tls-dtls13-31. (work in Progress)
24. INFSO D.4 Networked Enterprise and RFID INFSO G.2 Micro and Nanosystem:
Internet of Things in 2020, A Roadmap for the Future (2009)
25. Shelby, Z., Hartke, K., Bormann, C.: The Constrained Application Protocol
(CoAP). RFC 7252, June 2014. https://doi.org/10.17487/RFC7252. https://rfc-
editor.org/rfc/rfc7252.txt
52 A. Rizos et al.
26. Singh, M., Rajan, M.A., Shivraj, V.L., Balamuralidhar, P.: Secure MQTT for Inter-
net of Things (IoT). In: 2015 Fifth International Conference on Communication
Systems and Network Technologies, pp. 746–751, April 2015. https://doi.org/10.
1109/CSNT.2015.16
27. Tiloca, M., Selander, G., Palombini, F., Park, J.: Group OSCORE - Secure Group
Communication for CoAP. Internet-Draft draft-ietf-core-oscore-groupcomm-04,
Internet Engineering Task Force, March 2019. https://datatracker.ietf.org/doc/
html/draft-ietf-core-oscore-groupcomm-04. (work in Progress)
28. Ukil, A., Bandyopadhyay, S., Bhattacharyya, A., Pal, A., Bose, T.: Lightweight
security scheme for IoT applications using CoAP. Int. J. Pervasive Comput. Com-
mun. 10(4), 372–392 (2014)
Towards the Creation of a Threat
Intelligence Framework for Maritime
Infrastructures
1 Introduction
and collected data through an open industry platform, VERACITY [4] which is
specifically designed for maritime industry.
The rest of the paper is organized as follows: Sect. 2 offers some background
information about cyber attacks against maritime ecosystem; Sect. 3 introduces
the MAINFRAME framework and briefly describes its components while refer-
encing the state of the art practises; Sect. 4 describes MAINFRAME’s operation
in a LNG carrier use case; and finally, Sect. 5 draws the conclusions giving some
pointers for future work.
2 Maritime Security
One of the most popular cyber attacks that has the largest financial impact on
maritime industry is conducted via Business Email Compromise (BEC) [5], a
social engineering attack. BEC operators have a solid understanding of various
maritime specific operations and they use their knowledge to take advantage of
employees with low awareness of cyber attacks. For example, last year researchers
from Dell SecureWorks have identified a hacking group called GOLD GALLEON,
which is responsible for many wide-scale BEC attacks targeting maritime compa-
nies, such as ship management services, port services and cash to master services
[6]. Ransomware is another type of cyber threat affecting the maritime industry.
An exemplary victim of such cyber attack is COSCO (China Ocean Shipping
Company) US branch [7] which was attacked by a ransomware that resulted in
the breakdown of their telephone network, email servers, even the US website of
the company being offline in July 2018.
One of the most attractive targets in the maritime sector is Industrial Con-
trol Systems (ICS), which are critical parts of a vessel’s operations. Should cyber
attackers bypass the security perimeter of a vessel’s IT network they can inject
malware to ICS systems. During the last years, several attacks using ICS mal-
ware have been reported, with the most recent one targeting Triconex Safety
56 N. Pitropakis et al.
Instrumented System (SIS) controllers made by Schneider Electric [8]. The mal-
ware replaces the Logic of SIS controllers, an action which can: (i) use SIS to
shutdown the process, resulting in financial losses due to the process downtime
and the plant start-up procedure after the shutdown, (ii) reprogram the SIS
to allow an unsafe state, which will increase the risk of a hazardous situation
with physical consequences (e.g. impact to equipment, product, environment and
human safety) and (iii) reprogram the SIS to allow an unsafe state, while using
the Distributed Control System (DCS) to create an unsafe state or hazard, which
will have an impact on human safety, the environment, or equipment damage,
the extent of which depends on the physical constraints of the process and the
plant design. Interestingly, instances of STUXNET have been found on vessels
using the same Siemens industrial controllers as the Uranium enrichment plants
of Iran [9].
3 MAINFRAME Technologies
3.1 VERACITY
someone can access all DNV GL’s digital services and applications, but also
services from third parties such as the security solutions that comprise MAIN-
FRAME. VERACITY also includes a community for developers to make it easier
to develop new applications and analytics. On the platform’s marketplace, asset
owners can subscribe to industrial applications and data analytics services that
can help them make better use of data to optimize their safety and performance.
These applications will be provided not only by DNV GL, but by a host of other
qualified providers of data, data analytics, digital solutions and cyber security
tools. VERACITY will make easier the implementation and the interoperability
of the rest of the MAINFRAME’s components. All of the components will use
VERACITY as the main pillar of interaction. For example, VERACITY will
be capable to host extra functionalities such as the Event Detection and Event
Monitoring components that will orchestrate the cyber threats detection, thus
augmenting the existing VERACITY’s functionalities.
Data Processing’s (DProc) task is the collection of data from a range of
sources and silos of the maritime ecosystem. After the data has been ingested,
VERACITY checks its quality and then provides the necessary cleaning process.
The latter will permit the normalization and aggregation of data from multiple
sources in a common format and its addition to the VERACITY’s database,
where the data is physically protected and encrypted at the same time. This
step is quite critical as the maritime industry lacks on standardization when it
comes to data acquisition and IT systems on-board vessels. Consequently, the
maritime companies can have a better overview of what is really happening in
physical and cyber level, allowing the analysis of events and possible correlation
between them. For example, access in a restricted area might be correlated with
an installed malware that lead to an infection of the systems. Cyber events along
with cyber physical events will offer a better understanding of processes in mar-
itime environment, thus increasing the cyber security situational awareness at
any given time. Dproc also includes processes that are meant for secure oper-
ations, where strong controls are enforced to limit access to trusted personnel
only. MAINFRAME, will augment VERACITY’s capabilities by reinforcing the
cyber security part.
Data Fabric (DFab) offers important services to users. The user can control
what data to share, with whom and the amount of time. As the data is extracted
from proprietary silos and legacy IT systems, they are being uploaded in a
cloud-based infrastructure, giving the ability to perform automated analytics,
which can lead to comparisons between assets and benchmarking. The data the
user chooses to share on the platform is isolated from other products and other
businesses’ data unless the user chooses otherwise. Consequently, every maritime
business can benefit in terms of training, research and analytics tools. Training
and research are focused on cyber security, data quality assessment and reliability
of sensor data. MAINFRAME’s goal is to strengthen the cyber security part, as
the potential user will be able to access to threat intelligence data related to cyber
threats and cyber security incidents. The threat intelligence data, correlated with
the rest of the collected data from silos and legacy systems, will offer a better
58 N. Pitropakis et al.
view and understanding about all occurred incidents as the physical, cyber and
cyber physical approaches will work together towards a common goal. Regarding
the analytics tools, the threat detection and threat virtualization components
will offer the ability to detect and visualize cyber security incidents. Additionally,
the analytics about intrusions and metrics about the robustness of every system
will help the potential users strengthen their defences against cyber threats.
Configures
Feedback
Configures
Event Event
Honeypots Deep Learning
Monitoring Detecon
Configures
Data
Data Fabric
Processing
Configures
Maritime Infrastructure
Critical Infrastructure
Computer IoT
SCADA Switches Firewall
Terminals Devices
resources, deep learning methods can be employed [15]. Instead, for those systems
that are computationally constrained, statistical-based methods are used. For
autonomous devices, Bayesian networks have been studied [16], and have been
proven as a feasible approach, and presented that they can determine whether
the attack has originated in physical or cyber domain.
3.3 Honeypots
Honeypots are decoy systems, deployed alongside production systems with the
intent of tricking the potential attackers into breaking the security of the decoy
systems. It is a system which value is in getting its security violated. This is one
of the few methods available to a security professional to deceive the attackers
and study their behaviour. The main objectives of honeypots are to divert mali-
cious traffic away from important systems, get early warning of a current attack
before critical systems are hit, and gather information about the attacker and
the attack methods. By deceiving an attacker into carrying out his/her attack on
a non-critical, well-monitored system, valuable insight can be gained into their
attack methods, and information can be gathered for forensic or legal purposes.
Deception is so crucial to detecting lateral movement, uncovering privilege esca-
lation, and building threat intelligence, that any kind of honeypots are valuable.
MAINFRAME, makes use of a variety of open source honeypots that will be
implemented on top of SIEM. In specific the honeypots include: (a) IoT (Inter-
net of Things) decoys: Honeything [17] emulates the TR-069 WAN management
protocol, as well as a RomPager web-server, with vulnerabilities. Other IoT
decoys can be created by emulating embedded telnet/FTP servers, for exam-
ple with BusyBox; (b) SCADA/ICS decoys: ConPot [18] emulates a number of
operational technology control systems infrastructure, including protocols like
ModBus, DNP3 and Bacnet. It comes with a web-server that can emulate a
SCADA HMI as well; and (c) Generic: Honeydrive [19] is a GNU/Linux dis-
tribution that comes pre-installed with a lot of active defence capabilities. The
purpose of MAINFRAME is to cover the majority of the activities in a maritime
ecosystem, thus using decoys for IoT devices, SCADA/ICS and general-purpose
solutions that will help SIEM to increase its robustness towards a variety of
attacks and cyber threats. The SIEM endpoint instance/device will correlate
and report back data collected from Honeypots in near real time. Additionally,
those data will populate the VERACITY’s knowledge base, so as to be used by
other clients in favour of their protection.
counter measurements, e.g. patches, attack signatures, antivirus design but also
assist network forensic investigations. Attacks on honeypots can also serve as an
early warning system for administrators, providing more time to react to attacks
in progress. Honeypots have been popular in traditional IT systems and the com-
munity has tried to maximize their use. Jasek et al. [20] stated that it is feasible
to use honeypots in order to detect APT campaigns. In a similar way of think-
ing, Pitropakis et al. [21] suggested the use of honeypots in combination with
the creation of virtual personas willing to achieve cyber attack attribution. How-
ever, the use of honeypots is not so popular in the maritime ecosystem. In 2013
Solomon [22] suggested the use of honeypots that can act as a trap against mali-
cious parties. Bou-Harb et al. [23] in 2017 also suggested the use of honeypots as
a countermeasure of cyber physical systems which can be attacked in maritime
transportation. Honeypots can be monitored either internally or externally an
activity that is of paramount importance. Being deployed inside the monitored
honeypots, internal monitoring mechanisms are able to provide a semantic-rich
view on various aspects of system dynamics (e.g., system calls). However, their
very internal existence makes them visible and tangible [24]. On the other hand,
external honeypot monitoring systems are invisible to cyber attackers by they
are not capable of capturing any internal activity such as system calls. While
Sebek is one of the most important honeypot monitoring system used in a variety
of high-interaction honeypots internally, the authors in [24] proposed an exter-
nal virtualisation-based system, called VMscope, which is able to monitor the
system internal events of VM-based honeypots from outside the honeypots. The
authors also provided a proof-of-concept prototype by leveraging and extending
one key virtualization technique called binary translation. Based on their exper-
iments with real-world honeypots, VMscope is proved robust against advanced
countermeasures that can defeat existing internally-deployed honeypot monitors.
3.4 Blockchain
In MAINFRAME, the deployment of blockchain is based on the open source
Hyperledger Fabric (HLF) project [10] which is a global open source collabo-
rative initiative to advance cross-industry distributed ledger and is hosted by
the Linux Foundation. Hyperledger Fabric is a platform for distributed ledger
solutions underpinned by a modular architecture delivering high degrees of confi-
dentiality, resiliency, flexibility and scalability. It is designed to support pluggable
implementations of different components and accommodate the complexity and
intricacies that exist across the economic ecosystem. It also allows secure &
private transactions between businesses in supply chains and other use cases.
MAINFRAME implements a smart contract engine with a greater focus on
this enterprise-by-design concept than in other smart contract engines such as
Ethereum [25]. It delivers greater scalability than other blockchain solutions by
allowing for consensus mechanisms which meet the requirements of the business,
meaning fewer nodes are needed to validate transactions while still maintaining
practical byzantine-fault tolerance. This implementation allows full data supply
chain auditing as the use of this type of technology prevents malicious actors from
62 N. Pitropakis et al.
editing the transaction history, which can help to identify the group responsible
in case of a breach.
With data breaches and identity theft becoming more sophisticated and fre-
quent, companies need a way to take ownership of their identity, thus making
blockchain technology and protocols well suited for Enabling Decentralized IDs
(DID). Additionally, blockchain offers a secure encrypted digital hub (ID Hubs)
that can interact with user’s data while honouring user privacy and control,
resulting in Privacy by design, built in from the ground up. Traditional iden-
tity systems are mostly geared toward authentication and access management.
A self-owned identity system will add a focus on authenticity and aid maritime
community to establish trust. In a decentralized system trust is based on attes-
tations: claims that other entities endorse – which helps prove facets of one’s
identity. In addition to that DIDs and ID Hubs will enable developers to gain
access to a more precise set of attestations while reducing legal and compliance
risks by processing such information. Finally, the proposed solution will result in
a robust decentralized identity ecosystem that is accessible to all users, as it will
be built on top of standard, open source technologies, protocols, and reference
implementations.
State-of-the-Art. During the last decade, the popularity of the blockchain has
been increasing and has reached important notoriety not only in scientific and
IT journals but also in general public media. Since Blockchains began attract-
ing the attention of the financial, security and IT communities, several other
blockchain implementations have appeared (e.g. Ethereum [25]). As an alter-
native to Bitcoin and Ethereum, The Linux Foundation has proposed a new
blockchain project called Hyperledger. This project is a blockchain framework
to develop new services and applications based on a permissioned ledger. Even
though Hyperledger can be used for a wide spectrum of applications, one of
the most popular is for Smart Contracts development. The Hyperledger project
consists of five blockchain frameworks: Fabric, Iroha, Sawtooth, Burrow and
Indy (not released yet). In the case of Sawtooth, the implementation considers
a new consensus algorithm called Proof of Elapsed Time (PoET) that solves
the power consumption issue that Bitcoin has. On the other hand, Iroha is
a simple blockchain scheme designed to be easy to use for multiple business
applications. A more sophisticated blockchain implementation is Burrow. This
blockchain implementation is a smart contract machine based on the Ethereum
Virtual Machine (EVM) specifications. Finally, Fabric is a blockchain modular
framework that gives the flexibility to change different components by plug and
play. Moreover, the blockchain replication process between the nodes is cost-
efficient and capable to process about 3.500 tps [10], thanks to their consensus
algorithm based on Practical Byzantine Fault Tolerant. This makes Hyperledger
Fabric one of the best options for customizing a blockchain implementation.
Towards the Creation of a Threat Intelligence Framework 63
The collected data will be subject to linguistic analysis and go through a series of
processes employing a pipeline of tools in the following order: sentence-breaking,
tokenization, local grammars, lexicons, part of speech tagging, text normaliza-
tion, and eventually parsing. Additionally, there will be an extended use of “Reg-
ular Expressions”, which are expected to have a significant impact on the quality
of the results, whereas specialized gazetteers of relevant entities will also aid the
extraction. SMC aims at combining the social media concepts/named entities
extracted in a smart and scalable way, given the large and highly heterogeneous
nature of the data structure. These information will further feed VERACITY’s
threat intelligence database with more data. It should be noted that the whole
process will be carried out on publicly available data respecting always the users’
privacy.
identifying user behavior patterns [33]. However, despite all these activities, lit-
tle has been done on improving visual threat analytics capabilities of SIEMs
specifically and informed decision making is vital for specific domain needs.
LNG Carrier
3. Escalate Privileges
6. Correlaon
Honeypot
5. Monitoring
2. Gain Access 4. Enter Honeypot
1. Scan Aack
7. Aack Detecon
Temperature Sensor
9. Stop Aack
8. VTA Supported Detecon
Management
Veracity
Console
Creang Audit Trail
Blockchain
5 Conclusion
Driven by the increasing use of ICT systems in maritime environments and the
lack of focus on the cyber threats, we proposed a new framework, whose goal is
to aid threat intelligence collection and analysis in the maritime ecosystem. Our
solution is built on top of a commercially successful product, strictly designed for
maritime needs. On top of it, we suggested the addition of up to date intrusion
detection solutions, honeypot systems and visualization approaches that will
maximize the utilization of the previous components. Willing to secure the big
amount of transactions, our solutions makes also use of hyperledger blockchain
technology to achieve that, while enabling the use of EIDs.
The process of integrating so many different components, especially when
some of them have to be developed is a very challenging task. However, this is
not the end, rather than the beginning of systematic orchestration of defen-
sive mechanics that will help to protect the cyber maritime infrastructures
against malicious parties. MAINFRAME through the LNG Carrier case study
will also highlight the interdependences that exist between cyber and cyber-
physical threats as autonomous ships become more popular over the years. The
key to securing cyber maritime infrastructures remains the same and is the threat
intelligence gathering and sophisticated analysis.
Acknowledgments. This work has been partially supported by the Research Center
of the University of Piraeus.
Towards the Creation of a Threat Intelligence Framework 67
References
1. Greenberg, A.: The untold story of NotPetya, the most devastating cyberattack in
history. Wired, August 2018
2. Taddeo, M., Floridi, L.: Regulate artificial intelligence to avert cyber arms race.
Nature 556(7701), 296–298 (2018)
3. Apostolou, B., Apostolou, N., Schaupp, L.C.: Assessing and responding to cyber
risk: the energy industry as example. J. Forensic Investig. Account. 10(1) (2018)
4. DNVGL: Veracity. https://www.dnvgl.com/data-platform/index.html/. Accessed
09 Feb 2019
5. Mansfield-Devine, S.: The imitation game: how business email compromise scams
are robbing organisations. Comput. Fraud Secur. 2016(11), 5–10 (2016)
6. O’Donnell, L.: Gold galleon hacking group plunders shipping industry. https://
threatpost.com/gold-galleon-hacking-group-plunders-shippingindustry/131203/.
Accessed 09 Feb 2019
7. Shapo, V.: Cybersecurity implementation aspects at shipping 4.0 and industry 4.0
concepts realization. J. Sci. Perspect. 2(4), 1–12 (2018)
8. Johnson, B., Caban, D., Krotofil, M., Scali, D., Brubaker, N., Glyer, C.: Attackers
deploy new ICS attack framework “TRITON” and cause operational disruption to
critical infrastructure. Threat Research Blog (2017)
9. Marks, P.: Why the Stuxnet worm is like nothing seen before. News Science (2010)
10. Cachin, C.: Architecture of the hyperledger blockchain fabric. In: Workshop on
Distributed Cryptocurrencies and Consensus Ledgers, vol. 310 (2016)
11. OSSEC: Ossec. https://www.ossec.net/. Accessed 09 Feb 2019
12. Pitropakis, N., Lambrinoudakis, C., Geneiatakis, D.: Till all are one: towards a
unified cloud IDS. In: Fischer-Hübner, S., Lambrinoudakis, C., Lopez, J. (eds.)
TrustBus 2015. LNCS, vol. 9264, pp. 136–149. Springer, Cham (2015). https://
doi.org/10.1007/978-3-319-22906-5 11
13. IBM: Artificial intelligence for a smarter kind of cybersecurity. https://www.ibm.
com/security/artificial-intelligence/. Accessed 09 Feb 2019
14. Carasso, D.: Exploring Splunk. CITO Research, New York (2012)
15. Wei, J., Mendis, G.J.: A deep learning-based cyber-physical strategy to mitigate
false data injection attack in smart grids. In: Joint Workshop on Cyber-Physical
Security and Resilience in Smart Grids (CPSR-SG), pp. 1–6. IEEE (2016)
16. Bezemskij, A., Loukas, G., Gan, D., Anthony, R.: Detecting cyber-physical threats
in an autonomous robotic vehicle using Bayesian networks, pp. 98–103 (2017)
17. IBM: Honeything. https://www.ibm.com/security/artificial-intelligence/.
Accessed 09 Feb 2019
18. Conpot: Conpot. http://conpot.org/. Accessed 09 Feb 2019
19. BruteForce Lab: HoneyDrive. https://bruteforcelab.com/honeydrive/. Accessed 09
Feb 2019
20. Jasek, R., Kolarik, M., Vymola, T.: APT detection system using honeypots. In:
Proceedings of the 13th International Conference on Applied Informatics and Com-
munications (AIC 2013), pp. 25–29. WSEAS Press (2013)
21. Pitropakis, N., Panaousis, E., Giannakoulias, A., Kalpakis, G., Rodriguez, R.D.,
Sarigiannidis, P.: An enhanced cyber attack attribution framework. In: Furnell, S.,
Mouratidis, H., Pernul, G. (eds.) TrustBus 2018. LNCS, vol. 11033, pp. 213–228.
Springer, Cham (2018). https://doi.org/10.1007/978-3-319-98385-1 15
22. Solomon, J.F.: Maritime deception and concealment: concepts for defeating wide-
area oceanic surveillance-reconnaissance-strike networks. Naval War Coll. Rev.
66(4), 87 (2013)
68 N. Pitropakis et al.
23. Bou-Harb, E., Kaisar, E.I., Austin, M.: On the impact of empirical attack models
targeting marine transportation. In: 2017 5th IEEE International Conference on
Models and Technologies for Intelligent Transportation Systems (MT-ITS), pp.
200–205. IEEE (2017)
24. Jiang, X., Wang, X.: “Out-of-the-Box” monitoring of VM-based high-interaction
honeypots. In: Kruegel, C., Lippmann, R., Clark, A. (eds.) RAID 2007. LNCS,
vol. 4637, pp. 198–218. Springer, Heidelberg (2007). https://doi.org/10.1007/978-
3-540-74320-0 11
25. Wood, G.: Ethereum: a secure decentralised generalised transaction ledger.
Ethereum Proj. Yellow Pap. 151, 1–32 (2014)
26. Keim, E.D., Kohlhammer, J., Ellis, G.: Mastering the information age: solving
problems with visual analytics. Eurographics Association (2010)
27. Andrienko, N., et al.: Viewing visual analytics as model building. In: Computer
Graphics Forum. Wiley Online Library (2018)
28. Andrienko, G., Andrienko, N., Fuchs, G.: Understanding movement data quality.
J. Locat. Based Serv. 10(1), 31–46 (2016)
29. Endert, A., et al.: The state of the art in integrating machine learning into visual
analytics. In: Computer Graphics Forum, vol. 36, pp. 458–486. Wiley Online
Library (2017)
30. Shiravi, H., Shiravi, A., Ghorbani, A.A.: A survey of visualization systems for
network security. IEEE Trans. Vis. Comput. Graph. 18(8), 1313–1329 (2012)
31. Fischer, F., Keim, D.A.: NStreamAware: real-time visual analytics for data streams
to enhance situational awareness. In: Proceedings of the Eleventh Workshop on
Visualization for Cyber Security, pp.65–72. ACM (2014)
32. Chen, S., Guo, C., Yuan, X., Merkle, F., Schaefer, H., Ertl, T.: OCEANS: online col-
laborative explorative analysis on network security. In: Proceedings of the Eleventh
Workshop on Visualization for Cyber Security, pp. 1–8. ACM (2014)
33. Chen, S., et al.: User behavior map: visual exploration for cyber security session
data (2018)
Connect and Protect: Requirements
for Maritime Autonomous Surface Ship
in Urban Passenger Transportation
1 Introduction
According to the most recent report from the Norwegian Shipowners’ Associa-
tion, exactly half of the global shipping companies will implement autonomous
ships by 2050, while Rolls-Royce aims to operate autonomous unmanned ocean-
going ships by 2035 [25]. In this direction, the International Maritime Organi-
zation (IMO) started to address the regulatory scope for autonomous ships [8].
Norway is leading the autonomous shipping industry by opening several testing
areas for the development of this technology, in addition to the production of
Yara Birkland, the worlds first all-electric and autonomous cargo ship [27], and
other projects aiming to operate autonomous passenger ferries in different loca-
tions [5,28]. Many other initiatives all around the globe are taking place towards
the development of autonomous ships; for instance, in 2018, Rolls-Royce and a
Finish ferry operator demonstrated the world’s first fully autonomous ferry in
Finland [26].
The Norwegian Forum for Autonomous Ships (NFAS) has provided defini-
tions for autonomous ships, their context, and functions in [33]. A classification
of autonomous maritime systems was suggested, depending on the operational
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 69–85, 2020.
https://doi.org/10.1007/978-3-030-42048-2_5
70 A. Amro et al.
2 Related Work
Although the guidelines exclude ships smaller than 20m, we believe that the
suggested guidelines related to communication are relevant for the APS. Addi-
tionally, DNV GL published several documents discussing aspects of autonomous
ships. In their position paper they discussed the expected change in navigation,
the regulatory scope, safety assurance, and social and ethical assurance [21].
Another related document from DNV GL is the class guidelines for autonomous
and remotely operated ships [19]. In this document, DNV GL discussed several
aspects including navigation functions, communication functions and cybersecu-
rity considerations.
Several works discuss the lack of a regulatory framework that governs the
operation of autonomous ships and suggests solutions to adapt to such tech-
nology. The Danish maritime authorities published a report on the regulatory
barriers to the use of autonomous ships, suggesting suitable steps toward tack-
ling these barriers, such as creating new laws for autonomous ships or amend-
ing existing ones [17]. Another work surveyed relevant regulations that might
affect the operational capacity of autonomous ships [23]. The authors discussed
regulations like SOLAS, COLREGS (International Regulations for Preventing
Collisions at Sea), and others in detail, and pointed out that the regulations
in their current form limit the deployment of autonomous ships. The work pre-
sented in [23] suggested generic communication requirements in order to satisfy
certain regulations such as the availability of delay-free, reliable, fast and secure
communication between the ship and control center.
A general system context for the operation of a MASS as shown in Fig. 1 was
suggested by NFAS. A brief description of the context components and their
relevance to the APS is given below:
refer to this entity as Shore Control Center (SCC), others as Remote Control
Center (RCC); herein we adopt the latter term. An RCC functions as an
observer, by monitoring the APS status, but in some cases it might be forced
to take control of the ship in order to avoid accidents. For this reason, it was
concluded that certain manning requirements are important for the RCC to
operate [36]. Additionally, a single or a chain of RCCs might be expected to
serve several ships concurrently. The location of the RCC might be on shore
or it can reside on-board another vessel (e.g. an escort vessel).
– Emergency Control Team (ECT): a team which is expected to intervene
in case of emergencies endangering the passengers or the surrounding environ-
ment. For instance, a passenger falling into water, or the ship not responding
to remote commands and heading on a collision course.
– Shore Sensor System (SSS): A collection of sensors are expected to be
mounted on shore to aid some functions of the APS. For instance, ship auto-
matic docking, charging, and other functions related to passenger embarking
and disembarking.
– VTS/RIS: Ships are expected to establish contact with Vessel Traffic Ser-
vices (VTS) for guidance and reporting. Moreover, the European Parliament
has defined activities towards establishing harmonized River Information Ser-
vices (RIS) for inland waterways to facilitate navigation [13].
– Aids to Navigation (AtoN): Collection of systems expected to provide
real-time information for the ship navigation system regarding weather, other
ships, location awareness, etc. Examples of such systems are the Automatic
Identification System (AIS), the Global Navigation Satellite System (GNSS),
Radar, LIght Detection and Ranging (Lidar), etc.
– Other Ships: The APS is expected to communicate with other ships in the
area for sharing navigational information using several agreed upon commu-
nication systems, such as Very high frequency (VHF), the more advanced
VHF Data Exchange System (VDES) or AIS.
– Port Services: Some services, such as electric charging, maintenance, pas-
senger embarking and disembarking, might be provided to the APS at the
port or quay.
Other components in Fig. 1, such as the Maritime Rescue Coordination Cen-
tre (MRCC), Global Maritime Distress and Safety System (GMDSS), and Ser-
vice vessels (Pilot, tug, etc.) are less relevant to the case of the APS, due to the
smaller size of its operational area.
Regulations
Title Section/Chapter Scope APS Applicaple
SOLAS International voyages ✗
ISM ✗
ISPS ✗
GMDSS ✗
UNCLOS Sea ✗
STCW ✗
MARPOL ✗
SAR ✗
COLREG Sea Connected ✓*
Standards
IEC 61162 1 (NMEA 0183) Serial Communication ✓
3 (NMEA 2000) ✓
450 Ethernet ✓
460 Ethernet and Security ✓
IEC 61850 90-4 LAN Engineering ✓
MSC.252 83 Integrated Navigation System ✓
IEC 62443 3-3 Security of Industrial Control Systems ✓
ISO/IEC 27000 27001 Information Security Management ✓
Systems
27002 ✓
IEC 62940 Communication between on-board ✓
systems and external computer
systems
✓*: Require modifications
In the case of APS in urban transportation, the most related regulations are
the Convention on the International Regulations for Preventing Collisions at Sea
76 A. Amro et al.
In order for the APS to operate safely, it must support functions that include
navigation, machinery and passenger management, and communications. In this
paper we focus on the communication functions and cybersecurity considera-
tions for the APS to perform its intended functions, with an increased focus on
navigation. DNV GL discussed the navigation functions that are expected of a
vessel in autoremote operation [19]. These are listed below:
been identified for the APS to perform its intended functions: 1. External com-
munication including connection with the RCC and external systems and stake-
holders; 2. internal communication between on-board ship components; and 3.
communication with other vessels in the vicinity. This subsection discuss the
communication requirements for each communication category in addition to
general requirements that apply across all categories. Additionally, this subsec-
tion discusses cybersecurity requirements mapped to the relevant NIST frame-
work function as suggested by Bureau Veritas [15]. Each requirement in this
section is titled with a three level coding scheme. The first level is related to the
domain (communication (C) or Cybersecurity (S)). The second level is related
to the sub-domain. The communication sub-domains are external (X), internal
(N), with other ships (O) or general (G). The cybersecurity sub-domains are
identification (I), protection (P), detection (D), response and recovery (R). The
third level refers to the relative numbering of the requirement within its category.
Communication Requirements: This subsection discusses external and inter-
nal communication requirements, in addition to the communication with other
ships and other general communication requirements.
– External Communication
First, a dedicated physical space must be allocated separately from the con-
trolled vessel, which can be on the shore or on-board another ship. The
required level of reliability, availability, and security of the communication link
will increase with increased control of the RCC over the APS, depending on
the latter’s autonomy level. Additional communication with off-ship systems
is required. Examples of off-ship systems that are leveraged for operational
purposes are SSS, AtoN, VTS and RIS communication (cf Sect. 3.1). Addi-
tionally, other systems may require access to the ship’s systems, to provide
services such as maintenance, processing insurance claims, etc. Communica-
tion with external stakeholders is expected by the APS either by automated
systems on the vessel itself, or by the personnel on the RCC. The requirements
for the aforementioned communication are discussed below:
• C-X-1: The link’s minimum acceptable network latency and maximum
required bandwidth should be calculated, documented and implemented.
MUNIN provided minimum accepted requirements of latency and band-
width [34]. In total 4 Mbps accumulated link is considered the minimum
link bandwidth for ship to shore communication. The required bandwidth
is expected to be larger in the case of APS due to the implementation
of new technologies with high data requirements such as the lidar. For
instance, the targeted lidar for implementation in the Autoferry project
[1] requires local transfer rate between 9–20 Mbps. Although the amount
of data to be transmitted to the RCC is expected to be much less, in
case of an increased control of the RCC over the vessel, the full lidar data
might be expected for transmission. Additionally, the accepted latency
suggested by MUNIN ranges from 0.05 s for ship to ship communication
up to 2.5 s for HD video.
Connect and Protect: Requirements for MASS 79
• C-X-2: A dedicated, permanent and reliable link for emergency push but-
tons for passengers should exist. Such button should be used to indicate
passenger related emergency and is expected to initiate intervention of
the available ECT (cf Sect.3.1) in the area or to change the autonomy
level to provide the RCC full control of the APS if appropriate.
• C-X-3: The link with the RCC should be fault-tolerant so that it operates
at full capacity even in case of failure in a single component.
• C-X-4: Traffic in the link with the RCC should be prioritized according
to a pre-defined prioritization policy to enable traffic with higher priority
to be forwarded in case of reduced bandwidth. DNV GL suggested a pri-
oritization policy so that the traffic is prioritized in the following order,
from highest to lowest priority: 1. Control messages for emergency (e.g.
MRC activation); 2. commands for remote control of key vessel functions;
3. situational awareness data for remote control of key vessel functions;
4. supervision data; 5. maintenance data.
• C-X-5: The operator should be able to seamlessly switch and distribute
different vessel data between the different communication channels with-
out a negative effect on the operations e.g. situational awareness data on
one channel, the rest on another.
• C-X-6: Communication links should operate according to appropriate
QoS requirements and adapt with signal degradation. The QoS require-
ments are case dependent based on the implemented systems on board
the APS. For instance, a rule could be established that delay sensitive
systems (i.e. collision avoidance) should be carried through an appropri-
ate communication channel that provides the lowest delay whereas delay
tolerant systems (i.e. HD video) could be channeled through a communi-
cation channel with higher but still appropriate delay.
• C-X-7: The network should integrate monitoring and notification sys-
tems for real-time or near real-time link quality analysis, based on data
collection and aggregation subsystems which satisfy intrinsic and contex-
tual Quality of Information requirements to support such real-time/near
real-time situational awareness and incident response. The notification
functionality is expected to be integrated within the ship’s CAM.
• C-X-8: The operator should have independent troubleshooting capabili-
ties over each one of the communication links. Troubleshooting one link
should not interrupt the operations of another.
• C-X-9: Communication link with RCC should be established using
redundant communication channels, including main and backup chan-
nels, preferably using different communication technologies and service
providers. The communication architecture presented by MUNIN was
mainly focusing on deep sea operations. This entails the application of
satellite communication for carrying ship to shore operations as a primary
communication channel; this is different compared to inland or short sea
shipping such as the APS, where high communication requirements are
needed. In this case, mobile communication or Wi-Fi channels can be
primarily used [35].
80 A. Amro et al.
– Internal Communication
• C-N-1: The Communication network design should comply with the appli-
cable requirements in the relevant standards (cf Table 1).
• C-N-2: A Segregated network design should exist to avoid failure cascad-
ing. DNV GL suggested a specific network arrangement that applies net-
work segregation [19]. They suggested that the following systems should
not be connected to the same network: 1. Navigation system; 2.Communi-
cation system; 3. Machinery control and monitoring system; 4. Safety sys-
tems; 5. Control systems that serve redundant vessel services; 6. Auxiliary
systems not related to vessel key functions; 7. Other systems from differ-
ent system suppliers. Suggested network segmentation methods include
air-gap, VLAN, firewalls etc.
• C-N-3: A redundant network design should exist with automatic transi-
tion/activation/restoration between the main and backup system compo-
nents.
• C-N-4: It should be possible to divert connectivity to local resources upon
loss of remote resources. (e.g in case of distributed network or cloud ser-
vices providing data storage, backup local storage for critical data are
expected to be implemented).
• C-N-5: Connectivity to several systems on-board, such as passenger man-
agement system, alert system (CAM), log book, and local sensors should
exist. The passenger management system provides certain services to
the passengers on-board such as voice communication, trip status, and
internet-access. Local sensors may include weather sensors, positioning
sensors and others.
• C-N-6: If several wireless communication links are expected to operate
closely on-board with a risk of interference, a frequency coordination plan
should be made and documented and then tested on board.
– Communication with other vessels
• C-O-1: The APS should be able to communicate with other vessels. For
such communication, line of sight (LOS) communication system mainly
based on AIS or digital VHF with range of at least two kilometers should
be used. This communication includes position and route advertisement
which is essential for safe navigation and collision avoidance.
– General Communication Requirements
• C-G-1: Important communicated data should be recorded and logged to
be analyzed when needed. DNV GL proposed the minimum data that is
required to be recorded [19]: 1. The status of the vessel’s key functions
including the communication links; 2. Alerts; 3. Manual orders or com-
mands; 4. All input and output data to or from the decision support and
automation systems. In case the data is recorded on board, an early alert
should be raised in case storage capacity exceeds a certain threshold and
it should be possible for it to be transferred to shore.
• C-G-2: The network components and equipment should be type-approved
in compliance with the related certification policy. For technologies imple-
mented in autonomous vessels to be certified by DNV GL, type approval
Connect and Protect: Requirements for MASS 81
– Ship Registry: secure authentication controls should exist for ship certifi-
cation and revocation of certificates.
– IT Service Providers: controls regarding authorization and access control
should exist.
– System installation: controls to verify proper and secure systems installa-
tion according to a defined list of configuration parameters should exist.
– Maintenance: access to the system to provide software and/or hardware
maintenance services should be controlled, monitored, and verified.
– Financial services: controls should exist to protect processes related to
passengers payments.
– Insurance services: controls should exist to secure access or disclosure of
certain data in case of accidents.
Connect and Protect: Requirements for MASS 83
References
1. Autonomous all-electric passenger ferries for urban water transport. https://www.
ntnu.edu/autoferry
2. IACS Rec 164 - communication and interfaces - new Nov 2018. IACS
3. International association of classification societies. http://www.iacs.org.uk/
4. Nordic boat standard. https://www.sdir.no/en/guides/nordic-boat-standard/
5. Projects carried out by members of NFAS. http://bit.ly/NFASProjects
6. IACS Rec 158 - physical security of onboard computer based system - new Oct
2018. http://www.iacs.org.uk/download/8782
7. IACS Rec 159 - network security of onboard computer based systems - new Sept
2018. http://www.iacs.org.uk/download/8652
8. IMO takes first steps to address autonomous ships. http://bit.ly/IMOAutonomous
9. Tracking in public spaces. http://bit.ly/DatatilsynetTracking
10. Trondheim blir smartby. http://bit.ly/Trondheimkommune
11. Focus on risks 2018, November 2017. http://bit.ly/sdirRisks2018
12. An, K.: E-navigation services for non-solas ships. Int. J. e-Navig. Marit. Econ. 4,
13–22 (2016)
13. Andrés, S., Piniella, F.: Aids to navigation systems on in land waterways as an
element of competitiveness in ULCV traffic. Int. J. Traffic Transp. Eng. 7(1) (2017)
84 A. Amro et al.
14. ANSSI: Information systems defence and security: France’s strategy (2011)
15. Bureau Veritas: Guidelines for autonomous shipping (2017). http://bit.ly/
BureauVeritas641NI2017
16. TMS Committee: Maritime cyber risk management in safety management systems
(2017)
17. Danish Maritime Authority: Analysis of regulatory barriers to the use of
autonomous ships. Danish Maritime Authority, Final Report, December 2017
18. Datatilsynet: The anonymisation of personal data. http://bit.ly/Datatilsy
netAnonymisation
19. DNV GL: DNVGL-CG-0264: Autonomous and remotely operated ships (2018)
20. DNV GL: DNVGL-CP-0231: Cyber security capabilities of control system compo-
nents (2018)
21. DNV GL - Maritime: Remote-controlled and autonomous ships position paper
(2018)
22. Havdal, G., Heggelund, C.T., Larssen, C.H.: Design of a small autonomous pas-
senger Ferry. Master’s thesis, NTNU (2017)
23. Komianos, A.: The autonomous shipping era. Operational, regulatory, and quality
challenges. TransNav: Int. J. Mar. Navig. Saf. Sea Transp. 12 (2018)
24. L
ebkowski, A.: Design of an autonomous transport system for coastal areas.
TransNav: Int. J. Mar. Navig. Saf. Sea Transp. 12 (2018)
25. Levander, O., Marine, R.R.: Ship intelligence-a new era in shipping. In: The Royal
Institution of Naval Architects, Smart Ship Technology, International Conference
Proceedings, pp. 26–27 (2016)
26. MI News Network: Rolls-Royce and Finferries demonstrate world’s first fully
autonomous ferry, December 2018. http://bit.ly/marineinsightRollsRoyce
27. Norwegian Shipowners’ Association: Maritime outlook 2018. Technical report, Nor-
wegian Shipowners’ Association (2018)
28. Olsen, S.: Autonom ferge ballstadlandet. http://bit.ly/lofotenmatpark
29. IM Organization: Convention on the international regulations for preventing colli-
sions at sea. COLREGS (1972)
30. Patraiko, D.: The development of e-navigation. TransNav Int. J. Mar. Navig. Saf.
od Sea Transp. 1(3) (2007)
31. Porathe, T., Burmeister, H.C., Rødseth, Ø.J.: Maritime unmanned navigation
through intelligence in networks: the MUNIN project. In: 12th International Con-
ference on Computer and IT Applications in the Maritime Industries, COMPIT
2013, Cortona, 15–17 April 2013, pp. 177–183 (2013)
32. Rødseth, Ø., Burmeister, H.: MUNIN deliverable d10.1: impact on short sea ship-
ping (2015). http://www.unmanned-ship.org/munin/wp-content/uploads/2015/
10/MUNIN-D10-1-Impact-on-Short-Sea-Shipping-MRTK-final.pdf
33. Rødseth, Ø., Nordahl, H.: Definitions for autonomous merchant ships. In: Norwe-
gian Forum for Unmanned Ships (2017)
34. Rødseth, Ø.: Munin deliverable 4.3: evaluation of ship to shore communication links
(2012). http://www.unmanned-ship.org/munin/wp-content/uploads/2014/02/d4-
3-eval-ship-shore-v11.pdf
35. Rødseth, Ø.J., Kvamstad, B., Porathe, T., Burmeister, H.C.: Communication
architecture for an unmanned merchant ship. In: 2013 MTS/IEEE OCEANS-
Bergen, pp. 1–9. IEEE (2013)
36. Rødseth, Ø.J., Tjora, Å.: A system architecture for an unmanned ship. In: Pro-
ceedings of the 13th International Conference on Computer and IT Applications
in the Maritime Industries (COMPIT) (2014)
Connect and Protect: Requirements for MASS 85
37. Sedgewick, A.: Framework for improving critical infrastructure cybersecurity, ver-
sion 1.1. Technical report, National Institute of Standards and Technology (2019)
38. Sikora-Fernandez, D., Stawasz, D., et al.: The concept of smart city in the theory
and practice of urban development management. Rom. J. Reg. Sci. 10(1), 86–99
(2016)
39. SINTEF: Test site opens for unmanned vessels. http://bit.ly/sintefTestSites
40. Skille, A., Lorentzen, S.: Foreslår førerløs passasjerferge i Trondheim. http://bit.
ly/nrkTrondheim
41. Yoon, I.: Technology assessment - autonomous ships, September 2018. https://doi.
org/10.13140/RG.2.2.36778.88009
Simulation-Based Evaluation of DDoS Against
Smart Grid SCADAs
Abstract. The goal of this paper is to simulate the effects of different Dis-
tributed Denial of Service (DDoS) attack scenarios which might be launched
against smart grid Supervisory Control and Data Acquisition systems, i.e.
SCADAs. We will analyze attacks which are launched from compromised
Remote Terminal Units (RTUs) located in the process environment. We created
an ICS testbed and industrial DDoS simulator environment consisting of a single
C&C server and a configurable number of bots. We simulated scenarios with
different numbers of hacked RTUs trying to overwhelm the SCADA with
unwanted messages. We analyzed the effects of DDoS-type attacks against
SCADAs with different internal queue architectures used to manage the
incoming messages, i.e. no queues, single queue and separate queue for each
connected RTU.
1 Introduction
2 State-of-the-Art Review
In our review of the state-of-the-art we analyze the latest developments in ICS security,
denial of service attacks in ICS, as well as ICS security simulation testbeds.
and mostly focus on ICS instead of the SCADA only. Reference [16] surveys the latest
cybersecurity risk measurement methods and research in ICS. Attack detection is dis-
cussed in [24]. More specifically, the authors present a mathematical framework,
monitoring limitations and centralized and distributed attack detection and identifica-
tion. The Distributed Network Protocol’s security features are assessed in [25].
testbeds [10] usually had a simple system architecture not yet aligned with the IEC
62443 layered security model [14]. The PowerCyber testbed was designed to closely
resemble real-life power grid communications and was therefore equipped with field
devices and SCADA software. It was designed to support both education and research
purposes [11]. The authors of a 2015 survey on ICS testbeds explains that the extreme
availability requirements of real-life ICS make it challenging to conduct security tests
in practice, i.e. on systems already deployed in industrial settings. They argue that this
is the reason why researchers and practitioners turn to ICS testbeds and go on to
identify and asses 30 such testbeds [12]. An open, virtual testbed for ICS is presented in
[28]. The authors of [27] build a modular testbed, mimicking a SCADA network and
controlling real sensors and actuators via the Modbus protocol. The authors use their
testbed to evaluate the effects of different DDoS scenarios on SCADA operations.
An ICS cybersecurity testbed written in Python and consisting of virtual devices and
process de-vices is presented in [28]. The Smart City Testbed (SCT) developed at the
Washington State University (WSU) was used to analyze potential cyber intrusions and
anomaly detection systems (ADS) in substations, which are key components of smart
grids [30]. The scalable and reconfigurable, virtual SCADA security testbed presented
in [31] was used to evaluate the effects a DDoS-type false data injection attack would
have on normal SCADA operation.
3 Problem Definition
The primary goal of this research is to develop an ICS testbed for investigating the
potential effects of DDoS attacks originating from the process (i.e. operational tech-
nology - OT) environment. The secondary goal is to investigate the impact of DDoS
attacks against different SCADA queue management solutions.
It was the authors’ intention to investigate the extent of impact a few hijacked
RTUs could have on the operational performance of complex ICS, which consisted of
additional services consuming data forwarded by the SCADA. It was planned to
specifically analyze the behavior of those services, which require high CPU, RAM
and/or storage capacities. One such service is the calculation engine present in electric
power systems, which runs energy and/or distribution management system functions
(EMS/DMS), e.g. topology analysis, load flow, contingency analysis, short circuit
calculations.
4 System Architecture
The key elements of the ICS environment presented in Fig. 1 were the following:
• The Supervisory Control and Data Acquisition (SCADA) service collects, trans-
forms and visualizes data from the process environment and allows operators to
issue commands.
• The Network Model Service (NMS) is the single source of truth for the electric
power system’s network model, which it serves to the other services whose func-
tionality depends on having up-to-date network model information.
• The Weather Forecast Service (WFS) collects and serves weather forecast infor-
mation to the other elements of the system.
• The Calculation Engine (CE) extends the SCADA with network optimization
functionality, e.g. topology analysis, load flow, contingency analysis, short circuit
calculations. It relies on data collected from the NMS, SCADA and WFS.
• The Time Series Database (TSDB) stores and provides access to historical values,
e.g. all past measurements, commands and network model changes.
• Remote Terminal Units (RTU) are industrial computers deployed inside the process
environment, which collect measurement values from sensors and forward control
commands to actuators.
• The User Interface (UI), or the Human Machine Interface allows the operators to
monitor and control the underlying physical process.
2. Single message queue. This architecture assigns one thread to each RTU to accept
messages, which are stored in a single, internal message queue. When the RTU
threads put the received messages into the single queue, they notify a single mes-
sage processing thread to take over and process the content.
3. Multiple message queues. SCADA solutions in this group assign a separate mes-
sage queue and two threads to each RTU (Fig. 3). One thread is tasked to receive
messages, and the other does the data processing.
5 Experiments
During the test all ICS services operated regularly and there were no visualization
delays in the SCADA human-machine interface (HMI) and it was available to receive
and forward (operator-issued) commands. All messages were processed by the SCADA
service in a timely manner. All values were forwarded to the Historian and stored. The
calculation engine executed all calculations in a timely manner.
We analyzed the logs of the SCADA with a single message queue and observed
that its message queue of messages waiting for processing contained 1000 messages in
the 10th min of the simulation. After one hour, there were 7800 messages in the
message queue, i.e. all those messages were delayed. None of the services failed, but
the values were visualized with delay in the SCADA HMI. The length of the message
queue was sufficient, so there were no messages lost. All messages were archived in the
Historian.
The behavior of the SCADA with multiple message queues was similar to the
single-queue architecture. Message processing delays were shorter, as we observed
maximum queue length of 2800 at the end of the 5-hour simulation run.
Even the multiple-queue SCADA failed to properly respond to the data loads
generated in the 5-bot simulation scenario. At the end of the 5-hour simulation, the
messages queues handling the bots were up to 80,000 message-long. CPU load rose by
14\%. The calculation engine failed after 2 h, similarly to the previous scenario. It was
manually restarted, and the simulation run was not interrupted.
6 Conclusion
This paper presents an Industrial Control System (ICS) testbed and DDoS simulator
used for investigating the potential effects of attacks originating from one or more
manipulated Remote Terminal Units (RTUs) in the process environment.
Simulation-Based Evaluation of DDoS Against Smart Grid SCADAs 95
The presented testbed consisted of RTUs, SCADA and other services running in
the operational technology (OT) environment. The DDoS simulator had a centralized
system architecture consisting of a single command and control server and a config-
urable number of bots. The bots were designed to impersonate regular RTUs during the
RTU to SCADA reconnect phase. The ICS communication protocol used was DNP3.
The experiments conducted by the authors suggest that the presence of a single
manipulated or malfunctioning RTU has significant negative impact on the operational
capabilities of ICS. The negative impact is more pronounced for simple SCADA
architectures without an adequate message queuing solution. We showed that a
SCADA with no message queuing capabilities is incapable to respond to message
flooding from even a single RTU. Even a SCADA with a dedicated message queue for
each connected RTU would struggle if it would be bombarded from multiple manip-
ulated RTUs, as shown in our 5-bot experiment.
Additionally, we showed that even if the SCADA was capable to process the
unwanted messages resulting from a DDoS attack, the services consuming real-time
data received from the SCADA would eventually fail. In our experiments the calcu-
lation engine running power system analyses (e.g. topology analysis, load flow, con-
tingency analysis) was usually the first to fail.
As the DDoS simulator relies on a well-known ICS communication protocol, it can
be potentially modified, extended and used in the simulation of other attack scenarios.
Additionally, it can be extended with more elaborate DDoS traffic generation schemes
which could more easily avoid detection by firewalls and/or intrusion detection
systems.
The authors intend to extend this research and incorporate the results into a real-life
cyber-physical testbed, consisting of a mix of virtual and physical components, e.g.
sensors and actuators.
Acknowledgment. This work was supported by the Ministry of Education, Science and
Technological Development of the Republic of Serbia under grants III-42004 and TR33013. The
authors received funding from the EU’s Education, Audiovisual and Culture Executive Agency
(EACEA) under the Erasmus+ Capacity Building in the Field of Higher Education (CBHE) grant
number 586474-EPP-1-2017-1-RS-EPPKA2-CBHE-JP.
References
1. Amin, S., Cárdenas, A.A., Sastry, S.S.: Safe and secure networked control systems under
denial-of-service attacks. In: Majumdar, R., Tabuada, P. (eds.) HSCC 2009. LNCS, vol.
5469, pp. 31–45. Springer, Heidelberg (2009). https://doi.org/10.1007/978-3-642-00602-9_3
2. Ashok, A., Wang, P., Brown, M., Govindarasu, M.: Experimental evaluation of cyber
attacks on automatic generation control using a CPS security testbed. In: 2015 IEEE Power
& Energy Society General Meeting, pp. 1–5. IEEE (2015)
3. Barbosa, R.R.R., Pras, A.: Intrusion detection in SCADA networks. In: Stiller, B., De Turck,
F. (eds.) AIMS 2010. LNCS, vol. 6155, pp. 163–166. Springer, Heidelberg (2010). https://
doi.org/10.1007/978-3-642-13986-4_23
96 D. Gogic et al.
4. Chandia, R., Gonzalez, J., Kilpatrick, T., Papa, M., Shenoi, S.: Security strategies for
SCADA networks. In: Goetz, E., Shenoi, S. (eds.) ICCIP 2007. IIFIP, vol. 253, pp. 117–131.
Springer, Boston (2008). https://doi.org/10.1007/978-0-387-75462-8_9
5. Chikuni, E., Dondo, M.: Investigating the security of electrical power systems SCADA. In:
IEEE AFRICON 2007, Windhoek, South Africa, pp. 1–7. IEEE (2007)
6. Christiansson, H., Luiijf, E.: Creating a European SCADA security testbed. In: Goetz, E.,
Shenoi, S. (eds.) ICCIP 2007. IIFIP, vol. 253, pp. 237–247. Springer, Boston (2008). https://
doi.org/10.1007/978-0-387-75462-8_17
7. Davis, C.M., Tate, J.E., Okhravi, H., Grier, C., Overbye, T.J., Nicol, D.: SCADA cyber
security testbed development. In: 2006 38th North American Power Symposium, pp. 483–
488. IEEE (2006)
8. Ericsson, G.N.: Cyber security and power system communication—essential parts of a smart
grid infrastructure. IEEE Trans. Power Deliv. 25(3), 1501–1507 (2010)
9. Fernandez, J.D., Fernandez, A.E.: SCADA systems: vulnerabilities and remediation.
J. Comput. Sci. Coll. 20(4), 160–168 (2005)
10. Giani, A., Karsai, G., Roosta, T., Shah, A., Sinopoli, B., Wiley, J.: A testbed for secure and
robust SCADA systems. SIGBED Rev. 5(2), 4 (2008)
11. Hahn, A., et al.: Development of the PowerCyber SCADA security testbed. In: Proceedings
of the Sixth Annual Workshop on Cyber Security and Information Intelligence Research,
p. 21. ACM (2010)
12. Holm, H., Karresand, M., Vidström, A., Westring, E.: A survey of industrial control system
testbeds. In: Buchegger, S., Dam, M. (eds.) Nordic Conference on Secure IT Systems, vol.
9417, pp. 11–26. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-26502-5_2
13. Igure, V.M., Laughter, S.A., Williams, R.D.: Security issues in SCADA networks. Comput.
Secur. 25(7), 498–506 (2006)
14. International Electrotechnical Commission (IEC): Industrial communication networks -
Network and system security - Part 3-3: System security requirements and security levels.
IEC 62443-3-3, Geneva, Switzerland (2013)
15. Johnson, R.E.: Survey of SCADA security challenges and potential attack vectors. In: 2010
IEEE International Conference for Internet Technology and Secured Transactions, London,
United Kingdom, pp. 1–5. IEEE (2010)
16. Knowles, W., Prince, D., Hutchison, D., Disso, J.F.P., Jones, K.: A survey of cyber security
management in industrial control systems. Int. J. Crit. Infrastruct. Prot. 9, 52–80 (2015)
17. Kushner, D.: The real story of Stuxnet. IEEE Spectr. 3(50), 48–53 (2014)
18. Lee, R.M., Assante, M.J., Conway, T.: Analysis of the Cyber Attack on the Ukrainian Power
Grid. Defense Use Case, SANS ICS (2016)
19. Liu, S., Liu, X.P., El Saddik, A.: Denial-of-Service (DoS) attacks on load frequency control
in smart grids. In: 2013 IEEE PES Innovative Smart Grid Technologies Conference (ISGT),
Washington DC, USA, pp. 1–6. IEEE (2013)
20. Long, M., Wu, C.H., Hung, J.Y.: Denial of service attacks on network-based control
systems: impact and mitigation. IEEE Trans. Ind. Inform. 1(2), 85–96 (2005)
21. Nazir, S., Patel, S., Patel, D.: Assessing and augmenting SCADA cyber security: a survey of
techniques. Comput. Secur. 70, 436–454 (2017)
22. Nicholson, A., Webber, S., Dyer, S., Patel, T., Janicke, H.: SCADA security in the light of
Cyber-Warfare. Comput. Secur. 31(4), 418–436 (2012)
23. Pang, Z.H., Liu, G.P., Dong, Z.: Secure networked control systems under denial of service
attacks. IFAC Proc. Vol. 44(1), 8908–8913 (2011)
24. Pasqualetti, F., Dörfler, F., Bullo, F.: Attack detection and identification in cyber-physical
systems. IEEE Trans. Autom. Control 58(11), 2715–2729 (2013)
Simulation-Based Evaluation of DDoS Against Smart Grid SCADAs 97
25. Patwardhan, M.: DNP3: security and scalability analysis. Sacramento. California State
University (2012)
26. Qassim, Q.S., Jamil, N., Daud, M., Hasan, H.C.: Towards implementing scalable and
reconfigurable SCADA security testbed in power system environment. Int. J. Crit.
Infrastruct. 15(2), 91–120 (2019)
27. Queiroz, C., Mahmood, A., Hu, J., Tari, Z., Yu, X.: Building a SCADA security testbed. In:
Third IEEE International Conference on Network and System Security, Gold Coast,
Queensland, Australia, pp. 357–364. IEEE (2009)
28. Reaves, B., Morris, T.: An open virtual testbed for industrial control system security
research. Int. J. Inf. Secur. 11(4), 215–229 (2012)
29. Sommestad, T., Ericsson, G.N., Nordlander, J.: SCADA system cyber security—a
comparison of standards. In: IEEE PES General Meeting, Minneapolis, Minnesota, USA,
pp. 1–8. IEEE (2010)
30. Sun, C.C., Hahn, A., Liu, C.C.: Cyber security of a power grid: state-of-the-art. Int. J. Electr.
Power Energy Syst. 99, 45–56 (2018)
31. Tesfahun, A., Bhaskari, D.L.: A SCADA testbed for investigating cyber security
vulnerabilities in critical infrastructures. Autom. Control Comput. Sci. 50(1), 54–62 (2016)
32. Wang, C., Fang, L., Dai, Y.: A simulation environment for SCADA security analysis and
assessment. In: 2010 IEEE International Conference on Measuring Technology and
Mechatronics Automation, Changsha, China, pp. 342–347. IEEE (2010)
33. Yuan, Y., Zhu, Q., Sun, F., Wang, Q., Başar, T.: Resilient control of cyber-physical systems
against denial-of-service attacks. In: 2013 6th International Symposium on Resilient Control
Systems (ISRCS), San Francisco, USA, pp. 54–59. IEEE (2013)
Identifying Safety and Human Factors
Issues in Rail Using IRIS and CAIRIS
1 Introduction
1
https://cairis.org.
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 98–107, 2020.
https://doi.org/10.1007/978-3-030-42048-2_7
Identifying Safety and Human Factors Issues in Rail Using IRIS and CAIRIS 99
A key contribution of this work is the use of Human Factors Analysis and
Classification System (HFACS) to augment IRIS framework and CAIRIS plat-
form to identify safety and security issues. This helps rail stakeholders better
understand the safety and human factors implications of security concerns, and
also helps discover inter-dependencies between security, safety and human factors
engineering techniques.
In Sect. 2, we describe the related work upon which our approach is based,
followed by the explanation of our approach in Sect. 3. We illustrate this app-
roach with a case study example in Sect. 4, before concluding and discussing
future directions for our work in Sect. 5.
2 Related Work
The rail infrastructure has long been managed in accordance with health and
safety standards, working within legislative requirements such as in United King-
dom the Railway Act 2005, under guidance and supervision from bodies like Rail-
way Safety and Standards Board (RSSB) and Office of Rail Regulation (ORR).
More recently, the shift to digitalisation stipulated by the European Railway
Traffic Management System (ERTMS) imposed by European Union has seen
the introduction of the Common Safety Method for Risk Evaluation and Assess-
ment (CSM-REA) in addition to UK specific safety concepts such as ‘As Low
As Reasonably Practicable’ (ALARP) in managing safety risks.
The evolving nature of the cyber threats have imposed a greater challenge
for security experts in rail [12]. As a result, the rail infrastructure needs to be
supported by codes of practice (CoPs) throughout its life cycle as a combination
of security and safety [6]. Security should be infused with safety at a design
phase by ensuring a combined risk assessment approach.
Similarly, the strong linkage between the human intent to violate rules and
imposed safety hazards described by [3] highlights the value of combining safety
with human factors. The Human Factors Analysis and Classification System
(HFACS) is a framework for eliciting possible accident and incident contribution
factors based on taxonomy of active and latent failures caused by human inter-
actions in rail [17]. The HFACS have been used by rail stakeholders to determine
the human error sources behind accidents and incidents. However, to date, there
has been no work on how it can be used to consider safety or security attributes
of rail system.
Hazards and accidents may occur due to security breaches, and dependability
– delivering services that can justifiably trusted – encompasses safety and some
major elements of security [5]. Safety is an attribute of dependability, with avail-
ability, reliability, integrity and maintainability; security refers to the availability
100 A. Altaf et al.
and integrity attributes and to confidentiality [13]. Thus the risk factors (prob-
ability of chances of damage) along with the dependability (trust and reliance
on system) are triggered by safety and security issues. Both safety and security
engineering communities are now working to better bridge their communities
[11], e.g. safety engineering consideration of security mindedness [6].
Previous work has considered human error as an intersecting concept between
cyber security and safety. Humans may cause harm by making mistakes (active
failures) or by inducing errors within system (latent failures) [7], with human
intent as a differentiating factor. If humans are benevolent (unintentional), they
may alert the safety engineers by causing hazards and accidents; if malevolent
(intentional), they may carry out threats and exploit vulnerabilities that com-
promise system security [16], thereby leading to a risk instigating a safety hazard.
The core IRIS concepts are illustrated in the UML class diagram in Fig. 1.
Vulnerabilities and threats contribute to potential risks, and threats are contin-
gent on attacker’s intent. This intent helps analysts identify the tasks and goals
they carry out or exploit, which can help determine human factors issues in
the form of human errors (active failures). Consequently, although not explicitly
Identifying Safety and Human Factors Issues in Rail Using IRIS and CAIRIS 101
designed with safety in mind, IRIS provides a foundation for integrating security,
safety and human factors.
3 Approach
We have devised an approach based on the IRIS framework, which leverages
security and usability engineering approaches to better understand the safety
implications of rail infrastructure under design. This approach is tool-supported
using CAIRIS. The approach takes input from security and human factors engi-
neers, as well as from rail stakeholders with safety expertise.
from a security engineer’s model of the same system. Attacker’s motivation and
capabilities play an important role in threat identification. Tasks and goals ful-
filled by attackers also provide an insight during threat modelling. The threats
identified are assigned security properties based on the goals of attacker.
Based on asset modelling and risk analysis, the concerned tasks and goals are
elicited. These form the basis of system and user level goals. Tasks and goals are
identified from the attacker’s perspective and also form the basis for obstacles
that model obstructions to system goals. Goal and task models can help the
security engineers to better understand the system threat model.
The risk model generated by CAIRIS determines the safety hazards, by showing
the linkage between the assets with their associated security attributes, vulner-
abilities, emergent threats and the possible risks. The main purpose of this type
of modelling is to identify the possible safeguards to be taken and minimise the
chances of occurrence of any hazardous events.
of passenger safety2 . The 2008 incident was logged as School Boy Hacks into
Polish Tram System in the ‘Repository of Industrial Security Incidents’ [2].
We gathered open source intelligence as an input to our approach. This
was based on several online articles written about the particular Polish Tram
Incident. We supplemented publicly available data with the Operational Con-
cept for European Railway Traffic Management System (ERTMS); this was
used to understand the system architecture, application levels, operating modes,
signalling principles and control. We also obtained feedback on the emerging
CAIRIS model from safety and human factors experts at Ricardo, who were
representative of the rail stakeholder that might provide input to our approach.
The completion of these tasks lead to the satisfaction of system goals (Modify
TV Remote Control, Access Railway Network and Redirect Railway Trams) on
the part of attacker.
The attack was conducted by exploiting system loop-holes. The exploitation
of these loop-holes were active failures on the part of security engineers. For
example, the vulnerability Reported Problems with Signalling System led to the
human factors issue of Violations as the operation and performance of signalling
system was not compliant with secured protocols and standards. This allowed
the attacker to perform the task of Record Signals, fulfilling the system goal
Access Railway Network. In this case, the major security goal defined by security
engineers which would have acted as an obstacle for attacker would have been
the use of Advanced Train Control Protocol System which would have denied
Adam an unauthorised access into the railway network. Thus, it would have
mitigated the cyber-attack, and ensured the safety of passengers.
Table 1 shows how the vulnerabilities, threats and risks identified can be cate-
gorised to determine the human factors issues based on HFACS along with safety
hazards. These human factors issues also help us to verify the system usability
for risks, by the satisfaction of user goals depending on certain procedures, com-
petencies, permissions and training needs analysis (TNA) to achieve those goals
and complete defined tasks.
For example, the risk of Injury of Railway Staff or Passenger which is linked
to threat of Poland Railway Network Intrusion, may lead to safety hazard of Loss
of Life. In this case, the human factors issue observed using the HFACS frame-
work is the poor design of Technological Environment due to Poor Architectural
Design and Lack of Risk Assessment, which has life-threatening consequences.
This illustrates how the timely evaluation of technological environment using
checklists and task factors can minimise the chances of risk occurrence.
106 A. Altaf et al.
Acknowledgements. The work described in this paper was funded by the BU stu-
dentship Integrating Safety, Security, and Human Factors Engineering in Rail Infras-
tructure Design & Evaluation. We are also grateful to Ricardo for their support.
References
1. Randstad Rail - Generic Risk Assessment Log. https://www.randstad.co.uk
2. RISI - The Repository of Industrial Security Incidents (2008). https://www.
risidata.com
3. Alper, S.J., Karsh, B.T.: A systematic review of safety violations in industry. Accid.
Anal. Prev. 41(4), 739–754 (2009)
4. Atzeni, A., Cameroni, C., Faily, S., Lyle, J., Flechais, I.: Here’s Johnny: a method-
ology for developing attacker personas. In: 2011 Sixth International Conference on
Availability, Reliability and Security, pp. 722–727. IEEE, Vienna (August 2011)
5. Avizienis, A., Laprie, J.C., Randell, B., Landwehr, C.: Basic concepts and taxon-
omy of dependable and secure computing. IEEE Trans. Dependable Secure Com-
put. 1(1), 11–33 (2004)
6. Bloomfield, R., Bishop, P., Butler, E., Stroud, R.: Security-informed safety: sup-
porting stakeholders with codes of practice. Computer 51(8), 60–65 (2018)
7. Brostoff, S., Sasse, M.A.: Safe and sound: a safety-critical approach to security, p.
10 (2001)
8. Faily, S.: Designing Usable and Secure Software with IRIS and CAIRIS. Springer,
Cham (2018). https://doi.org/10.1007/978-3-319-75493-2
9. Faily, S., Flechais, I.: User-centered information security policy development in a
post-stuxnet world. 2011 Sixth International Conference on Availability, Reliability
and Security, pp. 716–721. IEEE, Vienna (August 2011)
10. Gollmann, D.: Computer Security, 2nd edn. Wiley, Hoboken (2007)
11. Jonsson, E., Olovsson, T.: On the integration of security and dependability in
computer systems, p. 6 (1998)
12. Boudi, Z., Koursi, E.M.E., Ghazel, M.: The new challenges of rail security. French
Institute of Science and Technology for Transport, Development, and Networks
IFSTTAR-COSYS-ESTAS (2016). University of Lille Nord de France, F-59000
Lille, Villeneuve d’Ascq, France. Journal of Traffic and Logistics Engineering
13. Piètre-Cambacédès, L., Bouissou, M.: Cross-fertilization between safety and secu-
rity engineering. Reliab. Eng. Syst. Saf. 110, 110–126 (2013)
14. Reason, J.: Human Error. Cambridge University Press, Cambridge (1990)
15. Wiegmann, D.A., Shappell, S.A.: A Human Error Approach to Aviation Accident
Analysis: The Human Factors Analysis and Classification System, 1st edn. Rout-
ledge, Aldershot (2003)
16. Young, W., Leveson, N.G.: An integrated approach to safety and security based
on systems theory. Commun. ACM 57(2), 31–35 (2014)
17. Zhou, J.L., Lei, Y.: Paths between latent and active errors: analysis of 407 railway
accidents/incidents’ causes in China. Saf. Sci. 110, 47–58 (2018)
SECPRE Workshop
How Not to Use a Privacy-Preserving
Computation Platform: Case Study
of a Voting Application
Jan Willemson1,2(B)
1
Cybernetica AS, Ülikooli 2, 51003 Tartu, Estonia
[email protected]
2
STACC, Ülikooli 2, 51003 Tartu, Estonia
1 Introduction
Data processing is a field offering both threats and opportunities. On one hand,
having access to larger amount of high-precision data allows to take better-
informed policy decisions and as a result increase the quality of life of the whole
society. On the other hand, having access to personal information may give rise
to malicious profiling, manipulation, blackmailing or other types of misuse.
Thus a balance is required between the two extremes of making all data
public (destroying individual privacy) and closing all data up (destroying data
utility).
One possible equilibrium is provided by secure computation mechanisms
that allow to generate aggregate results while still protecting the individual
records. Being originally proposed in early 1980s, the respective methods have
evolved over the decades, resulting in a number or practically applicable frame-
works. Some of the currently actively developed examples include Sharemind [7],
SPDZ [9], ObliVM [21], Chameleon [26], etc.
Unfortunately, building an application on top of such a framework does not
yet guarantee that the application itself is secure. There are many places where
things can go wrong if done carelessly.
A tempting area to deploy secure computation mechanisms is electronic vot-
ing. At the first sight, its problem setting closely resembles the one of secure
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 111–121, 2020.
https://doi.org/10.1007/978-3-030-42048-2_8
112 J. Willemson
computation. A potentially large number of voters each have their private input
(political preference), and their joint interest is to compute an aggregate end
result (the tally).
The major approach to join the two worlds has been using homomorphic
encryption which has been studied and developed since 1980s [3,4]. However,
this approach is quite limited in the choice of voting protocols it is able to
implement. Since homomorphic encryption allows performing only one kind of
protected operation (say, addition), the resulting protocols can not go much
further from simple vote counting. Implementing more involved paradigms like
preferential voting or supporting more complex ballots becomes very challenging.
Deploying a fully-fledged secure computation framework as a basis for a flex-
ible electronic voting solution is a natural idea. However, until recently there
have been only a few incomplete attempts in this direction. In [13], Gang merely
states the basic idea, and in [24], Nair et al. implement a simple Java applica-
tion to add secret-shared votes. Gjøsteen and Strand take a different approach
utilising recent advances in fully homomorphic encryption [16]. However, fully
homomorphic encryption is still too inefficient to be applied on a large scale.
The first fully functional solution for secure computation based electronic vot-
ing was proposed and implemented by Dang-awan et al. in 2018 [10]. They have
built their proof-of-concept application on top of Sharemind1 [7], an established
secure computation platform with good development tools, including high-level
SecerC programming language for creating secure applications [6]. Regrettably,
Dang-awan et al. made a number of mistakes in several stages of design and
development, resulting in a completely insecure application.
In this paper, we will be going over their main flaws. Besides the direct
protocol-analytic value, we find the result also very educational for the developers
of both secure computation and electronic voting applications.
The paper is organised as follows. We begin with a short overview of the state
of the art in both secure computation and electronic voting in Sect. 2, followed
by a general overview of the system architecture of Dang-awan et al. [10] in
Sect. 3. Next we analyse both the voting and tallying processes in Sects. 4 and 5,
respectively. Finally, we draw some conclusions in Sect. 6.
1
https://sharemind.cyber.ee/.
How Not to Use a Privacy-Preserving Computation Platform 113
3 System Architecture
Remote electronic voting systems generally comprise of the following compo-
nents.
– Client software working in the voter’s environment and being responsible
for displaying the options, getting the voter preference and securing it (by
encrypting, signing and/or other means).
– Voting server being responsible for collecting and storing the votes (typi-
cally in a secured state).
– Tallying server is where the votes are opened (e.g. decrypted) and tabu-
lated.
To prove that the required security properties of the system hold, frequently
various auditing components are implemented in addition. In a more general
114 J. Willemson
sense, we can also consider operating systems, network connections, local legis-
lation, etc. to be part of the picture, but out treatment will not go into these
details.
Dang-awan et al. [10] start from the observation that securing the voting and
tallying servers is a critical prerequisite for a trustworthy remote electronic voting
system. Indeed, a breach in a server-side component has a potential to allow for
a large-scale vote manipulation attack to be implemented unnoticed [29].
In order to decrease the need to rely on a single server, Dang-awan et al.
propose to distribute the voting server between different parties and run secure
multi-party computation (SMC) protocols between them to achieve the required
functionality [10].
On the high level, a representative-based architecture is used [12] (see Fig. 1).
In this architecture, voters act as (input) parties submitting their votes in
a secret shared form to computation servers (also called nodes), of which there
are three in the standard configuration of Sharemind. Each server also has a
database back-end to store the shares, and a Node.js front-end to implement
communication routines with the other parties.
4 Voting
In the system proposed by Dang-awan et al., vote casting is implemented via
a Javascript library loaded into the browser. The library secret shares the vote
(even though this operation is repeatedly confused with encryption in [10]) and
sends the shares to Sharemind computing nodes.
The first problem we observe is that integrity of the votes is not protected by
signatures or any other strong cryptographic mechanism. Instead, the authors
propose a naı̈ve cast-as-intended verification protocol.
The vote (consisting of the voter ID, position ID and candidate ID) is check-
summed using CRC32, and this check sum is then hashed with SHA-1. The
resulting hash is displayed to the voter as a receipt. After the vote shares are
received by the Sharemind nodes, they perform the same operation and the
resulting SHA-1 hash is returned to the voter for comparison. If the comparison
succeeds, the voter should be assured of correct casting.
There are many flaws in this protocol. Perhaps the biggest conceptual prob-
lem is that all the communication between the voter and the central system
(including displaying the check-sums) is performed through a single web browser.
While this is definitely convenient, it creates a single point of attack. When the
browser gets compromised (a scenario that is unfortunately very much possible),
it can manipulate the displayed information arbitrarily. As a result, the voter
can not distinguish whether the hash-check-sum displayed to her really matches
the vote stored on the servers, or has it been maliciously changed before being
shown to her.
It is exactly for this reason that remote cast-as-intended verification needs
and independent channel. It may be implemented in various ways like the code
sheets plus SMS as in Norway [15], or using an independent auditing device as
in Estonia [17]. In any case it is clear that just relying on one medium for both
vote casting and verification is not sufficient.
Second, applying SHA-1 after CRC32 does not add any security as SHA-1 is
a deterministic function (but it does make the hashes longer, thus more difficult
to compare by a human).
Third, no random salt is added before hashing. This means that the pre-
images can be easily found by full inspection. To give some back-of-the-envelope
estimates, let’s consider the university student council election given as a use case
by the authors of [10]. There would probably be about few thousand voters (say,
up to 10,000), a few positions (say, about 10), and a few dozens of candidates
(say, up to 100).
All in all, a vote has in the order of magnitude 10 million possible values. Since
both CRC32 and SHA-1 are designed to be very fast to evaluate, pre-computing
116 J. Willemson
a table of 10 million hashes is an easy task. As all the hashes are put on a publicly
accessible bulletin board (basically a webpage) for verification purposes, anyone
with the pre-computed table can efficiently find out how everyone of the voters
voted. This is definitely something that a well-designed election system should
avoid in order to counter vote selling and other coercion attacks.
The double hashing proposed in [10] also leads to other problems. Note that
the output of CRC32 is just 32 bits long and
√
232 = 216 = 65536 .
/****************************************
app_save_vote.sc:
How sent encrypted values are saved
****************************************/
for (uint i = 0; i < size(candidateId); ++i) {
table = arrayToString(
declassify(electionId[i]));
tdbInsertRow(datasource, table,
{voterId, positionId[i],candidateId[i]});
print("ROW INSERTED");
message = bl_str(
arrayToString(declassify(voterId))
+arrayToString(declassify(positionId[i]))
+arrayToString(declassify(candidateId[i])));
hash[i] = CRC32(message); // hashes returned
}
}
5 Tally
The tallying procedure of elections is essentially a histogram computation, and
this Dang-awan et al. actually implement in a privacy-preserving manner. How-
ever, privacy is not the only requirement of the tally process. Perhaps even more
important is integrity, i.e. making sure that vote counting was not manipulated
by anyone.
The biggest problem in [10] is using Sharemind in its out-of-the-box, three-
server passive security mode. What passive security means here is the ability to
withstand an attacker who is only observing one of the computing nodes, but is
not trying to actively interfere with it. However, this model is too weak for the
voting use case.
Just to give a small illustrating example, recall that standard Sharemind
uses additive secret sharing [7], i.e. a value x ∈ Z232 is divided into shares
x1 , x2 , x3 ∈ Z232 so that
x1 + x2 + x3 = x mod 232 ,
where the share xi is held by the computing party Pi . What a malicious party
can do is e.g. increasing his share of one of the votes and decreasing another at
the same time, resulting in the same change in the values of the vote sums. This
would lead to a serious voting result trustworthiness violation as any computing
node would be able to manipulate it undetected.
Of course, tally integrity concerns are inherent in any election system. This is
why a large variety of approaches towards verification have been proposed in the
research community (see, e.g. [18] for a good overview). Ideally, tally correctness
should be checkable by everyone, or at least by a large number of designated
118 J. Willemson
independent auditors. The proposal by Dang-awan et al. does not foresee any of
such mechanisms.
Ironically, the ability to find pre-images of hashes displayed on the bulletin
board gives a way for anyone to compute the tally independently. However, this
happens with the price of total vote privacy loss, which is something we do not
want either.
If a multi-party computation engine like Sharemind is used to implement the
voting server, measures ensuring security against active manipulation attacks
should be deployed [11,20].
On top of that, modern electronic voting systems target software indepen-
dence, a state of affairs where verifying security properties of the system should
not rely on assumptions about the underlying software platform [27]. A practi-
cal way of approaching this target is requiring strong independently verifiable
cryptographic audit trail of all the critical operations.
In the case of vote tallying, there are two main approaches proposed and
implemented that can achieve this property. First, homomorphic tallying allows
combining votes under encryption so that the end result is the encryption of the
final count. Second, votes can also be directly decrypted giving non-interactive
zero-knowledge proofs of decryption. To facilitate independent auditing and pro-
tect vote privacy at the same time, mix-nets need to be applied in this case. We
refer to [5] for a recent overview on these techniques.
Secure computations on top of secret sharing can actually be implemented in
a homomorphic way. In fact, the tally routine of Dang-awan et al. makes use of
homomorphic properties of Sharemind’s additive secret sharing. But the missing
piece of the puzzle is a software-independent cryptographic trail that can be
verified for integrity by independent auditors. Developing such a component is a
necessary prerequisite for a voting system to be considered secure in late 2010-s.
6 Conclusions
Acknowledgments. The research leading to these results has received funding from
the Estonian Research Council under Institutional Research Grant IUT27-1 and the
European Regional Development Fund through the Estonian Centre of Excellence in
ICT Research (EXCITE) and the grant number EU48684.
References
1. Report of the National Workshop on Internet Voting: Issues and Research
Agenda (March 2001), Internet Policy Institute. https://www.verifiedvoting.org/
downloads/NSFInternetVotingReport.pdf
2. Archer, D.W., Bogdanov, D., Pinkas, B., Pullonen, P.: Maturity and performance
of programmable secure computation. IEEE Secur. Priv. 14(5), 48–56 (2016).
https://doi.org/10.1109/MSP.2016.97
3. Benaloh, J.C., Fischer, M.J.: A robust and verifiable cryptographically secure elec-
tion scheme (extended abstract). In: 26th Annual Symposium on Foundations of
Computer Science, Portland, Oregon, USA, October 21–23, 1985, pp. 372–382.
IEEE Computer Society (1985). https://doi.org/10.1109/SFCS.1985.2
4. Benaloh, J.C., Yung, M.: Distributing the power of a government to enhance the
privacy of voters (extended abstract). In: Halpern, J.Y. (ed.) Proceedings of the
Fifth Annual ACM Symposium on Principles of Distributed Computing, Calgary,
Alberta, Canada, August 11–13, 1986, pp. 52–62. ACM (1986). https://doi.org/
10.1145/10590.10595
120 J. Willemson
5. del Blanco, D.Y.M., Alonso, L.P., Alonso, J.A.H.: Review of cryptographic schemes
applied to remote electronic voting systems: remaining challenges and the upcom-
ing post-quantum paradigm. Open Math. 16(1), 95–112 (2018)
6. Bogdanov, D., Laud, P., Randmets, J.: Domain-polymorphic language for privacy-
preserving applications. In: Proceedings of the First ACM Workshop on Language
Support for Privacy-enhancing Technologies, PETShop 2013, pp. 23–26. ACM,
New York (2013). https://doi.org/10.1145/2517872.2517875
7. Bogdanov, D., Laur, S., Willemson, J.: Sharemind: a framework for fast privacy-
preserving computations. In: Jajodia, S., Lopez, J. (eds.) ESORICS 2008. LNCS,
vol. 5283, pp. 192–206. Springer, Heidelberg (2008). https://doi.org/10.1007/978-
3-540-88313-5 13
8. Cramer, R., Damgård, I.B., Nielsen, J.B.: Secure Multiparty Computation and
Secret Sharing. Cambridge University Press, Cambridge (2015)
9. Damgård, I., Pastro, V., Smart, N., Zakarias, S.: Multiparty computation from
somewhat homomorphic encryption. In: Safavi-Naini, R., Canetti, R. (eds.)
CRYPTO 2012. LNCS, vol. 7417, pp. 643–662. Springer, Heidelberg (2012).
https://doi.org/10.1007/978-3-642-32009-5 38
10. Dang-awan, R., Piscos, J.A., Chua, R.B.: Using Sharemind as a tool to develop an
internet voting system with secure multiparty computation. In: 2018 9th Interna-
tional Conference on Information, Intelligence, Systems and Applications (IISA),
pp. 1–7. IEEE (July 2018)
11. Eerikson, H., Orlandi, C., Pullonen, P., Puura, J., Simkin, M.: Use your brain!
Arithmetic 3PC for any modulus with active security. Cryptology ePrint Archive,
Report 2019/164 (2019). https://eprint.iacr.org/2019/164
12. Frikken, K.B.: Secure multiparty computation. In: Atallah, M.J., Blanton, M. (eds.)
Algorithms and Theory of Computation Handbook, Volume 2: Special Topics and
Techniques, pp. 14:1–14:16. CRC Press, Boca Raton (2009)
13. Gang, C.: An electronic voting scheme based on secure multi-party computation.
In: 2008 International Symposium on Computer Science and Computational Tech-
nology, vol. 1, pp. 292–294 (December 2008)
14. Gentry, C.: Fully homomorphic encryption using ideal lattices. In: Mitzenmacher,
M. (ed.) Proceedings of the 41st Annual ACM Symposium on Theory of Comput-
ing, STOC 2009, Bethesda, MD, USA, May 31–June 2, 2009, pp. 169–178. ACM
(2009). https://doi.org/10.1145/1536414.1536440
15. Gjøsteen, K.: The Norwegian Internet voting protocol. In: Kiayias, A., Lipmaa,
H. (eds.) Vote-ID 2011. LNCS, vol. 7187, pp. 1–18. Springer, Heidelberg (2012).
https://doi.org/10.1007/978-3-642-32747-6 1
16. Gjøsteen, K., Strand, M.: A roadmap to fully homomorphic elections: stronger
security, better verifiability. In: Brenner, M., et al. (eds.) FC 2017. LNCS, vol.
10323, pp. 404–418. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-
70278-0 25
17. Heiberg, S., Willemson, J.: Verifiable Internet voting in Estonia. In: Krimmer, R.,
Volkamer, M. (eds.) 6th International Conference on Electronic Voting: Verifying
the Vote, EVOTE 2014, Lochau/Bregenz, Austria, October 29–31, 2014, pp. 1–8.
IEEE (2014). https://doi.org/10.1109/EVOTE.2014.7001135
18. Jonker, H., Mauw, S., Pang, J.: Privacy and verifiability in voting systems: meth-
ods, developments and trends. Comput. Sci. Rev. 10, 1–30 (2013). https://doi.
org/10.1016/j.cosrev.2013.08.002
19. Krimmer, R.: The evolution of e-voting: why voting technology is used and how it
affects democracy. Ph. D. thesis, Tallinn University of Technology, doctoral Theses
Series I: Social Sciences (2012)
How Not to Use a Privacy-Preserving Computation Platform 121
20. Laud, P., Pankova, A., Jagomägis, R.: Preprocessing based verification of multi-
party protocols with honest majority. PoPETs 2017(4), 23–76 (2017). https://doi.
org/10.1515/popets-2017-0038
21. Liu, C., Wang, X.S., Nayak, K., Huang, Y., Shi, E.: ObliVM: a programming frame-
work for secure computation. In: 2015 IEEE Symposium on Security and Privacy,
SP 2015, San Jose, CA, USA, May 17–21, 2015, pp. 359–376. IEEE Computer
Society (2015). https://doi.org/10.1109/SP.2015.29
22. Madise, Ü., Martens, T.: E-voting in Estonia 2005. The first practice of country-
wide binding Internet voting in the world. In: Krimmer, R. (ed.) Electronic Voting
2006: 2nd International Workshop, Co-organized by Council of Europe, ESF TED,
IFIP WG 8.6 and E-Voting.CC, August 2–4, 2006, Castle Hofen, Bregenz, Aus-
tria. LNI, vol. 86, pp. 15–26. GI (2006). http://subs.emis.de/LNI/Proceedings/
Proceedings86/article4547.html
23. Martins, P., Sousa, L., Mariano, A.: A survey on fully homomorphic encryption: an
engineering perspective. ACM Comput. Surv. 50(6), 83:1–83:33 (2017). https://
doi.org/10.1145/3124441
24. Nair, D.G., Binu, V.P., Kumar, G.S.: An improved e-voting scheme using secret
sharing based secure multi-party computation (2015)
25. Puiggalı́, J., Cucurull, J., Guasch, S., Krimmer, R.: Verifiability experiences in
government online voting systems. In: Krimmer, R., Volkamer, M., Braun Binder,
N., Kersting, N., Pereira, O., Schürmann, C. (eds.) E-Vote-ID 2017. LNCS, vol.
10615, pp. 248–263. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-
68687-5 15
26. Riazi, M.S., Weinert, C., Tkachenko, O., Songhori, E.M., Schneider, T., Koushan-
far, F.: Chameleon: a hybrid secure computation framework for machine learning
applications. In: Proceedings of the 2018 on Asia Conference on Computer and
Communications Security, ASIACCS 2018, pp. 707–721. ACM, New York (2018).
https://doi.org/10.1145/3196494.3196522
27. Rivest, R.L.: On the notion of ‘software independence’ in voting systems. Philos.
Trans. R. Soc. A Math. Phys. Eng. Sci. 366(1881), 3759–3767 (2008)
28. Shamir, A.: How to share a secret. Commun. ACM 22(11), 612–613 (1979)
29. Springall, D., et al.: Security analysis of the Estonian Internet voting system. In:
Proceedings of the 2014 ACM SIGSAC Conference on Computer and Communi-
cations Security, pp. 703–715. ACM (2014)
30. Yao, A.C.: Protocols for secure computations (extended abstract). In: 23rd Annual
Symposium on Foundations of Computer Science, Chicago, Illinois, USA, Novem-
ber 3–5, 1982, pp. 160–164. IEEE Computer Society (1982). https://doi.org/10.
1109/SFCS.1982.38
A Proposed Privacy Impact Assessment
Method Using Metrics Based
on Organizational Characteristics
1 Introduction
As IT applications are steadily increasing, via the Internet, more and more people are
using them failing to estimate the positive, or even worse neglecting the potential
negative consequences. Thus, the major problem that arises is how companies can
protect both customers’ and employees’ personal data in order to avoid privacy
breaches [36]. This is an interesting field of study which should be thoroughly
researched due to the fact that a huge amount of personal information is collected,
stored, edited, communicated and published on the Internet [1, 6]. More specifically,
when using the Internet-based applications the users risk their privacy, since their
personal data may be exposed to others.
In order to prevent privacy breaches, several laws, standards, regulations and
directives [2] have been applied to most developed countries. The intent is to compel
organizations to fully inform their users and obtain their prior consent before collecting,
storing or processing their personal data in any way. At the same time, privacy
principles [3], privacy requirements [4, 5] and security requirements [4, 5], are also
helpful since they assist the development of an integrated security and privacy pro-
tection framework.
It is therefore necessary to have mechanisms that facilitate the assessment of the
impact of modern IT systems and applications on customers’ privacy. This assessment
should be carried out during the design stage of an IT system or application, and
definitely throughout its lifecycle. Scientists have been working towards this direction,
proposing methodologies that can assist organizations to measure the impact of privacy
breaches. However, there is still a definite lack of reliable metric systems that com-
panies can adopt.
Considering that there is still no effective method for the quantification of privacy
breach impact, the contribution of this paper is that it introduces an approach for
assessing the privacy level of a system. If the impact of a privacy breach is quantified,
the personal data can be effectively protected, thus supporting individuals against the
digital age challenges and helping them to minimize possible intrusions in their private
lives. An additional objective is to facilitate a uniform management of security and
privacy requirements in the sense that specific security countermeasures can also assist
in the satisfaction of privacy requirements. To this respect the proposed methodology
integrates the results of risk analysis methods (i.e. CRAMM, Octave, VSRisk etc.) in
terms of the criticality of the data assets and of their potential impact on the organi-
zation if a security incident (loss of confidentiality, integrity, availability) occurs.
The paper is organized as follows: Sect. 2 provides an overview of the literature on
privacy impact assessment methodologies. Based on the literature review, Sect. 3
provides an overview of the different types of data sets maintained by organizations,
together with the associated privacy principles and requirements. It continuous with the
description of the proposed Privacy Impact Method together with the derived metrics
for the quantification of the criticality of the privacy principles for an organization.
Section 4 draws the conclusions giving some pointers for future work.
2 Literature Review
In the advent of computer science era, individuals use computers on a daily basis to
satisfy their “digital needs”, for instance to perform electronic transactions via the net.
To do so, they do not hesitate to provide the personal data required for accessing the
applications. Yet, can people be really protected when they “offer” their personal data
so willingly? To answer this question, it is first necessary to estimate the consequences
from a potential privacy breach, employing a Privacy Impact Assessment (PIA)
method. Having estimated the impact, the stakeholders may adopt remedial actions for
eliminating or minimizing the consequences [7]. Furthermore, failure to apply a PIA
method may result in a breach of privacy laws - regulations.
The adoption of a PIA method can certainly benefit both individuals and organi-
zations [7, 11, 29]. As far as individuals are concerned, the first and most important
advantage is that they can ensure that companies processing their personal data are
compliant with standards, laws and guidelines. In addition, individuals can determine
how their information is collected, stored and processed, ensuring transparency. On the
124 E.-L. Makri et al.
other hand, there are also benefits for organizations applying a PIA method, as they
protect their clients, mitigating privacy breaches and ensuring compliance with the
legal framework. Furthermore, when organizations offer privacy-oriented services, they
build both their customers and employees’ trust. Last but not least, the financial benefits
should not be neglected. By reducing the ongoing costs of a project through the
minimization of the amount of information collected or processed, where possible [11],
the organization can obviously become more profitable.
Along with the positive aspects of applying a PIA method, there are also some
negative ones [29]. The implementation and the following publication of PIA results
may assist opponents to exploit them, putting the individuals’ privacy at risk. Besides, a
PIA method may impose delays and additional costs when a new project is imple-
mented. Another important concern is the limitation of the organizations’ flexibility.
That, in practical terms, means that organizations are committed to take actions in a
specific way, based on laws and regulations, without alternative choices, which, for
instance, would help them to finish specific tasks more quickly. However, the
advantages far outweigh the disadvantages, since a PIA is by all means a really good
way to protect users’ privacy and mitigate privacy risks.
Considering the aforementioned PIA benefits, it can be inferred that through the
application of a PIA method the most widely known privacy principles are maintained.
In 1980 [3] the OECD organization proposed eight privacy principles, which were
globally accepted, namely: purpose specification principle, collection limitation prin-
ciple, data quality principle, use limitation principle, openness principle, individual
participation principle, accountability principle, security safeguards principle. Their
aim was to minimize the risk of personal data disclosure and consist the basis of
privacy protection [12]. Cavoukian [13–15] strongly supports the notion of privacy-by-
design, according to which privacy should be maintained throughout the entire life-
cycle of an IT system, from the conception of a new system up to its implementation.
According to Oetzel and Spiekermann [16, 17] the notion of privacy-by-design is really
important in a PIA method as PIAs try to follow these privacy principles in order to
achieve privacy-by-design, which is one of the most crucial concerns of today’s pri-
vacy community.
In order for an organization to protect the personal data of an individual, Privacy
Enhancing Technologies (PETs) are employed to satisfy privacy requirements (data
protection, anonymity/ pseudonymity, unlinkability and unobservability), along with
the privacy principles mentioned above. PETs are various technical and procedural
measures that can prevent unnecessary or unwanted processing of users’ personal data
and can thus protect their privacy [12, 18].
The idea of a PIA method is relatively new. UK was the first European country to
use the PIA handbook which was developed and published by the Information Com-
missioner’s Office (ICO) in December 2007, followed by a revised edition in June 2009
[7, 9, 19]. In the ICO publication the basic stages of a PIA method were explained in
detail. In May of the same year, the European Commission’s Recommendation on
radio frequency identification (RFID) tags followed, where the Member States called to
ensure that industry, in collaboration with relevant civil society stakeholders, will
develop a framework for privacy and data protection impact assessment, providing
input to the Article 29 Data Protection Working Party [7, 10]. The above provide clear
A Proposed Privacy Impact Assessment Method Using Metrics 125
evidence that recently the need for applying a PIA method is gaining ground and that it
provides the foundation for taking privacy into consideration [27].
Recently, the European Commission co-funded the project PIAF (A Privacy Impact
Assessment Framework for data protection and privacy rights), which aims to
encourage the EU and its Member States to adopt a PIA policy [23]. The 22-month-
project was completed in December 2012 and resulted in a step-by-step guide to
privacy impact assessment so that privacy issues could be better addressed and personal
data could be more effectively protected from unwanted processing [24]. This project
highlights how important the application of a PIA method is in IT systems and that it is
mandatory to all EU Member States.
In May 2013, ICO conducted further research on the PIAs which led to the pub-
lication of the “Privacy impact assessment and risk management”, in which specific
improvements were suggested, the most important of which was the better integration
of PIAs with existing project management and risk management processes [20]. One
year later, in February 2014 [21], ICO published the Privacy Impact Assessment
(PIA) Code of Practice, which was updated in January 2017 to help organizations
comply with their data protection law obligations when changing the way they use
personal data [22].
In May 2017, the International Organization for Standardization (ISO) and the
International Electrotechnical Commission (IEC), which form the specialized system
for worldwide standardization, published a standard related to privacy impact assess-
ment defining PIAs [8], which is not significantly different from David Wright’s and
Paul De Hert’s definition [7]. Specifically, it will support that a PIA is more than a tool
and will be defined as a process beginning at the earlier stages of a system and
continuing throughout its development life cycle, thus achieving privacy-by-design [8].
The aim of the International Standard will be to provide guidelines on the process for
conducting a PIA method and on the structure and content of a PIA report. According
to the official published draft [8] “it will be relevant to those involved in designing or
implementing projects, including the parties operating data processing systems and
services that process PII”. In addition, it will be applicable to all public and private
bodies.
Prior to the upcoming standard, ISO produced a standard for PIAs in financial
services [28]. The aim of ISO 22307:2008 standard was to act as a tool for the internal
editing of personal data during the development of a proposed financial system (PFS).
It is used to mitigate the risks which appear when an organization processes the
financial data of customers and consumers, business partners as well as citizens.
Except for organizations such as ICO and ISO/IEC, eminent scientists have made
research on PIA methodologies. They have proposed PIA methodologies, which can be
easily applied by an organization, giving accurate results at the same time. In 2012 [16]
and 2013 [17], Oetzel Marie Caroline and Spiekermann Sarah proposed a systematic
method for privacy impact assessment. Their PIA method, based on NIST security risk
assessment process (2002) [31] and the UK PIA process (ICO, 2009) [9], reduces the
complexity of the privacy laws landscape for practitioners, helping them to take privacy
management decisions for their applications. Furthermore, it defines privacy targets,
evaluates the level of protection they need, identifies threats and proposes controls.
126 E.-L. Makri et al.
Throughout the years, the rapid improvement of PIA methodologies highlights their
importance on privacy and data protection. However, there is no explicit way to
quantify the privacy impact. In 2011, Wright [25, 29] highlighted this need, by stating
that “Making privacy impact assessments mandatory is not the end of the story. Audits
and metrics are needed to make sure that PIAs are actually carried out and properly so
and to determine if improvements to the process can be made”. More recently, in 2013,
Wadhwa and Rodrigues [26] agreed with David Wright’s statement, which practically
means that the specific need still exists.
One of the main reasons that organizations adopt PIAs is to gain users’ trust.
A number of PIA tools, like AIPCA/CICA privacy risk assessment tool, the Security
and Privacy Impact Assessment (SPIA) tool of the University of Pennsylvania, the
GS1 RFID Privacy Impact Assessment (PIA) tool, the Vienna University intelligent
PIA tool for RFID applications and the Privacy Impact Assessment Tool for Cloud
Computing proposed by Tancock, Pearson, and Charlesworth [37], have been proposed
in order to assist companies assess privacy risks. However, none of them employs
metrics to quantify the impact of a privacy breach.
In July 2016, Agarwal [30] highlighted the fact that although there are a series of
modular and well-structured online PIA tools (GS1 tool, iPIA tool, SPIA tool, etc.),
they all fail to provide a metric to assess progress in the implementation of privacy
controls. In his research, he developed a structured metric to measure privacy risk.
Before Agarwal, Oetzel and Spiekermann [16, 17], had already proposed a qualitative
metric (low, medium, high) for measuring privacy risks, but their effort was quite
unstructured and difficult to measure explicitly [30]. In order to evaluate privacy risk,
Agarwal defined it as the product of impact and likelihood. To be more specific,
Agarwal assessed the impact using Solove’s taxonomy and the likelihood using Lip-
ton’s work. For the calculation of the impact, he used four different dimensions of
privacy, splitting them into categories and subcategories. For the likelihood, he used
actors (companies, 3rd parties, others) and data characteristics (amount of data, sen-
sitivity of data, value of data involved). This paper proposed a structured privacy risk
metric, but failed to delve deeper into the organizations’ characteristics which can have
a considerable negative impact on the users’ privacy.
In June 2015, Commission Nationale de l’Informatique et des Libertés (CNIL)
published a PIA methodology, which is in line with EU’s General Data Protection
Regulation (GDPR) [2]. According to CNIL [34], the PIA methodology rests on two
pillars: firstly, the fundamental principles and rights and, secondly, the management of
data subjects’ privacy risks. To be more specific, the methodology consists of four
steps: the definition and description of the content of the processing of personal data
under consideration, the identification of existing or planned controls, the evaluation of
privacy risks and the decision to validate the manner in which it is planned to comply
with privacy principles and treat the risks, or review the preceding steps. In December
2017, CNIL published a free and open source PIA software [35] in order to help data
controllers to follow their methodology.
In conclusion, it can be inferred that an effective way to measure the privacy impact
is by using metrics. Metrics can help organizations to calculate the significance of
threats and lead them to take measures to mitigate the risks. Despite the remarkable
efforts to define metrics by various researchers [16, 17, 30], so far, there has been no
A Proposed Privacy Impact Assessment Method Using Metrics 127
detailed PIA method to use metrics and, at the same time, take into account the
organization characteristics. Furthermore, there is no method that integrates security
and privacy assessment.
the data that an organization stores/process either internally (e.g. employees’ data) or
externally (e.g. users’ data) are classified in the following categories:
• Personal Data: There are several approaches to define personal data. The Data
Protection Act [32] defines the personal data as data relating to a living individual
who is or can be identified either from the data directly or in conjunction with other
information that is in, or is likely to come into, the possession of the data controller.
The EU Data Protection Directive (REGULATION (EU) 2016/679) [2] defines the
personal data as any information relating to an identified or identifiable natural
person (‘data subject’); an identifiable natural person is one who can be identified,
directly or indirectly, in particular by reference to an identifier such as a name, an
identification number, location data, an online identifier or to one or more factors
specific to the physical, physiological, genetic, mental, economic, cultural or social
identity of that natural person. For the needs of the proposed method, this data set
category will host data that an organization stores and processes in its systems and
are related to an identified or identifiable natural person. Some indicative examples
of personal data are: name, surname, age, address, telephone number, email,
education.
• Sensitive Personal Data: The sensitive personal data can be considered a subcat-
egory of personal data. Sensitive personal data, according to the legal framework,
need high protection. They are defined in Sect. 2 of the Data Protection Act [32] as
personal data consisting of information relating to the data subject with regard to
racial or ethnic origin, political opinions, religious beliefs or other beliefs of a
similar nature, trade union membership, physical or mental health or condition,
sexual life, the commission or alleged commission by the data subject of any
offence, or any proceedings for any offence committed or alleged to have been
committed by the data subject, the disposal of such proceedings or the sentence of
any court in such proceedings. For the needs of the proposed method, this data set
category will host data that an organization stores and processes in its systems and
are related to an identified or identifiable natural person and belong to any of the
aforementioned sub-categories.
• Operational Data: Operational data are the data that an organization stores and
processes, as a result of the use of its information systems. More specifically, this
data set category includes data generated by the information system itself, like
logging users’ actions. For instance, a log file may contain details about the attempts
of a user log in, for how long she was connected, the timeframe that she used a
specific application etc.
• Financial Data: This data set includes all the financial data of an organization
(related to its employees and/or users). Similarly to the operational data, they are not
provided by users or employees but, instead, they are “created” by the organization
on behalf of its users and/or its employees. More specifically, the financial data are
classified in two sub-classes: (a) Data related to the payroll of the organizations’
employees, (b) Data related to payments by organization’s users for the provided
services. It should be stressed, that the financial data, for both sub categories, are
generated by the organization.
A Proposed Privacy Impact Assessment Method Using Metrics 129
• Other Data: Any data that cannot be classified in any of the above categories will
be considered under this final data set. However, the type, use and criticality of the
data should be explicitly defined.
Fig. 1. A common security and privacy Fig. 2. Privacy audit methodology structure
methodology [4] [5]
130 E.-L. Makri et al.
More specifically, the first step is the most important one since the “Purpose
Specification Principle” defines the scope of data collection and use. If this privacy
principle is not satisfied the other privacy principles will not be applied in the right
way, violating the data privacy. The second step includes the satisfaction of “Data
Collection Limitation Principle” and “Data Quality Principle”. If the purpose from the
step 1 has been specified, the data collection and use must be limited and related to the
purpose. Moreover, the collected data should be accurate and kept updated. If these
privacy principles are not satisfied the upcoming privacy principles will not be applied
in the right way, violating the data privacy. The third step includes the satisfaction of
“Use, Retention and Disclosure Principle”. If the privacy principles from the step 2
have been satisfied, the data should be limited used, retained and disclosed according to
organizations’ policies. If the privacy principle in the third step is not satisfied the
upcoming privacy principles will not be applied in the right way, violating the data
privacy. The fourth and last step includes the satisfaction of “Security Safeguards
Principle”.
The other privacy principles include the satisfaction of “Openness Principle”,
“Individual Participation Principle” and “Accountability Principle”. These privacy
principles should be satisfied throughout the entire methodology.
Based on the hierarchy of the steps (as depicted in Fig. 2), Step 1 is the most
important one, Step 2 is more important than Step 3 and Step 3 is more important than
Step 4. The other privacy principles should be applied throughout the entire process.
should exhibit a similar sensitivity level for the organization. Some indicative data
subcategories are:
• Personal Data (Data which uniquely identify a person using IDs, personal or marital
status, business activities etc.)
• Sensitive Personal Data (Medical Data, convictions etc.)
• Financial Data (Data related to financial transactions, yearly tax etc.)
• Operational Data (Data generated during the execution of a service, i.e. cookies,
private log files of the organization etc.)
• Other Data
As already mentioned, the estimation of the organization’s data sensitivity, through
risk analysis, will be based on the impact that could be caused to the organization by a
potential security incident on an independent data subcategory. The overall impact for
the organization will depend on the partial impact caused by each data subcategory,
adopting in all cases the worst-case scenario.
Input to Metric: The organization’s Data Set, classified in data subcategories DS1,
DS2 … (DSn).
Formulation: Through Risk Analysis the risk factor for each data subcategory is
calculated, depending on the impact that a security incident could cause to the
organization.
while its importance is b (where b > a) if the organization maintains and processes a
considerable amount of sensitive data.
Input to Metric: The characteristics of the organization.
Formulation: According to the organization’s characteristics a vector will be used to
assess their impact on the security and privacy issues.
z1::N ¼ ½a numeric value for CH1 from the range : Value1 Value2 . . . ValueN
½a numeric value for CH2 from the range : Value1 Value2 . . . ValueN
½...
½a numeric value for CHN from the range : Value1 Value2 . . . ValueN
Final Output: A metric “Vector z1..N”, providing the importance of each organiza-
tional characteristic.
Example:
where: CH1,2,…,10: Characteristics 1,2,…,10 (e.g. Data Volume, Data Type, etc.)
Final Output: A metric “ki” is defined, representing the customized (specific to the
organization) criticality of its characteristics on privacy issues.
Severity PP ¼ app ki
Final Output: A metric “Severity PP”, representing the overall severity of each pri-
vacy principle.
organization. The value of each Table Cell in the following table is calculated in
accordance to the following formula:
It should be stressed that the derived table cell value for a specific privacy principle
and a specific data set, will not be necessarily the same for different organizations, since
it depends on the calculated ki value which is related to specific organizational char-
acteristics (Table 2).
To summarize, the final value of each Table Cell highlights the criticality of each
privacy principle for every data set maintained by the organization. The method
employed for the calculation of that criticality level, as already explained in the pre-
vious sections, takes into account the consequences that the organization may expe-
rience in case of a security or privacy violation incident on a specific data set, the
weighting of each privacy principle and the unique characteristics of each organization
(Table 1).
The resulting table values offer a strong indication of the security measures and
privacy enforcement mechanisms that the organization should adopt in order to
effectively protect its data. More specifically the value of each table cell can be
compared with the minimum or/and maximum value that the specific cell can take,
depending on the characteristics of the organization (Table 1), and if it is found to be
near to the maximum cell value the criticality of the privacy principle for the specific
data set is considered to be very high.
4 Conclusions
This paper has presented a Privacy Impact Assessment method that employs metrics
and takes into account the peculiarities and other characteristics of the organization.
The aim is to assist the organizations to estimate the criticality of potential privacy
breaches and, thus, to select the appropriate security measures for the protection of the
data that they collect, process and store. As far as the identification and assessment of
the risks, it is performed through the risk management or/and PIA methodology that the
organization decides to use while for the mitigation of the risks it is in our plans for
future work to develop a feedback channel that will enhance the proposed method in a
way that it will be able to recalculate the criticality by taking into account the
suitability/effectiveness of the implemented mitigation techniques. We will thus sup-
port the Plan-Do-Check-Act (PDCA) model of the ISO 27000 standard.
Additionally, we are considering the integration of the requirements set by the new
Data General Protection Regulation (GDPR).
Acknowledgment. This work has been partially supported by the Research Center of the
University of Piraeus.
References
1. Hong, W., Thong, J.Y.L.: Internet privacy concerns an integrated conceptualization and four
empirical studies. MIS Q. 37(1), 275–298 (2013). https://papers.ssrn.com/sol3/papers.cfm?
abstract_id=2229627
2. Regulation (EU) 2016/679 of the European Parliament and of the Council: The European
Parliament and the Council of the European Union, 27 April 2016. http://eur-lex.europa.eu/
legal-content/EN/TXT/PDF/?uri=CELEX:32016R0679&qid=1485368166820&from=en
3. OECD Privacy Principles: OECDprivacy.org, 1980. http://oecdprivacy.org/
138 E.-L. Makri et al.
4. Makri, E.L., Lambrinoudakis, C.: Towards a common security and privacy requirements
elicitation methodology. In: Jahankhani, H., Carlile, A., Akhgar, B., Taal, A., Hessami, A.,
Hosseinian-Far, A. (eds.) ICGS3 2015. CCIS, vol. 534, pp. 151–159. Springer, Cham
(2015). https://doi.org/10.1007/978-3-319-23276-8_13
5. Makri, E.L., Lambrinoudakis, C.: Privacy principles: towards a common privacy audit
methodology. In: Fischer-Hübner, S., Lambrinoudakis, C., López, J. (eds.) TrustBus 2015.
LNCS, vol. 9264, pp. 219–234. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-
22906-5_17
6. Bélanger, F., Crossler, R.E.: Privacy in the digital age: a review of information privacy
research in information systems. J. MIS Q. 35(4), 1017–1042 (2011). http://dl.acm.org/
citation.cfm?id=2208951
7. Wright, D., De Hert, P.: Introduction to privacy impact assessment. In: Wright, D., De Hert,
P. (eds.) Privacy Impact Assessment. Law, Governance and Technology Series, vol. 6,
pp. 3–32. Springer, Dordrecht (2012). https://doi.org/10.1007/978-94-007-2543-0_1
8. ISO/IEC FDIS 29134: Information technology—Security techniques—Privacy impact
assessment—Guidelines, Target publication, 30 May 2017. http://www.iso.org/iso/
catalogue_detail.htm?csnumber=62289, https://www.iso.org/obp/ui/#iso:std:iso-iec:29134:
dis:ed-1:v1:en
9. Information Commissioner’s Office (ICO): Privacy Impact Assessment Handbook, Wilm-
slow, Cheshire, December 2007, Version 2.0, June 2009
10. European Commission: Recommendation on the implementation of privacy and data
protection principles in applications supported by radio-frequency identification, C (2009)
3200 final, Brussels, 12 May 2009. http://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?
uri=CELEX:32009H0387&from=EN
11. Information Commissioner’s Office (ICO): Data Protection Act, Conducting privacy impact
assessments code of practice, February 2014. https://ico.org.uk/media/for-organisations/
documents/1595/pia-code-of-practice.pdf
12. Wang, Y., Kobsa, A.: Privacy-Enhancing Technologies (2008). http://www.cs.cmu.edu/afs/
cs/Web/People/yangwan1/papers/2008-Handbook-LiabSec-AuthorCopy.pdf
13. Cavoukian, A.: Creation of a Global Privacy Standard, November 2006. http://www.ipc.on.
ca/images/Resources/gps.pdf
14. Cavoukian, A., Taylor, S., Abrams, M.E.: Privacy by design: essential for organizational
accountability and strong business practices. In: Identity in the Information Society, Springer
(2010). http://link.springer.com/article/10.1007/s12394-010-0053-z
15. Cavoukian, A.: Privacy by design – the 7 foundational principles, Technical report, In-
formation and Privacy Commissioner of Ontario, January 2011. (revised version)
16. Oetzel, M.C., Spiekermann, S.: Privacy-by-design through systematic privacy impact
assessment - a design science approach. In: ECIS 2012 Proceedings, Paper 160 (2012).
http://aisel.aisnet.org/ecis2012/160
17. Oetzel, M.C., Spiekermann, S.: A systematic method for privacy impact assessments: a
design science approach. Eur. J. Inf. Syst. 23(2), 1–25 (2013)
18. van Blarkom, G.W., Borking, J.J., Olk, J.G.E.: PET, Handbook of Privacy and Privacy-
Enhancing Technologies, The Case of Intelligent Software Agents (2003). ISBN 90-74087-
33-7. http://www.andrewpatrick.ca/pisa/handbook/Handbook_Privacy_and_PET_final.pdf
19. Information Commissioner’s Office (ICO): Privacy Impact Assessment Handbook, Wilm-
slow, Cheshire, Version 1.0, December 2007
20. Information Commissioner’s Office (ICO): Privacy impact assessment and risk management,
May 2013. https://ico.org.uk/media/1042196/trilateral-full-report.pdf
A Proposed Privacy Impact Assessment Method Using Metrics 139
21. Information Commissioner’s Office (ICO): Conducting privacy impact assessments code of
practice, February 2014. https://ico.org.uk/media/for-organisations/documents/1595/pia-
code-of-practice.pdf
22. Information Commissioner’s Office (ICO): The Guide to Data Protection, January 2017.
https://ico.org.uk/media/for-organisations/guide-to-data-protection-2-7.pdf
23. European Commission PIAF: A Privacy Impact Assessment Framework for data protection
and privacy rights, January 2011–October 2012. http://www.piafproject.eu/Index.html
24. Wright, D., Wadhwa, K.: A step-by-step guide to privacy impact assessment, Second PIAF
workshop, Sopot, 24 April 2012. http://www.piafproject.eu/ref/A_step-by-step_guide_to_
privacy_impact_assessment-19Apr2012.pdf
25. Wright, D.: Should privacy impact assessments be mandatory? Commun. ACM 54(8),
121–131 (2011). https://doi.org/10.1145/1978542.1978568. http://cacm.acm.org/magazines/
2011/8
26. Wadhwa, K., Rodrigues, R.: Evaluating privacy impact assessments. Innov.: Eur. J. Soc. Sci.
Res. 26(1–2), 161–180 (2013). http://www.tandfonline.com/doi/abs/10.1080/13511610.
2013.761748, http://www.tandfonline.com/doi/pdf/10.1080/13511610.2013.761748?
needAccess=true
27. Brooks, S., Nadeau, E.: Privacy Risk Management for Federal Information Systems,
Information Technology Laboratory, NIST, Internal Report 8062, May 2015. http://csrc.nist.
gov/publications/drafts/nistir-8062/nistir_8062_draft.pdf
28. Ferris, J.M.: The ISO PIA standard for financial services. In: Wright, D., De Hert, P. (eds.)
Privacy Impact Assessment. Law, Governance and Technology Series, vol. 6, pp. 307–321.
Springer, Dordrecht (2012). https://doi.org/10.1007/978-94-007-2543-0_14
29. Wright, D.: Should privacy impact assessments be mandatory? Trilateral Research &
Consulting, 17 September 2009. http://www.ics.forth.gr/nis09/presentations/18-wright.pdf
30. Agarwal, S.: Developing a structured metric to measure privacy risk in privacy impact
assessments. In: Aspinall, D., Camenisch, J., Hansen, M., Fischer-Hübner, S., Raab, C.
(eds.) Privacy and Identity 2015. IAICT, vol. 476, pp. 141–155. Springer, Cham (2016).
https://doi.org/10.1007/978-3-319-41763-9_10
31. NIST (National Institute of Standards and Technology): Risk management guide for
information technology systems, NIST Special Publication 800-30 (2002)
32. Data Protection Act (1998). http://www.legislation.gov.uk/ukpga/1998/29/contents, http://
www.legislation.gov.uk/ukpga/1998/29/pdfs/ukpga_19980029_en.pdf
33. European Union Agency for Network and Information Security (ENISA): CRAMM (CCTA
Risk Analysis and Management Method). https://www.enisa.europa.eu/topics/threat-risk-
management/risk-management/current-risk/risk-management-inventory/rm-ra-methods/m_
cramm.html
34. Commission Nationale de l’Informatique et des Libertés (CNIL), Privacy Impact Assessment
(PIA) Methodology (how to carry out a PIA), June 2015. https://www.cnil.fr/sites/default/
files/typo/document/CNIL-PIA-1-Methodology.pdf
35. Commission Nationale de l’Informatique et des Libertés (CNIL): The open source PIA
software helps to carry out data protection impact assessment, January 2018. https://www.
cnil.fr/en/open-source-pia-software-helps-carry-out-data-protection-impact-assesment
36. De Capitani, S., di Vimercati, S., Foresti, G.L., Samarati, P., Privacy, D.: Definitions and
techniques. Int. J. Uncertainty, Fuzziness Knowl.-Based Syst. 20(6), 793–818 (2012)
37. Tancock, D., Pearson, S., Charlesworth, A.: A privacy impact assessment tool for cloud
computing. In: Second IEEE International Conference on Cloud Computing, pp. 667–676.
Indiana University, USA (2010)
A Conceptual Redesign of a Modelling
Language for Cyber Resiliency
of Healthcare Systems
1 Introduction
Security covers an increasingly broad range of domains that rarely interplay in other
contexts. For example, a healthcare system’s security design should address, not just
hardware and software vulnerabilities, but also other issues, such as equipment failures,
human errors, dependencies of healthcare services. In this sense, it essential to provide
a common language to address and manage this heterogeneity within the security
context. Such a language will allow the specification of a broad range of security
requirements of different stakeholders within the healthcare setting. Moreover, it can
allow the analysis of their resiliency as part of their security requirements elicitation,
meaning as early as possible in their design.
Healthcare systems stand for the organization of interacting elements arranged to
accomplish one or more healthcare purposes (based on [1]). Examples of healthcare
systems are implantable cardiac medical devices; medical ventilator and robotic X-ray.
Long life-cycles characterize healthcare systems. Over the usable lifespan of health-
care, their design and development methods change [47]. While an understanding of
the preventive security aspects of healthcare systems’ design is essential, issues
associated with other requirements and constraints when incidents occur are of more
significant concern for life-critical and context-aware systems. Healthcare systems are
increasingly networked, interconnected and software-dependent. With limited resour-
ces and an ever-evolving threat landscape, any new insight into the cyber resiliency of
healthcare systems and their design and implementation becomes crucial [22].
Cyber resiliency (also termed resilience) stands according to NIST SP 800-160,
V.2. for “the ability to anticipate, withstand, recover from, and adapt to adverse
conditions, stresses, attacks, or compromises on systems that use or are enabled by
cyber resources.” [40]. Based on the context (e.g., supply chain, environmental, psy-
chological, technological) with which it associates, resiliency can approach different
types of problems. In this paper, we focus only on cybersecurity resiliency, excluding
other contexts of resiliency.
One approach to allow the by-design cyber resiliency of maintaining security
requirements is the Model-driven engineering (MDE) [10, 47]. For healthcare systems
that have the patient-in-the-loop, model-based frameworks that explicitly model an
MCPS’s interaction with the environment and with the patient can contribute towards
safer development [5]. Similarly, modelled-based security approaches have shown the
benefits of considering security requirements from the early stages of systems devel-
opment [32, 34]. Such modelling approaches can potentially facilitate the development
of healthcare systems that consider the full cyber resiliency life-cycle (i.e., preparation,
identification, containment, eradication, recovery, lessons learned) [13].
Many security requirements modelling approaches are based on Goal-oriented
Requirements Engineering (GORE). Typically, they analyze a system considering its
organizational, operational and technical environment; to identify issues and oppor-
tunities. High-level goals are then modelled refined to address such issues and meet the
opportunities [15, 20]. In Security Requirements Engineering (SRE), relevant
requirements are then elaborated to meet those goals [2, 32, 34]. MDE and SRE may be
used in combination to support the resilience of healthcare systems and in particular, to
improve the awareness of redesign and reconfiguration capabilities of a healthcare
system, before its actual construction. After all, any of such activities, if not well
studied in advance, can harm the patients. Such cases contradict with the fundamental
medical goal of “at least not harm” [5], and hence, they should not be ignored in
healthcare systems engineering.
The main aim of this paper is to explore the consideration of cyber resiliency under
conditions of uncertainty where incidents challenge the achievement of a healthcare
system’s goals. In this paper, we present the first step towards the modelling language,
which will be part of a framework: redesigning a metamodel. Notice that we do not
offer a modelling language, but we do present underlying conceptual considerations
that led to the redesign of the language.
The research outcomes presented here aim to enhance the resilience management of
cybersecurity by proposing a cybersecurity-resilience unified model. Mainly, the
contribution of this paper comprises of:
– a combination of resiliency in the cybersecurity domain, extending the Secure
Tropos approach to cover resiliency concepts. We focus on the design of systems
considering cyber resiliency from the stage of requirements engineering;
142 M. Athinaiou et al.
– the proposed conceptual model presented as a UML class diagram, useful for the
development of other cybersecurity artefacts that support cyber resiliency. Such
artefacts can include processes, algorithms and tools. Such artefacts can support the
semi-automation of a cyber resiliency analysis;
– the demonstration of the pertinence of the conceptual model in regard to the
healthcare context, through a case study.
2 Background
problem’s scope with anti-requirements and their usage to aid the formation of security
requirements and the examination of relevant vulnerabilities and threats [27].
The Secure Tropos [34] approach is also an extended Tropos [9] version, which
provides means to elicit and analyze security requirements. It allows the expression of a
wide range of security, privacy and trust requirements in the form of constraints. Secure
Tropos is well-known for being a robust language for defining secure systems at the
organizational level. Its organizational approach to security allows its extension to
cover the healthcare context considering attacks that can have beyond cyber also
physical impact. Furthermore, existing automatic tools (i.e., SecTro [36]) ease the
design activities using this metamodel and can also be extended accordingly.
The Secure Tropos metamodel inspired the first design attempt of a cyber resiliency
modelling language for healthcare [34]. The initial design of the metamodel can be
found in [6]. The decision of a redesign stemmed from interviews with experts from the
Brighton and Sussex University Hospitals and MedStar Health as well as the appli-
cation of small case studies. From there, it became apparent that the metamodel needed
some enhancements. As a group, we agreed into three main redesign enhancements: the
incident, the healthcare context and the inclusion of constructs related to resiliency.
These enchantments led to the design of a second version of the metamodel, presented
in Fig. 1. The following subsections report on how this metamodel was redesigned.
5. Secure Tropos, has been already extended to cover different types of systems (e.g.,
cloud security requirements [33], trust [37], business processes [4]). Following this
paradigm, an extension can take place addressing the unique characteristics of
healthcare systems in relation to their cyber resiliency.
Having identified some of the advantages of extending Secure Tropos, coincide with
other security requirements approaches (e.g., KAOS [25], CORAS [8], SQUARE [31],
GBRAM [3]). However, these approaches tend to focus on the preventive aspect of
security. Resiliency stands for the ability to prepare for, respond to and recover from
cyber incidents. It helps a healthcare infrastructure to prepare for incidents, defend
against, limit their severity and ensure the continuation of operations despite an inci-
dent. Cyber resiliency has emerged as traditional cybersecurity measures are chal-
lenged, especially in the case of APTs [35]. When incidents do occur, the systems need
to be able to keep up with the changes and continue to pursue critical goals and
functions.
resiliency, we realized that there was a need to form constructs that support their
unification. Though, there was no clear way derived from the literature to support us
into making such a decision. Consequently, to face this obstruction, we plan to involve
practitioners for validation of the redesigned language.
The main semantic changes reflected in the three parts of the redesigned model
resulted from a systematic review of the scientific and standardization literature. The
purpose of the review was the sound derivation of a conceptual model. Here we cannot
present in detail the review process, but we discuss the main findings that resulted in
the redesign of the modelling language at a conceptual level. Every construct has a
variety of functions and implications, which can change over time and context. The
conceptual unification of cybersecurity and resiliency starts with the identification of
the basic constructs of the problem to be treated. The common terms identified in the
relevant literature are incident, healthcare, response and security. We briefly present
how they have been interpreted, offering useful components for the design of a con-
ceptual model of a modelling language.
The set of collected papers interprets the term incident in four different ways. The
majority of papers (7) consider an event such as updates, hardware failures, emer-
gencies, human errors, natural disasters, misuse and abuse cases as occurrences of
incidents [12, 14, 17, 21, 30, 44, 45]. In 4 papers an incident interpreted as a cyber-
security attack like hacking, ransomware and advanced persistent threat (APT) [7, 11,
12, 43]. Two (2) papers use the NIST SP 8000-61 definition either explicitly or
implicitly [16, 23] and 1 paper focuses on the effects of an occurrence on systems
functions and society as an incident [18]. Here it seems that an incident definition
exists, and each study chooses to focus on an aspect of an incident. Other studies seem
to choose a wider scope, that of event that also includes incidents and subsequently
cybersecurity incidents.
Healthcare overall appeared to have five different meanings. In 3 papers coincides
with the term hospital [16, 17, 44], in 6 papers with a form of a system, including
medical cyber-physical systems, electronic medical records systems and healthcare
information systems [7, 11, 14, 30, 43, 45], in 3 papers as a healthcare critical
infrastructure or a particular type (e.g. NHS) [7, 18, 45], in 2 papers addressed
healthcare organizations in general [21, 23] and 1 was focusing on healthcare infor-
mation [12]. From the above, it can be observed that the majority of papers interpret the
term healthcare as a type of healthcare system. It is important here to clarify, that the
A Conceptual Redesign of a Modelling Language for Cyber Resiliency 149
reason the number of papers corresponding to meanings (15) is greater than the set of
papers collected (13) is that in some papers the same term is used but is given multiple
meanings. The same holds for the rest of the terms and the corresponding number of
papers with similar interpretations.
When it comes to response, 4 papers address specific aspects/phases like detection,
forensics and post-incident activities [7, 12, 21, 30] 3 papers refer to all the phases of
incident response [14, 16, 43], 2 papers analyze response overarching manner ranging
from reactive on the one end and on the other to proactive adaptable responses to
incident characteristics [11, 44], in 2 papers response is studied within the planning
context in the form of an incident response plan (IRP) along with other types of plans
like emergency plan and business continuity plan [17, 23]. Response is also considered
closely associated with resilience and recovery in [45] and with management in [18].
The selected set of papers studies response from many aspects, usually related either
with its phases individually or as a whole and in other studies as broader positioning of
response within healthcare organizations.
The concept of security is one that is commonly associated with safety. Within this
set of papers that was the case only in [16] and even there, the proposed security
approach adjusts to feet cybersecurity needs. Examining relevant papers, it also seems
that security mostly in the past but also in the present focuses on information security
and confidentiality, integrity and availability properties [12, 16, 21, 30, 44, 45].
However, in more recent studies, cyber-physical aspects are studied as well as moving
from information technology-security to what is referred to in the broader literature as
operational technology-security [7, 18].
Specific aspects of security are also studies in the relevant literature. The concep-
tualization of security as vulnerable [45], the adaptability of security [11, 45] are two
such examples. Moreover, security is addressed from a socio-technical perspective
[43], as organization wide [14, 17]. In some cases, defense [44] and forensics [12] as
important elements of security are studied based on risk plans [23]. Thus, security
evolves as cyber risks do. The cyber risks become more sophisticated and dynamic, and
security interpretations and understanding reflect these changes.
This definition similar to NIST SP.800-61r2 is quite subtle because it not only
allows that an incident can be something that actually happened in the real world, but
also that it can be imaginary and does not really occur. The example of a false positive
alarm of an intrusion detection system can be treated as an incident even though it did
not occur. The second meaning describes incidents that occur in computer systems. In
this way, the term incident corresponds to incidents as a threat or as an actual attack.
An incident can be better understood through its likelihood and severity. In our
framework, we consider likelihood as evidence that maintains or rejects the occurrence
of an incident (MAIN and REJ). For example, the resilience entity all the actions the
surgeon takes should be recorded with nonrepudiation capability in the telesurgery
robotic system will help to enforce (MAIN) the security constraint protect against any
reasonably anticipated uses or disclosures of patient’s health-care data and discard
(REJ) the incident modify messages while packets are in-flight.
On the other hand, severity is introduced as the influence of an incident to a security
constraint. This relation allows us to model situations where a single incident impacts
on more than one security constraints. The occurrence of an incident contributes to
security constraint maintenance. In other words, a system under normal circumstances
achieves a security constraint. When an incident occurs, the system wants to maintain
security constraint achievement. The occurrence of an incident contributes to a security
constraint’s negatively in regard to its maintenance. Since the severity of an incident
restrains a security constraint when it occurs, in our model, we use only MAIN rela-
tions between an incident and security constraints. This relation stands for the main-
tenance of an incident and can result in a positive or negative contribution to a security
constraint’s maintenance.
On the other hand, a resilience mechanism is a tool or technique that can be adopted
in order to either prevent, mitigate or recover from an incident or is meant to implement
a security constraint. A resilience mechanism might operate by itself, or with others, to
provide a particular service. When an incident stands for a threat, then the preventive
aspect of a resilience mechanism is meaningful, whereas in cases where an incident is
an actual attack, then the mitigating and recovery aspects are relevant.
A healthcare and safety constraint (HSC) is a safety condition that the system has
to achieve and restricts a security constraint in order not to endanger a patient’s health
and/or well-being. In the modelling process, HSC constraints are modelled as variation
points of a resiliency plan. They are imposed by a healthcare actor that restricts the
achievement of a security constraint. HSC constraints are within the control of an actor.
This association with actors means that, differently than security constraints, HSC
constraints are conditions that an actor wishes to introduce to protect the patient in the
loop that characterizes healthcare systems. However, HSC constraints are examined
based on how they affect security entities and thus contribute towards the analysis of
resilience security requirements. HSC constraints can also be grouped according to the
safety objective towards the achievement they contribute. Safety objectives are broader
descriptions of safety principles or rules such as sterilization, calibration and
interoperability.
A Conceptual Redesign of a Modelling Language for Cyber Resiliency 151
4 Case Study
We take a scenario where a surgical system performs a surgical procedure (e.g., biopsy)
on a patient with manipulators and an endoscope. An endoscope is a long, thin, flexible
tube that has a light source and camera at one end. In our scenario, the surgical system
comprises a surgical robot, including a station, four robot arms mounted on the station
and a console for controlling the surgical robot. The surgical system also comprises a
data server for storing information from diagnostic imaging modalities (e.g. MRI, CT,
X-ray) which have been captured from a patient with the use of an ultrasonic diagnostic
device mounted on the distal end of a robot arm and a display unit. The display unit
simultaneously displays an endoscopic image and acquired from the endoscope and an
ultrasonic image acquired by the ultrasonic diagnostic device. This scenario provides a
simplified view of the stakeholders in the surgical system, the healthcare services
supported, and the concepts involved when a healthcare service is provided.
Due to space limitations in Figs. 3 and 4 we present a partial view that captures the
security and healthcare context, respectively. Particularly Fig. 4 depicts the healthcare
context along with goals, security constraints and resilience entities. The process starts
taking the security constraints from Secure Tropos and forming a conjunctive security
constraints tree (where the relation between sub-security constraints represents con-
junctive or disjunctive sub-security constraints). We have developed a simple security
constraints structure with parent goal “perform surgical procedure on a patient” that has
an AND decomposition (both of them need to be achieved for the parent goal to be
achieved) with the sub-goals “use console to control the surgical robot” and “utilize
154 M. Athinaiou et al.
patient data from the data server”. From the high-level security constraint “perform a
secure surgical procedure on a patient”, we can also extract leaf security constraints that
must be satisfied by resilience entities within the system. In our example, we prefer to
keep simplicity at this point, because the AND/OR decomposition are well known in
the existing literature.
Given ongoing attacks and expected incidents, we derive what security constraints
are relevant to these incidents and consequently, what are the security entities that we
need to consider. These considerations take the form of a three-layered incident model
that connects security constraints, incidents and resilience entities, as shown in Fig. 3.
This model is then used as input for the instantiation of Fig. 4. Moreover, by reviewing
healthcare process documents and relating them with resilience entities, different points
where a response will need to adjust to the ongoing conditions are specified. In our
example, some of such points are “ergonomic settings”, “laparoscopic procedure” and
“change device settings”. Taking one of these points, let us say “laparoscopic proce-
dure” a security practitioner that considers implementing a resilience entity such as
“encrypt data streams between the surgeon’s terminal and the robot” has to consider if a
laparoscopic operation is taking place at the same time. If so, the overhead or other
complication that encryption might result from having to be valued in relation to the
potential impact the response can have to the ongoing healthcare process and ulti-
mately, the patient.
With this simple case, we were able to demonstrate one of the additional capa-
bilities that the enhanced design can offer to cybersecurity practitioners of healthcare
environments. In particular, we looked at the constructs that relate to the healthcare
context and described at a high-level process through which such models can be
instantiated.
5 Conclusions
This paper focuses on cyber resiliency in relation to incidents that have recently arisen
or may arise for healthcare systems. The critical result of this revision of the modelling
language was the update of the metamodel to define more accurately the constructs
related to incidents, healthcare and resiliency. These enhancements were made to allow
security engineers to define a structure to support the resiliency of specific applications
relevant to their healthcare systems and incident conditions they face or prepare to
manage. In a case study for a robotic surgical system, we were able to demonstrate one
aspect of the application of the modelling language extensions. A detailed validation
needs to take place in future work. Because of the wide variety of physical and digital
capabilities of healthcare systems along with the potential impact they can have, we
believe that their cybersecurity needs to be studied further.
References
1. ISO/IEC/IEEE 15288:2015. https://www.iso.org/standard/63711.html. Accessed 12 July
2019
2. Li, T., Horkoff, J., Mylopoulos, J.: Integrating security patterns with security requirements
analysis using contextual goal models. In: Frank, U., Loucopoulos, P., Pastor, Ó.,
Petrounias, I. (eds.) PoEM 2014. LNBIP, vol. 197, pp. 208–223. Springer, Heidelberg
(2014). https://doi.org/10.1007/978-3-662-45501-2_15
3. Antón, A.I., Earp, J.B.: Strategies for developing policies and requirements for secure and
private electronic commerce. In: Ghosh, A.K. (ed.) E-Commerce Security and Privacy.
Advances in Information Security, vol. 2, pp. 67–86. Springer, Boston (2001). https://doi.
org/10.1007/978-1-4615-1467-1_5
4. Argyropoulos, N., Mouratidis, H., Fish, A.: Advances in Conceptual Modeling. Springer,
Cham (2015). https://doi.org/10.1007/978-3-642-33999-8
5. Arney, D., Pajic, M., Goldman, J.M., Lee, I., Mangharam, R., Sokolsky, O.: Toward patient
safety in closed-loop medical device systems. In: Proceedings of the 1st ACM/IEEE
International Conference on Cyber-Physical Systems - ICCPS 2010, pp. 139–148. ACM
Press, Stockholm (2010)
6. Athinaiou, M., Mouratidis, H., Fotis, T., Pavlidis, M., Panaousis, E.: Towards the definition
of a security incident response modelling language. In: Furnell, S., Mouratidis, H., Pernul, G.
(eds.) TrustBus 2018. LNCS, vol. 11033, pp. 198–212. Springer, Cham (2018). https://doi.
org/10.1007/978-3-319-98385-1_14
7. Boddy, A., Hurst, W., Mackay, M., Rhalibi, A.E.: A study into data analysis and
visualisation to increase the cyber-resilience of healthcare infrastructures. In: Proceedings of
the 1st International Conference on Internet of Things and Machine Learning - IML 1917,
pp. 1–7. ACM Press, Liverpool (2017)
156 M. Athinaiou et al.
8. Den Braber, F., Hogganvik, I., Lund, M.S., Stlen, K., Vraalsen, F.: Model-based security
analysis in seven steps a guided tour to the CORAS method. BT Technol. J. 25(1), 101–117
(2007)
9. Bresciani, P., Perini, A., Giorgini, P., Giunchiglia, F., Mylopoulos, J.: Tropos: an agent-
oriented software development methodology. Auton. Agents Multi-Agent Syst. 8(3), 203–
236 (2004)
10. Chapurlat, V., et al.: Towards a model-based method for resilient critical infrastructure
engineering how to model critical infrastructures and evaluate ist resilience? How to model
critical infrastructures and evaluate its Resilience? In: 2018 13th Annual Conference on
System of Systems Engineering (SoSE), pp. 561–567. IEEE, Paris (2018)
11. Chen, Q., Lambright, J.: Towards realizing a self-protecting healthcare information system.
In: 2016 IEEE 40th Annual Computer Software and Applications Conference (COMPSAC),
pp. 687–690. IEEE, Atlanta (2016)
12. Chernyshev, M., Zeadally, S., Baig, Z.: Healthcare data breaches: implications for digital
forensic readiness. J. Med. Syst. 43(1), 7 (2019)
13. Cichonski, P., Millar, T., Grance, T., Scarfone, K.: Computer Security Incident Handling
Guide: Recommendations of the National Institute of Standards and Technology. Technical
report NIST SP 800-61r2, National Institute of Standards and Technology (2012)
14. Cooper, T., Collmann, J., Neidermeier, H.: Organizational repertoires and rites in health
information security. Camb. Q. Healthc. Ethics 17(4), 441–452 (2008)
15. Dardenne, A., van Lamsweerde, A., Fickas, S.: Goal-directed requirements acquisition. Sci.
Comput. Program. 20(1–2), 3–50 (1993)
16. DeVoe, C., Rahman, S.S.M.: Incident response plan for a small to medium sized hospital.
Int. J. Netw. Secur. Appl. 5(2), 1–20 (2013)
17. Genes, N., Chary, M., Chason, K.W.: Case study. An academic medical centers response to
widespread computer failure. Am. J. Disaster Med. 8(2), 145–150 (2013)
18. Ghafur, S., Grass, E., Jennings, N.A., Darzi, A.: The challenges of cybersecurity in health
care: the UK National Health Service as a case study. Lancet Digit. Health 1(1), e10–e12
(2019)
19. Giorgini, P., Massacci, F., Zannone, N.: Security and trust requirements engineering. In:
Aldini, A., Gorrieri, R., Martinelli, F. (eds.) FOSAD 2004-2005. LNCS, vol. 3655, pp. 237–
272. Springer, Heidelberg (2005). https://doi.org/10.1007/11554578_8
20. Giorgini, P., Mylopoulos, J., Sebastiani, R.: Goal-oriented requirements analysis and
reasoning in the Tropos methodology. Eng. Appl. Artif. Intell. 18(2), 159–171 (2005)
21. He, Y., Johnson, C.: Challenges of information security incident learning: an industrial case
study in a Chinese healthcare organization. Inf. Health Soc. Care 42(4), 393–408 (2017)
22. Lee, I., et al.: Challenges and research directions in medical cyberphysical systems. Proc.
IEEE 100(1), 75–90 (2012)
23. Jalali, M.S., Russell, B., Razak, S., Gordon, W.J.: EARS to cyber incidents in health care.
J. Am. Med. Inf. Assoc. 26(1), 81–90 (2019)
24. Jürjens, J.: UMLsec: Extending UML for Secure Systems Development. In: Jézéquel, J.-M.,
Hussmann, H., Cook, S. (eds.) UML 2002. LNCS, vol. 2460, pp. 412–425. Springer,
Heidelberg (2002). https://doi.org/10.1007/3-540-45800-X_32
25. van Lamsweerde, A.: Goal-oriented requirements engineering: a guided tour. In: Proceed-
ings Fifth IEEE International Symposium on the Requirements Engineering, pp. 249–262.
IEEE Computer Society, Toronto (2000)
26. van Lamsweerde, A., Letier, E.: From object orientation to goal orientation: a paradigm shift
for requirements engineering. In: Wirsing, M., Knapp, A., Balsamo, S. (eds.) RISSEF 2002.
LNCS, vol. 2941, pp. 325–340. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-
540-24626-8_23
A Conceptual Redesign of a Modelling Language for Cyber Resiliency 157
27. Lin, L., Nuseibeh, B., Ince, D., Jackson, M., Moffett, J.: Introducing abuse frames for
analyzing security requirements. J. Lightwave Technol. 371–372 (2003). IEEE Comput.
Soc, Monterey Bay, CA, USA
28. Lodderstedt, T., Basin, D., Doser, J.: SecureUML: a UML-based modeling language for
model-driven security. In: Jézéquel, J.-M., Hussmann, H., Cook, S. (eds.) UML 2002.
LNCS, vol. 2460, pp. 426–441. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-
45800-X_33
29. McDermott, J., Fox, C.: Using abuse case models for security requirements analysis. In:
Proceedings 15th Annual Computer Security Applications Conference (ACSAC 1999),
pp. 55–64. IEEE Computer Society, Phoenix (1999)
30. McGlade, D., Scott-Hayward, S.: ML-based cyber incident detection for Electronic Medical
Record (EMR) systems. Smart Health 12, 3–23 (2019)
31. Mead, N.R., Stehney, T.: Security quality requirements engineering (SQUARE) method-
ology. ACM SIGSOFT Softw. Eng. Notes 30(4), 1 (2005)
32. Meland, P.H., Paja, E., Gjre, E.A., Paul, S., Dalpiaz, F., Giorgini, P.: Threat analysis in goal-
oriented security requirements modelling. In: Computer Systems and Software Engineering:
Concepts, Methodologies, Tools, and Applications, pp. 2025–2042. IGI Global (2018)
33. Mouratidis, H., Argyropoulos, N., Shei, S.: Security requirements engineering for cloud
computing: the secure tropos approach. In: Karagiannis, D., Mayr, H., Mylopoulos, J. (eds.)
Domain-Specific Conceptual Modeling, pp. 357–380. Springer, Cham (2016). https://doi.
org/10.1007/978-3-319-39417-6_16
34. Mouratidis, H., Giorgini, P.: Secure tropos: a security-oriented extension of the tropos
methodology. Int. J. Softw. Eng. Knowl. Eng. 17(02), 285–309 (2007)
35. Mwiki, H., Dargahi, T., Dehghantanha, A., Choo, K.-K.R.: Analysis and triage of advanced
hacking groups targeting western countries critical national infrastructure: APT28, RED
October, and Regin. In: Gritzalis, D., Theocharidou, M., Stergiopoulos, G. (eds.) Critical
Infrastructure Security and Resilience. ASTSA, pp. 221–244. Springer, Cham (2019).
https://doi.org/10.1007/978-3-030-00024-0_12
36. Pavlidis, M., Islam, S., Mouratidis, H.: A CASE tool to support automated modelling and
analysis of security requirements, based on secure tropos. In: Nurcan, S. (ed.) CAiSE Forum
2011. LNBIP, vol. 107, pp. 95–109. Springer, Heidelberg (2012). https://doi.org/10.1007/
978-3-642-29749-6_7
37. Pavlidis, M., Islam, S., Mouratidis, H., Kearney, P.: Modeling trust relationships for
developing trustworthy information systems. Int. J. Inf. Syst. Model. Des. 5(1), 25–48
(2014)
38. Pavlidis, M., Mouratidis, H., Panaousis, E., Argyropoulos, N.: Selecting security mecha-
nisms in secure tropos. In: Lopez, J., Fischer-Hübner, S., Lambrinoudakis, C. (eds.)
TrustBus 2017. LNCS, vol. 10442, pp. 99–114. Springer, Cham (2017). https://doi.org/10.
1007/978-3-319-64483-7_7
39. Ransford, B., Clark, S.S., Kune, D.F., Fu, K., Burleson, W.P.: Design Challenges for Secure
Implantable Medical Devices. In: Burleson, W., Carrara, S. (eds.) Security and Privacy for
Implantable Medical Devices, pp. 157–173. Springer, New York (2014). https://doi.org/10.
1007/978-1-4614-1674-6_7
40. Ross, R., Graubart, R., Bodeau, D., McQuaid, R.: Systems Security Engineering Cyber
Resiliency Considerations for the Engineering of Trustworthy Secure Systems. Technical
report, NIST (2018)
41. Schumacher, M.: Toward a security core ontology. In: Security Engineering with Patterns:
Origins, Theoretical Models, and New Applications, pp. 87–96. no. 2754, LNCS, Springer,
New York (2003). https://doi.org/10.1007/b11930
158 M. Athinaiou et al.
42. Sindre, G., Firesmith, D.G., Opdahl, A.L.: A reuse-based approach to determining security
requirements. Requirements Eng. 10, 34–44 (2004)
43. Sittig, D., Singh, H.: A socio-technical approach to preventing, mitigating, and recovering
from ransomware attacks. Appl. Clin. Inf. 07(02), 624–632 (2016)
44. Wiant, T.L.: Information security policy’s impact on reporting security incidents. Comput.
Secur. 24(6), 448–459 (2005)
45. Williams, P.A.H.: Is cyber resilience in medical practice security achievable? In:
Proceedings of the 1st International Cyber Resilience Conference, pp. 105–111. Edith
Cowan University, Perth (2010)
46. Yu, E.S.K.: Modeling strategic relationships for process reengineering, Ph.D. thesis,
University of Toronto, Canada (1995)
47. Jiang, Z., Pajic, M., Mangharam, R.: Cyberphysical modeling of implantable cardiac medical
devices. Proc. IEEE 100(1), 122–137 (2012)
Shaping Digital Identities in Social Networks:
Data Elements and the Role of Privacy
Concerns
Abstract. Individuals today shape a digital identity through which they “in-
troduce” themselves to others in Social Network Sites (SNS). SNS embody
features that enable users to customize their digital identity at will and to dis-
close desirable elements of their personality. The processes by which users
shape their digital identity through information disclosure are largely unknown,
including the role of privacy concerns. In this paper we identify the data ele-
ments that users consider important for shaping their digital identity in SNS and
how privacy concerns shape this process.
In order to explore the above, we conducted an online survey research with
759 participants. Our findings reveal the elements that users consider as
important for shaping their digital identity. They also demonstrate that users’
privacy concerns do not seem to affect the amount of information users choose
to publish when shaping their digital identity. Finally, we show that particular
characteristics of social networking platforms affect the way that users shape
their digital identity and privacy behavior.
1 Introduction
Identity is a complex and multifaceted concept due to the dynamic way it is formed and
redefined in space and time, as well as due to the complexity that characterizes
interpersonal interaction as to how it is perceived by oneself and others (Subrah-
manyam and Šmahel 2011; Delahunty 2012).
Within classical anthropological theories, human identity is considered as static: a
distinct personality that corresponds to the physical body of a human being (Stone
1996). However, in practice people experience identity in a completely different way
than this conceptualization; they choose how to present oneself differently depending
on the context in which they act and their target audience each time (Goffman 1959).
Nowadays, social, cultural, economic and technological developments suggest that
identity is a “fluid” concept and takes shape within a social context equally fluid and
changing (Papacharissi 2010; Kimmons 2014). It is also argued that there is no sub-
stantial or real identity, but multiple identities or “versions” of the identity that rep-
resent respectively the multiple “selves” of the individual (e.g. private and professional)
(Subrahmanyam and Šmahel 2011; Barbour and Marshall 2012; Ramírez and Palu-ay
2015). These “selves” are formed and activated within unique timeframes (Gee 1999).
Thus, the term “identity” relates to the point of view that one has of himself/herself or
the way in which the views of oneself are presented in various aspects of art and the
media (Marwick 2013). Many studies examine the meaning of identity as the way in
which someone “introduces” himself to other people in contexts enabled by Web 2.0
technologies (Canary et al. 2003; Baym 2010; Boyd 2010).
A “digital identity” refers to the way in which individuals choose to introduce
themselves when they communicate and interact on the Internet (Canary et al. 2003;
Delahunty 2012; Michikyan et al. 2015). The use of the term “digital identity” com-
pared to the term “identity”, implies a distinction between the way users would present
themselves on the Internet and in the real life. However, people nowadays commu-
nicate with other people they know in real life and through social networks and are
continuously online through mobile devices; thus, it makes sense that they choose to
disclose for themselves true identity data (Marwick 2013). Especially within Web 2.0
and social network contexts individuals are supposed to voluntary create a virtual
presentation of themselves (Baym 2010; Subrahmanyam and Šmahel 2011; Fieseler
et al. 2014; Junco 2014). Considering that social media users constantly present
themselves online in various roles (e.g. user as: self-creator through posted content,
member of specific groups expressing social belonging, information-seeker, audience
of other users’ activities, self-presenter), they are expected to maintain these presen-
tations as unique personal identities (Fieseler et al. 2014; Bareket-Bojmel et al. 2016).
When it comes to real-life representations (e.g. personal and professional), it is easier
for the individual to identify himself or herself, try changes to his public profile and get
feedback without risk (Belk 2013; Junco 2014). On the contrary, online identities are
more complex to create and change. Social Networking Sites (SNS) comprise of virtual
collections of user profiles that are used to maintain ties and can be shared with other
users (Hughes et al. 2012). Thus, a digital identity comprises of the information that an
individual share publicly, but it also includes the conclusions drawn by other users.
People form an opinion for someone by looking at and interpreting every aspect of
his/her online profile, activity and interaction with others (Delahunty 2012; Marwick
2013; Wang 2013; Kimmons 2014; Hodkinson 2017).
Through the creation of virtual profiles, the use of SNS has led to the collection and
storage of huge amounts of information about people, part of which may refer to
personal and sensitive data (Zhang et al. 2010). Although users provide their infor-
mation for specific purposes and expect that it would be shared only with specific
recipients, in reality personal information shared in SNS is often used for commercial
and advertising interests or become the target of malicious attacks (Gross and Acquisti
2005; Barnes 2006), raising issues of privacy violations (Krishnamurthy and Wills
2008). This exploitation of SNS data reveals a key contradiction that users face: While
information disclosure is a prerequisite for anyone who wants to build a strong online
profile and take full advantage of the networking capabilities offered by the SNS, as the
Shaping Digital Identities in Social Networks 161
depending on the SNS. We did not find evidence that privacy concerns prohibit users
from disclosing the information that they believe are part of their digital identities.
Section 2 presents background studies through which we have identified the data
elements that comprise digital identities and we have formulated our hypotheses
regarding digital identity shaping and the role of privacy concerns and social network
site. Section 3 presents our research strategy including the design of the survey. In
Sect. 4 we present the results of the empirical analysis and our findings, while Sect. 5
concludes the paper.
Many researchers argue that the formation of digital identity resembles the process of
“public self-presentation” (Larsen 2007; Mehdizadeh 2010; Subrahmanyam and
Šmahel 2011; Wang 2013; Michikyan et al. 2015; Ramírez and Palu-ay 2015) and
refers to the way in which the users choose to introduce themselves when they com-
municate and interact on the Internet (Canary et al. 2003; Hodkinson 2017). Building a
digital identity and disclosing personal information is part of the modern social net-
working philosophy. The typical practices of communication, interaction and self-
presentation in social media have also been described as the usage of an “identity-kit”
(Belk 2013). Zhao et al. (2008) explored the formation of digital identity in online
environments that favor - or even require - a correspondence with the real identity.
Their findings suggest that people nowadays tend to show rather than say things about
themselves. According to this view, digital identity is not a personal trait but a social
“product”, which is not only shaped by what is shared by the users but also by what
other users are sharing or stating about them. The above is also confirmed by research
findings suggesting that young people use their digital literacy skills to construct their
digital identity (Alvermann et al. 2012; Subrahmanyam and Šmahel 2011).
The real name of an individual is a crucial element of digital identity in SNS
(Larsen 2007; Zhao et al. 2008). Furthermore, SNS users whose digital identity cor-
responds to the real one, are considered more trustworthy and, therefore, they broaden
their networking easier and faster. SNS encourage individuals to use their real name to
prevent frauds and to guarantee security. On the other hand, this allows SNS to exploit
users’ personal data for commercial purposes (Marwick 2013).
Relevant studies also suggest that, in addition to the use of their real name, the
digital identity of users is also formed by disclosing a series of personal data directly
related to their real identity (such as gender, job position, interests) and habits of one’s
everyday life (Stutzman 2006; Pempek et al. 2009; Ismail 2010; Shafie et al. 2012;
Marwick 2013). Photos are an essential element of a SNS digital identity; they act as
links between the past and the present self and therefore represent identity (Salimkhan
et al. 2010; Ismail 2010). Photos can transfer messages that the user wants to com-
municate. They show a desirable image of the users (Tufekci 2008) and hence users
choose to upload the photos that help them highlight positive traits and cover any
negative ones (Zhao et al. 2008). Another element of digital identity is the users’ level
of social connectivity and popularity, which is calculated based on the number of
Shaping Digital Identities in Social Networks 163
online friends, posts and participation in groups (Gross and Acquisti 2005; Zhao et al.
2008; Hoy and Milne 2010; Shafie et al. 2012).
Besides explicit elements that form a digital identity, such as hobbies, interests,
opinions, favorite music, movies and so on, there are also implicit ones. Research
shows that even the fonts and pseudonyms that users choose may symbolize aspects of
their identity that make them stand out from others especially when they use them to
enter new interest groups (Ismail 2010; Papacharissi 2010; Shafie et al. 2012; Marwick
2013; van Dijck 2013; Cover and Doak 2015; Hodkinson 2017).
Thus, a digital identity encompasses a wide range of offline authentication attributes
(e.g. name, gender, age etc.), combined with several online elements and behaviors
(e.g. profiles, usernames, passwords, shared or published content etc.) (Wessels 2012;
Ramírez and Palu-ay 2015; Eastin et al. 2016). All these attributes, elements and
behaviors create clear links and common references between social and technological
interpretation of identity (Wessels 2012).
Although existing literature does not examine digital identity elements per se, there
are various studies that help us specify the information that users commonly choose in
order to form their digital identities, including their names, photos, etc. In the next
sections, we examine aspects of the digital identity formation process and analyze some
factors that may influence this process.
Conclusively, personal data disclosure is part of the digital identity formation
process which is associated with popularity: the digital identity of popular users is
characterized by high percentage of other people’s involvement in shaping it (Zhao
et al. 2008; van Dijck 2013). This creates a dilemma for the user: limiting third-party
access to personal information will eventually limit the digital identity formation and,
therefore, the user’s popularity. This dilemma in its extreme means that not partici-
pating in social media leads to a cost, which is equivalent to the lack of a digital
identity and a complete absence of the user from the online realm (Raynes-Goldie
2010); Turkle (2011) state “online networks can lead to isolation anxiety” and Cover
and Doak (2015) argue that “identity construction is becoming increasingly relational,
built on high quantity, low quality interactions”.
3 Research Design
Acquisti 2005; Krasnova et al. 2009; Youn 2009; Shin 2010; Xu et al. 2011). The
concerns expressed by users themselves are related to various aspects of their private
and social life. Besides concerns on the information that a user chooses to disclose,
users also express concerns about the lack of control on information that others share
about them (Krasnova et al. 2009; Shin 2010; Madden 2012). This is related to possible
feelings of embarrassment or shame if that content becomes visible to people outside
the close friendly environment (Krasnova et al. 2009; Zhang et al. 2010).
Relevant research shows that users may choose pseudonyms or multiple profiles as
a mechanism for privacy protection, although they express that this may be ineffective
given that their real identity can be revealed through dozens of other ways, such as
social networking lists or other media posts that refer to them (Raynes-Goldie 2010).
Interestingly, there are mixed conclusions if privacy concerns affect users’ infor-
mation disclosure behavior. While some studies argue that privacy concerns do not
affect users’ disclosure behavior (Stutzman 2006; Madden et al. 2007; Tufekci 2008),
others show that users actually reduce the amount of information they disclose due to
privacy concerns and adjust their SNS presence depending on perceived privacy
invasion threats (Krasnova et al. 2009; Raynes-Goldie 2010; Madden 2012).
From our point of view, if privacy concerns determine information disclosure
behavior, this eventually affects the digital identity formation. Thus, we formulate the
following research hypothesis:
H1: Increased privacy concerns lead to reduced identity-shaping information
disclosure.
Taking advantage of the features that each SNS offers, users can publish their
information, communicate with friends and share common interests. Different SNS
platforms offer different services and functionalities (Papacharissi 2009; Hughes et al.
2012; van Dijck 2013), which can explain why some SNS became more popular than
others during time or in different regions (Skeels and Grudin 2009). According to
Karapanos et al. (2016) and van Dijck (2013) these differences alter the overall user
experience that the individual gains from each SNS. It is not only the social interaction
functionalities by themselves (e.g., provision of chat), but also the user interface that
affect the user experience. Especially because of these unique features the various SNS
platforms offer, it has become common the last few years for the majority of people to
be subscribers of more than one SNS (Brandtzaeg 2012).
Therefore, we expect that the SNS would be a determinant factor for information a
user discloses (and thus the respective privacy concerns) and the formation of digital
identity and assume that:
H2: The SNS that an individual use, determines the formation of digital identity.
H3: The SNS that an individual use, affects their privacy concerns.
Table 1. (continued)
Constructs Items Sources
Privacy Concerns – Social PCS1: I am often concerned that I don’t Krasnova
Threats (5-point Likert scale) have control over the actions of other users et al.
PCS2: I am often concerned that someone (2009)
might purposefully embarrass me on SNS
PCS3: It often worries me that other users
might purposefully write something
embarrassing about me on SNS
PCS4: I am often concerned that other users
might take advantage of the information they
learned about me through SNS
Privacy Concerns – PCO1: I am often concerned that the SNS Krasnova
Organizational Threats (5- provider could store my information for the et al.
point Likert scale) next couple of years (2009)
PCO2: I sometimes feel anxious that the SNS
provider might know too much about me
PCO3: I am often concerned that the SNS
provider could share the information I
provide with third parties (e.g. marketing, HR
or government agencies)
PCO4: It often worries me that third parties
(e.g. marketing, HR, government agencies)
could use the information they have collected
about me from the SNS for commercial
purposes
Digital Identity perception (5- IDP: How much does each piece of profile Pempek,
point Likert scale) information below help you to express who et al.
you are to others in SNS? (2009)
Name, Gender, Photo, Friend List, Email
Address, Physical Address, Country, Phone
Number, Birthday, Hometown, Relationship
Status, Major, Sexual Orientation, Academic
Classification, Academic Status, School
Information, Job/Occupation, Political
beliefs, Group Affiliation, Interests, Favorite
Music, Favorite Books, Favorite Movies,
Favorite TV Shows, Personal Statement,
Web page
Identity information disclosure IDA1: I have a comprehensive profile on Krasnova
– Amount (5-point Likert SNS et al.
scale) IDA2: I find time to keep my profile up to (2009)
date
IDA3: I keep my friends updated about what
is going on in my life through SNS
IDA4: When I have something to say, I like
to share it on SNS
(continued)
Shaping Digital Identities in Social Networks 167
Table 1. (continued)
Constructs Items Sources
Identity information disclosure IDH1: I am always honest in the information Krasnova
– Honesty (5-point Likert I provide on SNS et al.
scale) IDH2: I am always truthful when I write (2009)
about myself on SNS
Identity information disclosure CIE: In my SNS profile I disclose: Stutzman
(Yes/No) Name, Gender, Photo, Friend List, Email (2006)
Address, Physical Address, Country, Phone Zhao et al.
Number, Birthday, Hometown, Relationship (2008)
Status, Major, Sexual Orientation, Academic Pempek
Classification, Academic Status, School et al.
Information, Job/Occupation, Political (2009)
beliefs, Group Affiliation, Interests, Favorite Ismail
Music, Favorite Books, Favorite Movies, (2010)
Favorite TV Shows, Personal Statement, Shafie et al.
Web page (2012)
We utilized the construct Privacy Concerns on SNS, adapted from Krasnova et al.
(2009). The construct includes two dimensions: (a) Concerns about Organizational
threats, which refer to users’ concerns on how SNS’ organizations may use their
information and (b) Concerns about Social threats which refer to possible negative
outcomes from an undesired exploitation of users’ information from their social
environment. In this study we included the most popular SNS in terms of traffic
(ALEXA 2017), namely Facebook, Twitter, LinkedIn, Google+, Tumblr, Instagram,
Pinterest, YouTube.
perception and digital identity elements are excluded, as they are structured in a dif-
ferent way and they were used for descriptive conclusions. The results confirmed our
grouping. The quality criteria of the constructs are shown in Table 2.
During the analysis of the results, all responses were examined in two ways. The
entire set of responses was the first basis for drawing conclusions, especially fields
related to demographic statistics, and then this set was split into smaller individual
subsets, nine in total, one for each user-preferred SNS, as the one they use most and is
selected in order to complete the questionnaire.
Concerning the construct that measures privacy concern (PC), the first question
(PC1) shows a lower contribution to the formation of the variable. This is due to the
variation of responses from one SNS to another. If the answers for a specific SNS, such
as Facebook, are isolated, the question rate (PC1) increases exponentially.
To test H2 hypothesis, responses were analyzed in two different ways to ensure the
validity of the result. In order to illustrate how digital identity is being formed, users
were asked to state which Common Identity Elements they disclose in their SNS
profile. A checkbox list of answers was given, with 27 available options, including
“something else” option. At first, all the responses were transformed into an absolute
number that represents the number of items they disclose. In the One-Way Analysis of
Variance, the dependent variable was the number of items disclosed by the user and the
predictive factor was the SNS. Subsequently, the answers to this question have been
coded in four categories so as to create a rating scale for the disclosure of digital
identity data. Users with little or no disclosed data were ranked in the first group (0–5
elements), those who disclosed several data were ranked in the second group (6–12
elements), then those who disclosed many elements (13–19 elements) and, finally, the
users with the maximum number of disclosed data (20–26 elements). These four groups
were created from a basis of 30 different options. 26 items were given to the question, 3
new items were declared by the users (Foreign Languages, Nickname and Sporting
Team), and zero option (no items disclosed) is also included. Responses ranged from 0
to 26 choices. Chi Square control was then applied.
For the third research hypothesis (H3) we initially used a comparative table of the
averages of variables related to privacy concerns. Then, we performed three variance
analyses with one factor, one for each of the privacy variables, and the categorizing
factor was the SNS.
The participants replied to the questionnaire according to a specific SNS they use,
in order to study the possible impact of each platform on how users form their digital
identity. It would be interesting to also study the possible differentiation of the same
user’s behavior among different SNS. However, in this case the extent of the ques-
tionnaire would make its completion unattainable, with the risk of low participation. As
not all users all the SNSs, or the same combination of them, the collected data would
not be comparable. Our findings are also bounded by the self-report limitations that
apply to human behavior studies which use surveys.
Shaping Digital Identities in Social Networks 169
We obtained 759 observations (N = 759), of whom 59,8% were women and 40,2%
were men. The first three age categories recorded the majority of respondents; 36,9%
were between 26 and 35 years old, 29,1% were between 36 and 45 years old and 21,7%
were younger than 25 years old. 67.3% of the participants were graduates of a higher
education institution (BSc, MSc, PhD). Regarding the demographics of SNSs user
170 T. Papaioannou et al.
base, Facebook holds the first place, with 97,2% of the respondents, followed by
YouTube and Instagram.
The distribution of SNS’s in the sample is presented in the Fig. 2.
Although we expected that privacy concerns affect the amount of information users
disclose on SNS (H1), nonetheless our findings do not support this. There are studies
that also did not find a strong relationship among them (Acquisti and Gross 2006;
Stutzman 2006; Dwyer et al. 2007; Tufekci 2008). They explained this finding based
on the unpredictable way in which human perceptions are transformed into actual
behavior (Tufekci 2008). As an example, there is a rhetorical statement in an interview
from a user who wondered why to keep a profile SNS active if it does not give enough
information for its owner and ultimately does not fulfill its role, which is effective
networking (Tufekci 2008). Facing this scenario, a common users’ reaction is to limit
the audience in which information is visible, despite the amount already disclosed. On
the users’ side, this explanation is also due to the prevailing norms on the use of SNS.
People who use SNS in their everyday life, adopt a set of behaviors, which in this case
are related to the disclosure of personal information. A paradoxical reason for this
behavior is also suggested by Palen and Dourish (2003), who argue that someone can
increase the amount of information they disclose in order to restrict the target audience;
by stating specific details about personality and preferences, those who are not inter-
ested or do not agree distance themselves.
choose to introduce themselves to other users of the platform and hence the formation
of their digital identity. On Fig. 4, the distribution of responses for each social network
in relation to the amount of published information, is demonstrated.
Concerning H3 research hypothesis, there was a large difference in average values
between some SNS platforms. For example, Facebook’s average on each of the three
privacy-related variables is significantly higher than those on LinkedIn, Twitter,
YouTube, Instagram and Pinterest. Figure 5 shows the average of the three variables
for each SNS. From the line graph we conclude that the average for the variable of
general privacy concerns is considerably higher than those of the other two variables,
independently of the SNS. It also appears that privacy threats related to the social
environment are considerably lower, also regardless of the used SNS. Finally, the SNS
that seems to raise most privacy concerns is Facebook, while LinkedIn appears to raise
less privacy concerns.
Analysis of variance results showed that there are statistically significant variations
with all three combinations, when the dependent variable was each of the three con-
structs of privacy concerns. Statistically significant differences are again observed in
the same pairs of SNS. In the variable of privacy concerns regarding social threats, we
notice the lowest averages among all SNS. The same variable has the largest deviations
from the highest to the lower limit (Facebook and LinkedIn respectively.
174 T. Papaioannou et al.
Overall, we can conclude that the SNS that the user utilizes affects users’ privacy
concerns. The platforms that users trust to disclose much and honest identity infor-
mation, such as LinkedIn, are ranked very low in privacy concerns rating. This finding
seems expected if we consider the reason why someone uses a network like LinkedIn.
The information provided by the user should be complete and accurate for job-oriented
networking to be effective. On the contrary, Facebook collects vast amounts of
information and seems to worry users about privacy violations. Although the platform
integrates privacy protection tools, concerns about privacy issues record the highest
averages. The support of H3 research hypothesis is also reinforced by the differenti-
ation of the contribution of questions in the formation of each variable. The contri-
bution factors of each question, separately for each SNS, are higher than those recorded
in the aggregate results.
5 Discussion
This paper explores digital identities’ formation in SNSs, by identifying the elements
that users perceive as constituents of their digital identities, as well as the role of
privacy concerns for forming them. As identification and analysis of the data elements
that constitute digital identities in SNS is scarce in the literature, this study provides an
understanding of how users perceive and form these identities.
Results of this research show that users perceive as pillar information for their
digital identity, in sequence of significance: their name, gender, photo, interests, job,
favorite music, academic classifications, and others (see Fig. 2). Our findings
Shaping Digital Identities in Social Networks 175
demonstrate that privacy concerns do not seem to affect the information that users
disclose when they form their digital identities. The paradox of this finding can be
explained by the unpredictable nature of the human personality and the way it is
expressed. This fluidity of human behavior in general, and especially in terms of
privacy behaviors, was reflected in the research that led to the formulation of the
privacy paradox. From a psychological viewpoint, the privacy paradox occurs when
users realize that they must trade-off between their privacy concerns and impression
management (Utz and Kramer 2009). This gives us an indication of why users in our
research differentiate the personal data they disclose when privacy concerns related to
their social environment are mediated. The ability of people to adapt to the circum-
stances and their tendency to make choices in pursuit of the best desired result outline
those cases where the user ignores presumable negative effects.
Through this research we have also identified how users’ perceptions and behaviors
differentiate across different SNS, including the diversity of information that users
choose to publish for shaping their digital identity in them. The SNS was found to be
determinant for both the information users choose for their digital identity and for their
privacy concerns. The large number of social networking platforms available today can
confirm the validity of this finding. Each of the SNS brings together specific features
and targets different audiences. The usage purpose, the user interface and the trust of
the user to the platform are factors that play an important role in shaping the conditions,
under which users adopt a social network. Furthermore, each SNS offers a different
combination of personal identity elements, from which the user can choose what to
disclose in order to form a unique digital identity. Consequently, users behave online in
a specific context, different for each social networking platform.
To the best of our knowledge, this is the first paper that examines digital identity
formation in SNS in relation to the type of SNS and privacy concerns. As mentioned in
the introduction, shaping digital identity using social networks is an internal process,
which is then influenced by such factors. Literature mentions as determinants of
identity formation, inter alia, user self-esteem/self-confidence (Forest and Wood 2012;
Emery et al. 2014; Marshall et al. 2015), the “Big Five” model of personality (Marshall
et al. 2015), narcissism (Mehdizadeh 2010), perceived benefits (Xu et al. 2013; Khan
and Shaikh 2007; Forman et al. 2008; Boyd and Ellison 2007), privacy sensitivity
(Yang and Wang 2009; Xu et al. 2013), information control (Xu et al. 2013) and
subjective rules such as social culture, trust and intimacy with other users (Lewis et al.
2008; Xu et al. 2013). However, these are factors that have been already researched and
validated by previous works. The combination of social media chosen by the user and
privacy concerns as regulators of shaping the digital identity is proposed in this paper
for the first time. The aforementioned factors can be further studied in the future, in
combination with those proposed in the present research, with a view to drawing more
detailed conclusions about human behavior online.
There are also limitations in our research. First, participants answered the questions
according to the SNS they chose. In a more advanced research approach, we could
study the differentiation of every user’s behavior among the social networks that they
use. However, in this case other methodological problems would arise, as the large
extent of the questionnaire would take a long time to complete, at the risk of low
participation. Moreover, all users do not use all social networks, nor the same
176 T. Papaioannou et al.
combination of them, which would lead to the collection of a very large number of
different cases that are not comparable to each other. Second, in our research, responses
for specific SNSs had few answers because they are not popular in Greece. Snapchat,
Tumblr and Flickr had less than 2% of the sample and therefore could not provide
evidence to support or not the corresponding findings of the literature.
Overall, the results of this research are consistent with findings in relevant literature and
contribute with new insights: Whilst privacy concerns about “social threats” affect the
amount of personal information disclosure, we have found that this is not the case with
general privacy concerns and those related to “organizational threats”. As Tufekci
(2008) argues, it seems that users are more concerned about being honest with their real
friends and prefer to keep unwanted audiences away using privacy settings. Moreover,
users tend to disclose a lot of their personal information, overcoming their privacy
concerns, in order to state their preferences clearly and avoid misconceptions. In the
same time, we need to consider a possible lack of concern about organizations, gov-
ernment etc., comparing to the social environment. The findings of this study can
motivate further research in users’ online behavior and can be of value for SNS
designers as they reveal the difficulty to understand the role of privacy within social
networking.
Further, we provide evidence that users have different privacy concerns per SNS
platform and choose different elements to formulate their digital identity in each one of
them. This finding can provide significant insights for the SNS providers and the
protective mechanisms that they enable. This study indicates that users worry differ-
ently about their privacy depending on the type of SNS. SNS providers can benefit
from this finding by differentiating their authentication methods to reassure their users
(e.g., our data implies that Facebook would benefit by enforcing stronger authentication
mechanisms as a remedy for the increased privacy concerns).
We also examined the role of privacy concerns for individuals’ perception and
information disclosure. Our findings confirm that users do have privacy concerns when
using SNS, which can be a useful finding for entities that design privacy awareness
campaigns (e.g., ENISA). These privacy concerns however were not found to prohibit
users from disclosure information that they perceive as important for expressing who
they are in SNS platforms.
Findings of this study can also trigger future research in the understanding of
human behavior across SNS. It would be interesting to examine how the same user may
differentiate her digital identity formation when moving from one SNS to another.
Acknowledgement. This research has been partially supported by the SECREDAS project,
which is co-funded by the ECSEL Joint Undertaking of the European Union under grant
agreement number 783119 and PDM&FC.
Shaping Digital Identities in Social Networks 177
References
Acquisti, A., Gross, R.: Imagined communities: awareness, information sharing, and privacy on
the Facebook. In: Danezis, G., Golle, P. (eds.) PET 2006. LNCS, vol. 4258, pp. 36–58.
Springer, Heidelberg (2006). https://doi.org/10.1007/11957454_3
Alexa Web Search: The top 500 sites on the web: Global (2017). Accessed 11 Apr 2017
Alvermann, D.E., Marshall, J.D., McLean, C.A., Huddleston, A.P., Joaquin, J., Bishop, J.:
Adolescents’ web-based literacies, identity construction, and skill development. Lit. Res.
Instr. 51(3), 179–195 (2012)
Barbour, K., Marshall, D.: The academic online: constructing persona through the World Wide
Web. First Monday 17(9) (2012). https://doi.org/10.5210/fm.v0i0.3969
Bareket-Bojmel, L., Moran, S., Shahar, G.: Strategic self-presentation on Facebook: personal
motives and audience response to online behavior. Comput. Hum. Behav. 55, 788–795 (2016)
Barnes, S.B.: A privacy paradox: social networking in the United States. First Monday 11(9)
(2006). https://firstmonday.org//article/view/1394/1312
Baym, N.K.: Personal Connections in the Digital Age. Polity, Malden (2010)
Belk, R.W.: Extended self in a digital world. J. Consum. Res. 40(3), 477–500 (2013)
Boyd, D.: Social network sites as networked publics: affordances, dynamics, and implications.
In: Papacharissi, Z. (ed.) A Networked Self: Identity, Community, and Culture on Social
Network Sites, pp. 39–58. Routledge, New York (2010)
Boyd, D.M., Ellison, N.B.: Social network sites: definition, history, and scholarship. J. Comput.-
Mediat. Commun. 13(1), 210–230 (2007)
Brandtzaeg, P.B.: Social networking sites: their users and social implications-A longitudinal
study. J. Comput.-Mediat. Commun. 17(4), 467–488 (2012)
Buchanan, T., Paine, C., Joinson, A.N., Reips, U.D.: Development of measures of online privacy
concern and protection for use on the Internet. J. Am. Soc. Inf. Sci. Technol. 58(2), 157–165
(2007)
Canary, D., Cody, M., Manusov, V.: Interpersonal Communication: A Goals-Based Approach,
pp. 2–49. Bedford/St. Martin’s, Boston (2003)
Cover, R., Doak, S.: Identity offline and online. In: International Encyclopedia of the Social &
Behavioral Sciences, pp. 547–553. Elsevier (2015)
Delahunty, J.: ‘Who am I?’: exploring identity in online discussion forums. Int. J. Educ. Res. 53,
407–420 (2012)
Dwyer, C., Hiltz, S.R., Passerini, K.: Trust & privacy concern within social networking sites: a
comparison of Facebook & MySpace. In: AMCIS 2007 Proceedings (2007)
D’Augelli, A.R.: Identity development and sexual orientation: toward a model of lesbian, gay,
and bisexual development. In: Trickett, E.J., Watts, R.J., Birman, D. (eds.) The Jossey-Bass
Social and Behavioral Science Series. Human Diversity: Perspectives on People in Context,
pp. 312–333. Jossey-Bass, San Francisco (1994)
Eastin, M.S., Brinson, N.H., Doorey, A., Wilcox, G.: Living in a big data world: predicting
mobile commerce activity through privacy concerns. Comput. Hum. Behav. 58, 214–220
(2016)
Emery, L.F., Muise, A., Dix, E.L., Le, B.: Can you tell that I’m in a relationship? Attachment and
relationship visibility on Facebook. Pers. Soc. Psychol. Bull. 40, 1466–1479 (2014)
Fieseler, C., Meckel, M., Ranzini, G.: Professional personae-how organizational identification
shapes online identity in the workplace. J. Comput.-Mediat. Commun. 20(2), 153–170 (2014)
Forest, A.L., Wood, J.V.: When social networking is not working: individuals with low self-
esteem recognize but do not reap the benefits of self-disclosure on Facebook. Psychol. Sci. 23,
295–305 (2012)
178 T. Papaioannou et al.
Forman, C., Ghose, A., Wiesenfeld, B.: Examining the relationship between reviews and sales:
the role of reviewer identity disclosure in electronic markets. Inf. Syst. Res. 19(3), 291–313
(2008)
Gee, J.: An Introduction to Discourse Analysis: Theory and Method. Routledge, Abingdon
(1999)
Goffman, E.: The Presentation of Self in Everyday Life. Doubleday, New York (1959)
Gross, R., Acquisti, A.: Information revelation and privacy in online social networks. In:
Proceedings of the 2005 ACM Workshop on Privacy in the Electronic Society, Alexandria,
VA, November 2005, pp. 71–80 (2005)
Hodkinson, P.: Bedrooms and beyond: youth, identity and privacy on social network sites. New
Media Soc. 19(2), 272–288 (2017)
Hoy, M.G., Milne, G.: Gender differences in privacy-related measures for young adult Facebook
users. J. Interact. Advert. 10(2), 28–45 (2010)
Hughes, D.J., Rowe, M., Batey, M., Lee, A.: A tale of two sites: Twitter vs. Facebook and the
personality predictors of social media usage. Comput. Hum. Behav. 28(2), 561–569 (2012)
Ismail, S.: An evaluation of students’ identity-sharing behavior in social network communities as
preparation for knowledge sharing. Int. J. Adv. Sci. Arts 1(1), 14–24 (2010)
Joinson, A.N.: “Looking at”, “Looking up” or “Keeping up with” people? Motives and uses of
Facebook. In: Proceedings of CHI. ACM, New York (2008)
Joinson, A., Reips, U., Buckhanan, T., Paine Schofield, C.: Privacy, trust and self-disclosure
online. Hum.-Comput. Interact. 25, 1–24 (2010)
Junco, R.: Engaging Students Through Social Media: Evidence-Based Practices for Use in
Student Affairs, p. 113. Wiley, Hoboken (2014)
Karapanos, E., Teixeira, P., Gouveia, R.: Need fulfillment and experiences on social media: a
case on Facebook & WhatsApp. Comput. Hum. Behav. 55, 888 (2016)
Khan, J.I., Shaikh, S.S.: Computing in social networks with relationship algebra. J. Netw.
Comput. Appl. 31, 862–878 (2008)
Kimmons, R.: Social networking sites, literacy, and the authentic identity problem. TechTrends
58(2), 93–98 (2014)
Krasnova, H., Gunther, O., Spiekermann, S., Koroleva, K.: Privacy concerns and identity in
online social networks. Identity Inf. Soc. 2(1), 39–63 (2009)
Krishnamurthy, B., Wills, C.E.: Characterizing privacy in online social networks. In: Proceedings
of the First Workshop on Online Social Networks, pp. 37–42, Seattle, WA, August 2008
(2008)
Larsen, M.C.: Understanding social networking: on young people’s construction and co-
construction of identity online. Internet Res. 8, 18–36 (2007)
Lewis, K., Kaufman, J., Christakis, N.: The taste for privacy: an analysis of college student
privacy settings in an online social network. J. Comput.-Mediat. Commun. 14, 79–100 (2008)
Madden, M., Fox, S., Smith, A., Vitak, J.: Digital Footprints: Online Identity Management &
Search in the Age of Transparency. Pew Internet & American Life Project, Washington, DC
(2007)
Madden, M.: Privacy management on social media sites. Pew Internet Report, pp. 1–20 (2012)
Marshall, T.C., Lefringhausen, K., Ferenczi, N.: The Big Five, self-esteem, and narcissism as
predictors of the topics people write about in Facebook status updates. Pers. Individ. Differ.
85, 35–40 (2015)
Marwick, A.: Online identity. In: A Companion to New Media Dynamics, pp. 355–364.
Blackwell Publishing Ltd. (2013)
Mehdizadeh, S.: Self-presentation 2.0: narcissism and self-esteem on Facebook. Cyberpsychol.
Behav. Soc. Netw. 13(4), 357–364 (2010)
Shaping Digital Identities in Social Networks 179
Michikyan, M., Dennis, J., Subrahmanyam, K.: Can you guess who I am? Real, ideal, and false
self-presentation on Facebook among emerging adults. Emerg. Adulthood 3(1), 55–64 (2015)
Norberg, P.A., Horne, R., Horne, A.: The privacy paradox: personal information disclosure
intentions versus behaviors. J. Consum. Aff. 41(1), 100–126 (2007)
Palen, L., Dourish, P.: Unpacking “privacy” for a networked world. In: Proceedings of the ACM
Conference on Human Factors in Computing Systems, pp. 129–136. Association for
Computing Machinery, New York (2003)
Papacharissi, Z.: The virtual geographies of social networks: a comparative analysis of Facebook,
LinkedIn and ASmallWorld. New Media Soc. 11(1–2), 199–220 (2009)
Papacharissi, Z.: A Networked Self: Identity. Community and Culture on Social Network Sites.
Routledge, Abingdon (2010)
Pempek, T.A., Yermolayeva, Y.A., Calvert, S.L.: College students’ social networking
experiences on Facebook. J. Appl. Dev. Psychol. 30(3), 227–238 (2009)
Ramírez, G., Palu-ay, L.: “You don’t look like your profile picture”: the ethical implications of
researching online identities in higher education. Educ. Res. Eval. 21(2), 139–153 (2015)
Raynes-Goldie, K.: Aliases, creeping and wall cleaning: understanding privacy in the age of
Facebook. First Monday 15(1–4) (2010). https://firstmonday.org/article/viewArticle/2775/
2432
Ross, C., Orr, E.S., Sisic, M., Arseneault, J.M., Simmering, M.G., Orr, R.R.: Personality and
motivations associated with Facebook use. Comput. Hum. Behav. 25(2), 578–586 (2009)
Salimkhan, G., Manago, A., Greenfield, P.: The construction of the virtual self on MySpace.
Cyberpsychol.: J. Psychosoc. Res. Cyberspace 4(1) (2010). https://cyberpsychology.eu/
article/view/4231/3275
Shafie, L., Nayan, S., Osman, N.: Constructing identity through Facebook profiles: online
identity and visual impression management of university students in Malaysia. Soc. Behav.
Sci. 65, 134–140 (2012)
Shin, D.H.: The effects of trust, security and privacy in social networking: a security-based
approach to understand the pattern of adoption. Interact. Comput. 22(5), 428–438 (2010)
Skeels, M., Grudin, J.: When social networks cross boundaries: a case study of workplace use of
Facebook and LinkedIn. In: Proceedings of the GROUP 2009, pp. 95–104 (2009)
Stone, A.A.: The War of Desire and Technology at the Close of the Mechanical Age. MIT Press,
Cambridge (1996)
Stutzman, F.: An evaluation of identity-sharing behavior in social network communities. Int.
Digit. Media Arts J. 3(1), 10–13 (2006)
Subrahmanyam, K., Šmahel, D.: Constructing identity online: identity exploration and self-
presentation. In: Subrahmanyam, K., Smahel, D. (eds.) Digital Youth. Advancing Respon-
sible Adolescent Development, pp. 59–80. Springer, New York (2011). https://doi.org/10.
1007/978-1-4419-6278-2_4
Tufekci, Z.: Can you see me now? Audience and disclosure regulation in online social network
sites. Bull. Sci. Technol. Soc. 28(1), 20–36 (2008)
Turkle, S.: Alone Together: Why We Expect More from Technology and Less from Each Other.
Basic Books, New York (2011)
Utz, S., Kramer, N.: The privacy paradox on social network sites revisited: the role of individual
characteristics and group norms. Cyberpsychol. J. Psychosoc. Res. Cyberspace 3(2) (2009).
Article 1. https://cyberpsychology.eu/article/view/4223/3265
Van Bavel, J.J., Pereira, A.: The partisan brain: an identity-based model of political belief. Trends
Cogn. Sci. 22(3), 213–224 (2018)
Van Dijck, J.: ‘You have one identity’: performing the self on Facebook and LinkedIn. Media
Cult. Soc. 35(2), 199–215 (2013)
180 T. Papaioannou et al.
Wang, S.S.: “I share, therefore I am”: personality traits, life satisfaction, and Facebook check-ins.
Cyberpsychol. Behav. Soc. Netw. 16(12), 870–877 (2013)
Wessels, B.: Identification and the practices of identity and privacy in everyday digital
communication. New Media Soc. 14(8), 1251–1268 (2012)
Xu, H., Dinev, T., Smith, J., Hart, P.: Information privacy concerns: linking individual
perceptions with institutional privacy assurances. J. Assoc. Inf. Syst. 12(12), 798 (2011)
Xu, F., Michael, K., Chen, X.: Electron. Commer. Res. 13(2), 151–168 (2013)
Yang, S., Wang, K.: The influence of information sensitivity compensation on privacy concern
and behavioral intention. Data Base Adv. Inf. Syst. 40(1), 38–51 (2009)
Youn, S.: Determinants of online privacy concern and its influence on privacy protection
behaviors among young adolescents. J. Consum. Aff. 43(3), 389–418 (2009)
Zhang, C., Sun, J., Zhu, X., Fang, Y.: Privacy and security for online social networks: challenges
and opportunities. IEEE Netw. 24(4), 13–18 (2010)
Zhao, S., Grasmuck, S., Martin, J.: Identity construction on Facebook: digital empowerment in
anchored relationships. Comput. Hum. Behav. 24, 1816–1836 (2008)
GDPR Compliance: Proposed Technical
and Organizational Measures
for Cloud Providers
1 Introduction
Cloud computing is a technology with huge spread in many aspects and areas of ICT.
A lot of research has been conducted over the past years on security, privacy and trust
issues of cloud environments and especially on shared clouds. This research is still
open, imposing, even today, obstacles in the adoption of cloud. In addition to that, from
May 2018 cloud providers should comply with GDPR. The General Data Protection
Regulation (GDPR) has a clear goal: to introduce a higher, more consistent level of
personal data protection across the European Union, which will give citizens back
control over their personal data and simplify the regulatory environment for business.
The regulation, applies to all companies that hold or process EU residents’ data,
including cloud computing users, providers and their sub-contractors. The existing
National legal framework, based on the 95/46 EU Data Protection Directive, has not
achieved harmonization of personal data protection rules between member states.
These variations, and at times conflicting rules, are complicating businesses’ require-
ments and procedures, especially as data increasingly flows across borders in today’s
digital age. By implementing it as a regulation, the GDPR aims to ensure that the same
data protection rules will apply uniformly across the EU. In addition, while many of the
GDPR’s concepts and principles have been based on the 95/46 Data Protection
Directive, it introduces significant new rules and enhancements. The emphasis is on
how personally identifiable information (PII) is handled and protected by institutions
within the EU—and, in certain cases, outside the EU. For the cloud providers, the new
obligations are extensive and challenging. The purpose of this paper is to give a brief
guidance on what a cloud provider should consider and what further actions to take in
order to comply with GDPR [2].
Section 2 describes in detail the requirements for GDPR compliance of cloud
computing environments. Section 3, presents the GDPR roles (data controller and data
processor) in a cloud environment and discusses the applicability of GDPR compliance
requirements for each cloud architecture (IAAS, PAAS, SAAS). Furthermore, Sect. 3
proposes countermeasures for satisfying the aforementioned requirements. Section 4
concludes the paper.
2 GDPR Requirements
collected. Cloud controllers must keep personal data accurate and up to date. When the
data are not any more required, in relation to the initial processing purpose, they should
be immediately erased, thus conforming to the storage limitation principle. Finally,
integrity and confidentiality should be reassured to avoid unauthorized or unlawful
processing or/and accidental loss, destruction or damage.
To conform with the above cloud providers must maintain full documentation of
personal data held, where it came from and with whom they are shared with, including
the reason of processing. Data minimization should be considered in the organization
and the purpose of collecting information should be defined in the security policy.
Scheduled data re-evaluation should be performed periodically. Furthermore, to ensure
purpose limitation it is necessary to perform periodic audits to cloud clients and
employees. Also, periodic data accuracy compliance checks should be done. Finally,
vital for cloud providers is to apply storage limitation scan mechanisms and transfer
restriction. To ensure integrity and confidentiality, data encryption, encrypted net-
works, firewall, data fragmentation, and anonymization techniques should be utilized.
Pseudoanonymization, a privacy enhancing technique, should also be implemented if
possible, avoiding immediate linkability of data to the data subjects. In terms of
accountability and lawfulness of processing, appropriate audit mechanisms on data
operations (access, edit, delete, export etc.) are proposed to be implemented. The
legitimate interest should be documented and included with accurate, clear and specific
terms in the Service Level Agreement – SLA.
2.3 Consent
Cloud providers that collect/process any form of personal data need always a legal
basis. In certain cases this legal basis can be the consent of the data subject. In other
words, the cloud controller needs at any time to be able to demonstrate that the data
subject has consented to the processing of his or her personal data. If the data subject’s
consent is given in the context of a written declaration which also concerns other
matters, the request for consent shall be presented in a manner which is clearly dis-
tinguishable from the other matters, in an intelligible and easily accessible form, using
clear and plain language. Any part of such a declaration which constitutes an
infringement of this Regulation shall not be binding. The data subject shall have the
right to withdraw his or her consent at any time. The withdrawal of consent shall not
affect the lawfulness of processing based on consent before its withdrawal. Prior to
giving consent, the data subject shall be informed thereof. It shall be as easy to
withdraw as to give consent.
The consent management mechanisms should be supported through some software
application which will support the provision, updating, revoking and maintaining of
users’ consents. Restriction on clear and plain language consent should be included in
order to be intelligible and easily accessible (e.g. native language of data subject).
Alerting for updating the users’ consents will be necessary when a change in the
purpose or manner of personal data processing is happening.
184 Z. Georgiopoulou et al.
Regarding information notices a suggestion is to adopt some tool for generating and
automatically sharing, template documents for the information notices, requests and
responses.
group of people (including analysis or prediction). The guidelines make it clear that the
definition is very broad and that the processing does not need to involve inference to be
caught – “simply assessing or classifying individuals based on characteristics such as
their age, sex, and height could be considered profiling, regardless of any predictive
purpose” [7]. The guidelines describe profiling as having three distinct stages each of
which fall within the GDPR definition of profiling: (1) data collection (2) automated
analysis to identify correlations and (3) applying the correlation to an individual to
identify characteristics of present or future behavior. A decision based solely on
automated processing is a decision with no human involvement in the decision process.
The guidelines warn that involving a human in the process to circumvent the rules on
solely automated decision making would not work, as the human involvement must be
meaningful and not just a token gesture. The individual needs to have the authority to
change the decision considering all the information available.
Individuals must be told when a decision has been taken solely using automated
decision making and they must have the right to request a review of the decision. The
review should be done by a person with appropriate authority and capacity to change
the decision and should involve a thorough review of all relevant data and any addi-
tional information provided by the individual. Organizations using automated decision
making should also carry our regular reviews and use appropriate procedures to prevent
errors.
of daily processes and comply with local legislations. The above processing can be
shared with sub-contractors only if the IaaS publishes the list of subcontractors
accompanied with full documentation and reasoning of sharing.
To support the transparency principle, audit mechanisms must be provided by cloud
providers, recording in an automatic way the requested resources, the users and the
sources of activity.
Purpose limitation in an Infrastructure as a Service cloud architecture can be
supported by splitting the infrastructure into individual clusters. Cloud hardware
resources provided to a data controller, should be isolated to avoid flow of personal
data information. Cloud providers should also be able to offer to their customers the
ability to create virtual cloud networks and thus facilitate communication between the
isolated resources and at the same time supporting isolation from public internet.
Accuracy, from the side of an IAAS cloud provider, should be supported by
offering relevant tools to their customers (data controllers). Software scanning must be
enabled to actively monitor data content, integrity and automatically generate alerts to
customer for malicious data. Encryption techniques can also help in the accuracy.
“GDPR can support integrity and confidentiality to ensure that the appropriate
security of personal data against unauthorized or unlawful processing and against
accidental loss, destruction or damage…” Article 5(1)(f). Hardware and network level
access control in a cloud infrastructure is proposed to comply with GDPR with the
concept of least privilege. Encryption on storage can also help in security of personal
data applying encryption on block level, object level and metadata with separate keys
and up to date technologies.
An IaaS is not obliged to have the consent of data subjects since they do not have
direct relationship with the service offered. The same logic applies to parental consent.
The only part that could be related to but without any required obligation is to take
extra measures in the authentication process to mitigate the unauthorized access
especially for non-adults.
In terms of Sensitive data an IaaS cloud provider should comply with all the
aforementioned data protection measures and have in place a software for scanning
data files in order to quickly identify sensitive information stored in their datastores.
Information notices is not an obligation of GDPR for cloud providers. Only for the
offered marketing and client support services they could employ a mechanism for
generating automatically templates for the documents that the cloud controllers need to
provide to authorities and to data subjects.
Subject access rectification and portability falls under IaaS oligarchy to GDPR.
IaaS providers must provide information regarding the data processed, the possible data
transfers with the relevant recipients, including information regarding rectification and
reassure. To support this, providers needs to have in place mechanisms for supporting
data portability through the appropriate export mechanisms.
The right to object, erasure and restriction is mostly relevant to cloud processors
and not providers-controllers themselves. The only case that an IaaS provider may
request a cloud processor to embed network restrictions and isolations for specific parts
of the infrastructure is after a data subject objects to a cloud processor. Also, a software
for secure erasure must be in place in the Infrastructure used.
Profiling and automated decision-taking is out of scope for an IaaS Cloud provider.
GDPR Compliance: Proposed Technical and Organizational Measures 189
Data subject consent management is not a requirement that a PaaS cloud provider
should comply with. Some big PaaS providers offer supporting tools to collect consent.
Parental consent is also not required.
PaaS cloud providers must employ tools for sensitive data identification and rele-
vant measures for classifying and protecting them. Data retention tools may be also
necessary. Rule based controls could be also provided to alert the administrators of the
cloud processor that a PaaS user stores information that has been classified as sensitive.
Regarding sensitive date it is strongly recommended to have audit controls against
global data privacy standards such as ISO 27018.
In terms of information notices, PaaS providers need to maintain full documenta-
tion of their platforms and of the security mechanisms they employ to support requests
for information notices from authorities.
PaaS Cloud users maintain the right of access, rectification and portability under the
enforcement of GDPR. All information hosted in PaaS environments must be exportable
in a universal and readable way through tools that the cloud service provider will supply.
Right to object is also applicable to PaaS providers and they must provide tools for
restriction of data storage, retention or deletion after the data subject objects.
Profiling and automated decision-taking is out of scope for a PaaS Cloud provider.
Furthermore, a PaaS cloud provider is subject to accountability, security and breach
notification. In the security measures, it is imperative to include firewalls and network
protection tools. Incident management should be also in place to actively monitor data
breaches and thus to support the notification of supervisory authorities in 72 h. Stan-
dard certification of the infrastructures provided could help (e.g. ISO 27001) (Table 2).
administrators when EU citizens are using the software to notify software owners that
they should comply to GDPR [8].
Data protection principles to SaaS cloud providers apply to all layers of cloud from
physical protection, infrastructure and up to software data protection. Measures
regarding physical protection must cover unauthorized access of personnel in the data
centers, including physical access control mechanisms like cards, cameras and bio-
metrics. On network level, encryption mechanisms are suggested, and it is necessary to
use firewalls, Intrusion Detection and Intrusion Prevention controls. Furthermore,
network segregation is required, mac filtering and network access control. Storage
encryptions techniques should be on level of hardware applying also on the same level
software scan tools. On the software data level, the best practices proposed by global
standards should be applied. Vulnerability assessments of the software and penetration
testing must be periodically conducted. Software must include audit mechanisms to log
and alert for data view, usage and edit with alert customization rules. It is also
imperative to apply encryption between communication and storage levels in terms of
database and hardware itself. Pseudonymization and anonymization is also proposed
when applicable [6].
Software delivered in form of SaaS is not necessary to embed consent management
techniques. It would be useful and probably recommended, but it is not obligatory since
it is out of the scope of the software itself. The same applies to children consent but it
would be useful to have a way for parental consent.
Sensitive data and relevant documentation of where it could be stored from soft-
ware partition level, up to physical infrastructure required would be more than obli-
gatory. It is also required to have in place encryption techniques and all the measures
referred above on data protection levels.
In terms of information notices there is a need to maintain full documentation of
their platforms and security mechanisms applied to support requests for information
notices from authorities.
Subject access, rectification and portability must be fully documented by data
controllers, but SaaS data processors must give to controller’s tools to maintain this
information. It is also imperative to have mechanisms for data export to support
portability in universal formats.
SaaS providers should avoid profiling users based on their sensitive information
whether these are directly collected from them or inferred as part of their undergoing
automated profiling. Data minimization principle should drive service design as data
controllers should be able to understand the minimum amount of data you will need for
it. The best way for doing that is to consider Data Protection by Design and by Default,
building services always examining what data are strictly needed, how to use them and
why. Do not experiment with algorithms and training models by collecting first data
and then decide how to use them, but rather only use well tested models that you know
will suffice to your scope, before deploying them [3].
Accountability, security and breach notification notices should be raised from SaaS
providers to let know the data controller about the leakage. Intrusion detection must be
included in the infrastructure of SaaS and relevant DPO must react. Full documentation
of data leakage and audit information should also be included (Table 3).
192 Z. Georgiopoulou et al.
4 Conclusions
The following Table 4 provides a summary of the requirements for cloud providers
acting as data Processors.
The following Table 5 provides a summary of the requirements for cloud providers
acting as data Controllers.
This paper has provided an overview of the security measures that could be
engaged in the process of GDPR compliance in cloud computing environments. For
future work we plan to research on what level trust management can influence GDPR.
Acknowledgment. This work has been partially supported by the Research Center of the
University of Piraeus.
References
1. Bird & Bird, Guide to the General Data Protection Regulation, January 2017
194 Z. Georgiopoulou et al.
2. Cloud Security Alliance, Code of Conduct for GDPR Compliance, November 2017. https://
downloads.cloudsecurityalliance.org/assets/research/gdpr/CSA_Code_of_Conduct_for_
GDPR_Compliance.pdf
3. Regulation (EU) 2016/679 of the European Parliament and of the Council, The European
Parliament and the Council of the European Union, 27 April 2016. https://eur-lex.europa.eu/
legal-content/EN/TXT/PDF/?uri=CELEX:32016R0679&qid=1485368166820&from=en
4. Microsoft, Accelerate GDPR compliance with the Microsoft Cloud, July 2017
5. Amazon Web Services, Navigating GDPR Compliance on AWS (2018)
6. Google Cloud Whitepaper, General Data Protection Regulation (GDPR), May 2018
7. Deloitte, Data Privacy in the cloud, September 2015
8. LexisNexis, GDPR and codes of conduct in SaaS, January 2019
9. Oracle Cloud Infrastructure, Oracle Cloud Infrastructure and the GDPR, European Union
General Data Protection Regulation, April 2018
On the Applicability of Security and
Privacy Threat Modeling for Blockchain
Applications
Dimitri Van Landuyt(B) , Laurens Sion, Emiel Vandeloo, and Wouter Joosen
1 Introduction
Threat modeling [24,25,29] involves the systematic investigation of threats to
software-intensive systems, and is performed by evaluating generic and well-
known threat types or threat categories. Threat elicitation approaches such as
Microsoft’s STRIDE [25] and LINDDUN [8] act upon an end-to-end system
model encoded in a Data Flow Diagram (DFD) and rigorously examine the
applicability and likelihood of threat types at an intricate level of granularity in
a software architecture. Given their versatility, these methods are increasingly
adopted in practice, and different tool implementations are in existence [19,24,
26]. Essentially distinguishing between user entities, processing elements, data
storage elements, and data flows, DFDs are created with abstractions that are
particularly well-aligned to traditional, centralized, multi-tier architectures.
This raises the key question whether existing threat elicitation methods and
tools can be applied successfully to more contemporary decentralized architec-
tures. In this paper, we evaluate this question in the context of distributed
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 195–203, 2020.
https://doi.org/10.1007/978-3-030-42048-2_13
196 D. Van Landuyt et al.
2 Background
This section first introduces threat modeling, and then provides a brief intro-
duction to the main principles behind blockchains and distributed ledgers.
Threat Modeling. Threat modeling refers to a number of architecture-centric
analysis approaches to identify security- and privacy-related design flaws, poten-
tial threats to the correct and fair workings of a software-intensive system.
Threat Modeling for Blockchain Applications 197
STRIDE [25] and LINDDUN [8] are acronyms1 of threat categories for respec-
tively security and privacy. There threat categories are further refined into more
concrete threat types in so-called threat trees. In these elicitative threat model-
ing approaches, the system under analysis is modeled as a Data Flow Diagram
(DFD) constituting of external entities, processes, data stores, and data flows
between these. Every element in a DFD is systematically investigated for each
of the threat types via a mapping table that indicates whether a threat type of
a specific threat category is applicable to an element of a DFD model element
type.
Threat modeling is widely used in industry as a cornerstone of secure devel-
opment life-cycles (SDLs) and the importance of security- and privacy-by-design
principles is increasingly recognized. Existing tool implementations [19,26] high-
light and exploit the elicitative nature of these approaches; i.e. they systemat-
ically iterate over the architectural elements encoded in the system DFD and
the threat trees to identify and generate concrete threats (this is the threat
elicitation step). In practical implementations, distinction is made between per-
element [8,25] and per-interaction [25,27] threat elicitation, in which threats are
identified respectively at the level of individual DFD elements or specifically at
the level of interactions in the system (data flows).
Blockchain Architecture and Principles. Blockchain architectures are
decentralized data stores in which many participants (nodes) hold a copy of
shared data structure, and cooperate in a peer-to-peer fashion to reach con-
sensus on a linked list of transactions (a digital transaction history, also called
a ledger ). Emerging from the context of digital crypto-currencies (and it best-
known instance Bitcoin [20]), a blockchain is a shared data store among many
participants, that is incrementally maintained and kept consistent through trans-
actions. Individual participants (nodes) collect and validate transactions and
package them in blocks.
To accept a block on the blockchain, distributed consensus is to be attained.
In proof-of-work (PoW) based blockchain systems, this consensus is reached
via the mechanism of mining. Miner participants are financially incentivized to
solve mathematical and computationally-intensive problems. This in turn pro-
vide them with authority to append a new block to the blockchain, which is then
shared in gossip-style among all peers in the peer-to-peer network.
1
STRIDE: Spoofing, Tampering, Repudiation, Information Disclosure, Denial of Ser-
vice, Elevation of Privilege;
LINDDUN: Linkability, Identifiability, Non-repudiation, Detectability, Disclosure of
information, Unawareness, Non-compliance.
198 D. Van Landuyt et al.
(i) exploiting race conditions in the consensus protocol, (ii) colluding among
participants (miners) to obtain disproportional gains or to gain control over the
network (51% attacks), and (iii) attacks to the network itself (denial of service,
isolating nodes, etc).
Privacy Threat Types. Blockchains are essentially public, transparent, and
shared data stores onto which transactions are processed in a pseudonymous
manner (at the basis of addresses). The public nature leads to a number of
privacy threat types related to identifying, linking, and accumulating information
about how these addresses are used. As shown in the bottom half of Table 1, we
distinguish between threats that act upon the information on the blockchain
(blockchain analysis threats), upon transaction meta-data (e.g., addresses or
pseudonyms used), and privacy threats related to key management.
Table 2. Categorical scale defined to assess the compatibility of STRIDE and LIND-
DUN with the identified blockchain-specific threat types.
4 Compatibility Assessment
To systematically investigate the compatibility of the existing threat modeling
frameworks with the threat types discussed in the previous section, we have
established the categorical scale presented in Table 2.
These criteria specifically aim at identifying (i) issues related to the expres-
sivity of the DFD notation and incompatibilities that may arise in threat elici-
tation itself (scores 0, 2), (ii) the a-priori compatibility of the threat categories
of STRIDE, respectively LINDDUN (scores 1, 3), or more specifically, (iii) com-
patibility with the existing threat trees (scores 4, 5).
Based on extensive expertise with the STRIDE and LINDDUN threat model-
ing approaches, and our assessment of the impact on existing tooling efforts [28],
200 D. Van Landuyt et al.
Fig. 1. DFD of miner activities. Empty entities represent different participants (peers)
in the distributed ledger.
5 Conclusion
References
1. Almashaqbeh, G., Bishop, A., Cappos, J.: ABC: a cryptocurrency-focused threat
modeling framework. arXiv preprint arXiv:1903.03422 (2019)
2. Androulaki, E., Karame, G.O., Roeschlin, M., Scherer, T., Capkun, S.: Evaluating
user privacy in bitcoin. In: Sadeghi, A.-R. (ed.) FC 2013. LNCS, vol. 7859, pp.
34–51. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-39884-1 4
3. Biryukov, A., Khovratovich, D., Pustogarov, I.: Deanonymisation of clients in Bit-
coin P2P network. In: Proceedings of the 2014 ACM SIGSAC Conference on Com-
puter and Communications Security, pp. 15–29. ACM (2014)
202 D. Van Landuyt et al.
4. Biryukov, A., Pustogarov, I.: Bitcoin over Tor isn’t a good idea. In: 2015 IEEE
Symposium on Security and Privacy, pp. 122–134. IEEE (2015)
5. Carson, B., Romanelli, G., Walsh, P., Zhumaev, A.: Blockchain beyond the hype:
what is the strategic business value. McKinsey & Company (2018)
6. Conti, M., Kumar, E.S., Lal, C., Ruj, S.: A survey on security and privacy issues
of Bitcoin. IEEE Commun. Surv. Tutorials 20(4), 3416–3452 (2018)
7. Courtois, N.T., Bahack, L.: On subversive miner strategies and block withholding
attack in Bitcoin digital currency. arXiv preprint arXiv:1402.1718 (2014)
8. Deng, M., Wuyts, K., Scandariato, R., Preneel, B., Joosen, W.: A privacy threat
analysis framework: supporting the elicitation and fulfillment of privacy require-
ments. Requirements Eng. 16(1), 3–32 (2011)
9. DuPont, J., Squicciarini, A.C.: Toward de-anonymizing Bitcoin by mapping users
location. In: Proceedings of the 5th ACM Conference on Data and Application
Security and Privacy, pp. 139–141. ACM (2015)
10. Eskandari, S., Clark, J., Barrera, D., Stobert, E.: A first look at the usability of
Bitcoin key management. arXiv preprint arXiv:1802.04351 (2018)
11. Eyal, I., Sirer, E.G.: Majority is not enough: Bitcoin mining is vulnerable. Commun.
ACM 61(7), 95–102 (2018)
12. Finney, H.: Best practice for fast transaction acceptance-how high is the risk (2011)
13. Gervais, A., Capkun, S., Karame, G.O., Gruber, D.: On the privacy provisions
of bloom filters in lightweight Bitcoin clients. In: Proceedings of the 30th Annual
Computer Security Applications Conference, pp. 326–335. ACM (2014)
14. Heilman, E., Kendler, A., Zohar, A., Goldberg, S.: Eclipse attacks on Bitcoin’s
peer-to-peer network. In: 24th USENIX Security Symposium (2015)
15. Karame, G.O., Androulaki, E., Capkun, S.: Double-spending fast payments in bit-
coin. In: Proceedings of the 2012 ACM Conference on Computer and Communica-
tions Security, pp. 906–917. ACM (2012)
16. Koshy, P., Koshy, D., McDaniel, P.: An analysis of anonymity in Bitcoin using
P2P network traffic. In: Christin, N., Safavi-Naini, R. (eds.) FC 2014. LNCS, vol.
8437, pp. 469–485. Springer, Heidelberg (2014). https://doi.org/10.1007/978-3-
662-45472-5 30
17. Kwon, Y., Kim, D., Son, Y., Vasserman, E., Kim, Y.: Be selfish and avoid dilemmas:
fork after withholding (FAW) attacks on Bitcoin. In: Proceedings of the 2017 ACM
SIGSAC Conference on Computer and Communications Security. ACM (2017)
18. Meiklejohn, S., et al.: A fistful of Bitcoins: characterizing payments among men
with no names. In: Proceedings of the 2013 Conference on Internet Measurement
Conference, pp. 127–140. ACM (2013)
19. Microsoft Corporation: Microsoft Threat Modeling Tool 2016 (2016). http://aka.
ms/tmt2016
20. Nakamoto, S., et al.: Bitcoin: a peer-to-peer electronic cash system (2008)
21. Nick, J.D.: Data-driven de-anonymization in Bitcoin. Master’s thesis, ETH-Zürich
(2015)
22. Reid, F., Harrigan, M.: An analysis of anonymity in the Bitcoin system. In: Alt-
shuler, Y., Elovici, Y., Cremers, A., Aharony, N., Pentland, A. (eds.) Security and
Privacy in Social Networks, pp. 197–223. Springer, New York (2013). https://doi.
org/10.1007/978-1-4614-4139-7 10
23. Rosenfeld, M.: Analysis of bitcoin pooled mining reward systems. arXiv preprint
arXiv:1112.4980 (2011)
24. Shevchenko, N., Chick, T.A., O’Riordan, P., Scanlon, T.P., Woody, C.: Threat
modeling: a summary of available methods (2018)
Threat Modeling for Blockchain Applications 203
25. Shostack, A.: Threat Modeling: Designing for Security. Wiley Publishing, Indi-
anapolis (2014)
26. Sion, L., Van Landuyt, D., Yskout, K., Joosen, W.: SPARTA: security & privacy
architecture through risk-driven threat assessment. IEEE (2018)
27. Sion, L., Wuyts, K., Yskout, K., Van Landuyt, D., Joosen, W.: Interaction-based
privacy threat elicitation. In: International Workshop on Privacy Engineering
(2018)
28. Sion, L., Yskout, K., Van Landuyt, D., Joosen, W.: Solution-aware data flow dia-
grams for security threat modelling. In: SAC 2018: Proceedings of the 33rd Annual
ACM Symposium on Applied Computing, pp. 1425–1432, April 2018. https://doi.
org/10.1145/3167132.3167285
29. Tuma, K., Calikli, G., Scandariato, R.: Threat analysis of software systems: a
systematic literature review. J. Syst. Softw. 144, 275–294 (2018)
Privacy, Security, Legal and Technology
Acceptance Requirements for a GDPR
Compliance Platform
Abstract. GDPR entered into force in May 2018 for enhancing user data
protection. Even though GDPR leads towards a radical change with many
advantages for the data subjects it turned out to be a significant challenge.
Organizations need to make long and complex changes for the personal data
processing activities to become GDPR compliant. Citizens as data subjects are
empowered with new rights, which however they need to become aware of and
understand. Finally, the role of data protection authorities changes as well as
their expectations from organizations. GDPR compliance being a challenging
matter for the relevant stakeholders calls for a software platform that can support
their needs. The aim of the Data govErnance For supportiNg gDpr (DEFeND)
EU Project is to deliver such a platform. To succeed, the platform needs to
satisfy legal and privacy requirements, be effective in supporting organizations
in GDPR compliance, and provide functionalities that data controllers request
for supporting GDPR compliance. Further, it needs to satisfy acceptance
requirements, for assuring that its users will embrace and use the platform. In
this paper, we describe the process, within the DEFeND EU Project, for eliciting
and analyzing requirements for such a complex platform, by involving stake-
holders from the banking, energy, health and public administration sectors, and
using advanced frameworks for privacy requirements and acceptance require-
ments. The paper also contributes by providing elicited privacy and acceptance
requirements concerning a holistic platform for supporting GDPR compliance.
1 Introduction
Since May 2018 the General Data Protection Regulation (GDPR) has become the
center of attention for practitioners, researchers, States, and citizens. The General Data
Protection Regulation enforces significant changes on the way that personal data is
being processed, the way that data protection authorities guide and audit data con-
trollers and on the individual rights of data subjects. Further, GDPR altered the terri-
torial scope of the European Data Protection framework, enforcing changes to service
providers who serve data subjects living in European member states.
For entities that process personal data (i.e., data controllers or data processors) the
enforcement of GDPR means the implementation of organizational and technical
changes, including the deployment of tools that allow demonstration of GDPR com-
pliance, the appointment of Data Protection Officers, the conduction of privacy impact
assessments, the training of staff, the implementation of data de-identification tech-
niques, and so on. According to the first official report on implementation of the GDPR,
provided by the European Data Protection Board (European Data Protection Board
2019), most organizations have increased their financial budget allocated to personal
data protection (30%–50%), increased the personnel allocated, while a total of 206.326
legal cases have been presented to the authorities from 31 member states (complaints,
data breaches, etc.). Reuters (2019) reports that organizations are still not ready in terms
of GDPR compliance, and many of them know very little about the Regulation and
whether or how it will affect them. A report by ISACA also presents a similar view
(approximately 65% of organizations reported not ready in terms of GDPR compliance
in May 2018) and elaborates on the technical, regulatory and legislative tools that should
be implemented to assist organizations in their compliance efforts (ISACA 2019).
We aim to address this research and industrial gap through the development of a
GDPR compliance platform that will deliver tools for organizations and interfaces for
data protection authorities and citizens to interact with the organizations that process
personal data. We do so, through the Data govErnance For supportiNg gDpr (DEFeND)
EU Project (Innovation Action) that is dedicated into delivering such a platform. Ten
organizations collaborate for the provision of the platform from Spain, UK, Italy,
Portugal, Bulgaria, Greece and France. The DEFeND platform will guide organizations
in fulfilling GDPR compliance through Privacy by Design and by Default tools, and in
supporting consent management, privacy analysis, security risk assessment, and data
breach management. The platform will also support citizens concerning personal data
management, awareness and breach notifications. Finally, it will support the interaction
of organizations with the respective data protection authorities.
In this paper, we present the software engineering methodology and results that
were followed to capture the needs of users and model the software requirements for a
GDPR Compliance Platform. Our software engineering approach spanned into multiple
aspects of user needs, including functional, security, privacy, legal and acceptance
requirements. We collected user needs focusing on four industrial sectors; namely
financial, health, public administration and energy management. In this paper however
we will emphasize on the financial sector and the respective lessons learned.
206 A. Tsohou et al.
The paper is structured into seven sections. Following this introduction, Sect. 2
provides a review of state of the art to reveal the industrial and academic needs
associated with a GDPR compliance platform. Section 3 presents our software engi-
neering approach and Sect. 4 presents our methodology to collect data for capturing
software requirements. Section 5 presents indicative software requirements that
resulted and Sect. 6 provides the knowledge that was learnt from this process and could
be informative for similar endeavors. Finally, Sect. 7 concludes the paper.
design phase of new systems and services, and also be enforced as a default setting
(Cavoukian 2011; Kurtz and Semmann 2018). While a number of methodologies for
privacy by design have been proposed during the last decade (e.g., (Kalloniatis et al.
2011; Deng et al. 2011; Faßbender et al. 2014; Notario et al. 2015), recent surveys
(e.g., (Kurtz and Semmann 2018)) exhibit a lack of technologies and/or tools to
implement the PbD principle in a holistic way. PbD principles have not yet gained
adoption in the engineering practice, mainly because a mismatch between the legal and
technological mindsets (Martin and Kung 2018) with the result being that engineers are
mostly relied on privacy policies for compliance.
The DEFeND project advances state-of-the-art by facilitating organisations to
implement a privacy management approach that takes into account the PbD principles,
enabling them to (re)design their processes with respect to their privacy requirements,
at an operational level.
Consent Management. Until recently, users were supposed to read privacy policies or
notices before giving their consent to the data controller for processing their data, but in
reality users never read them (McDonald and Cranor 2008). The cost of reading
privacy policies. ISJLP, 4, 543.), in which case consent becomes not informed (Tsohou
and Kosta 2017). Even if the users read the privacy policies, it is usually difficult to
follow the legal and technical terminology inside (often, lengthy) policy texts and
notices. With GDPR’s more strict requirements on: (a) the consent being specific;
(b) getting parents’ consent for processing children data; (c) respecting data subjects’
rights to revoke their consent, technologies and tools should provide users the possi-
bility to withdraw consent as easily as they gave it. State of the art technologies and/or
tools to implement the Lawfulness of Processing (ar. 6, GDPR) principle in a holistic
way do not exist or are still immature (Politou et al. 2018; Priyadharshini and Shyamala
2018).
The DEFeND project approaches consent management in a holistic way, delivering
a Privacy Data Consent (PDC) to users which will act as a contract among the data
controller and data subject, encapsulating all the necessary information regarding the
consent of the processing to their personal data. At operational level, the platform,
based on the PDC, will monitor and enforce data subject’s preferences, and will notify
users if any inconsistency will be identified.
Privacy Impact Assessment (PIA). The execution of PIAs (ar. 35, GDPR) should
ideally be supported by an information security risk management system to identify and
reduce the privacy risks of data subjects when their personal data are processed by data
controllers. Given that the guidelines of ISO/IEC 27005:2011do not include PIAs, and
that data protection standards such as BS 10012:2017, ISO/IEC 29151:2017, ISO/IEC
27018:2014, require PIA in addition to conducting information security risk assess-
ments, in 2017 ISO issued the ISO/IEC 29134:2017 standard with guidelines for PIA,
superseding ISO 22307:2008 (“Financial services - Privacy impact assessment”) and
related guidelines (WP29 Guidelines on Data Protection Impact Assessment 2017).
The DEFeND project will advance the current state of the art in Data Protection
Impact Assessment by providing an in-depth processing analysis based on a recognized
methodology and based on international standard. This analysis will be performed in an
208 A. Tsohou et al.
easy and user-friendly interface and it will not need a specific knowledge and expertise
in security and/or risk analysis to be performed.
The project will achieve its aim by introducing a new paradigm, which we call
Model-Driven Privacy Governance (MDPG). Such paradigm enables building (from an
abstract to a concrete level) and analysing privacy related models following a Privacy-
by-Design approach that spans over two levels, the Planning Level and the Operational
Privacy, Security, Legal and Technology Acceptance Requirements 209
Level, and across three management areas, i.e. Data Scope, Data Process and Data
Breach as shown in Fig. 1.
More specifically, at the planning level, the platform will support the development
of models of the organisational data that capture information required for GDPR
compliance such as identification of data and assets (art. 4), Organisational Info and
establishments (art. 4), Data Transparency, Lawfulness and Minimisation (art. 25),
personal data consent (art. 6, 7, 8, 13, 14) and data breach information (art. 34).
Concretely, the DEFeND platform will support the transformation of planning models
to operational models that are employed to perform analysis that supports Data min-
imisation, Data Protection, Impact Assessments (art. 35) and Privacy-by-Design and
Privacy-by-Default principles (art. 25). At the operational level, the project will bring
together security and privacy methodologies, encryption and anonymization tools and
policy enforcers.
These management areas could be seen as the main services that the platform will
provide to organizations and relevant stakeholders. Each one of these services assists
organisations to collect, analyse and operationalise different aspects and articles of the
GDPR and provide appropriate reporting capabilities.
To support those services, the platform consists of five (5) back-end components:
Data Assessment Component, Data Privacy Analysis Component, Privacy Specifica-
tion Component, Privacy Implementation and Monitoring Component, Data Breach
Component. Each component includes a number of modules aiming to deliver func-
tionalities (Fig. 2). The modules will be developed by enhancing software tools, ser-
vices and frameworks of the project partners. Moreover, the platform includes a
dashboard, which works as the main front-end between the platform and its users.
1
IEEE Guide for Software Requirements Specifications, IEEE Std 830-1984.
212 A. Tsohou et al.
Towards defining the requirements necessary to be used as basis for building the
DEFeND platform, we used a Human-Centered design (HCD), where incorporating the
user’s perspective into software development is considered of paramount importance in
order to achieve a functional and usable system (Maguire 2001). Based on widely
accepted methodologies that have been proposed in the area of user-adaptive systems
development, user data have been collected using questionnaire-based and interviews-
based approaches in order to assist the elicitation of requirements for the platform.
Further, focus groups were realized in order to validate the data collection instruments
and the elicited requirements. In particular, DEFeND partners identified the key
stakeholders, and for each user category, a questionnaire was prepared, aiming at
capturing the DEFeND user needs concerning various aspects; legal, functional,
security, privacy and acceptance aspects. In sequence, user needs were translated into
software requirements for all levels of the DEFeND platform, i.e., Data Scope Man-
agement (DSM), Data Process Management (DPM), and Data Breach management
(DBM). The overall approach is depicted in Fig. 3.
Fig. 3. Methodological approach for eliciting software requirements for DEFeND platform
Privacy, Security, Legal and Technology Acceptance Requirements 213
2
https://ec.europa.eu/eusurvey/runner/DEFeNDEndUser.
214 A. Tsohou et al.
participate in a focus group during a subsequent partners meeting. Indeed, DPOs and IT
managers from various organizations in the banking sector have participated either
physically or via conference call. The session included an initial introduction by one
partner of the project, a round table discussion with the participants and a final part with
question and responses. The objective of this session was to gather feedback from this
group of end-users on the questionnaire (with respect to structure, text of the questions,
format of the questions/answers, language used, etc.) which was shared with them in
advance. The result of this stage was a consolidated draft of the questionnaire for the
end users.
least one of the three criteria were (even partially) positive for a Chapter, we moved to
step 2, in which the specific Sections of that Chapter were examined in terms of
relevance applying the same three-step test. Where the responses to at least one of the
three criteria were (even partially) positive for a Section, we moved to step 3, in which
the individual Articles of that Section are examined in terms of relevance applying the
same three-step test.
Ultimately, the project has identified concrete, practical privacy and security legal
requirements that should ideally be met in relation to each theme of the DEFeND
platform for it to be able to support organisations in complying with the GDPR.
Considering both the 12 themes of the DEFeND platform and the GDPR requirements,
74 legal requirements have been compiled and distributed as depicted in Fig. 4.
Some indicative legal requirements in the areas ‘Developing a GDPR privacy plan’
and ‘Creating a third party management program’ are presented below in Table 2.
Privacy, Security, Legal and Technology Acceptance Requirements 219
In this section we present the requirements engineering challenges that the consortium
faced, the innovations that were applied, and the lessons learned from the process of
eliciting and consolidating requirements for a GDPR compliance platform.
years, the banks had implemented several tools and procedures to ensure compliance
with the GDPR.
In this context, the value of a unique platform like DEFeND could be in the
possibility of supporting a continuous GDPR Maturity Assessment, in order to identify
the most critical areas of compliance, plan the improvement actions and convey specific
reports to different actors, also considering the existing standard and the evolution of
best practices. However, our requirements elicitation process revealed that to leverage
those opportunities, it is important that the GDPR platform represents a sort of
orchestration engine, able to enforce a presidium on the different data protection
processes and able to seamlessly integrate with all the other systems and procedures
that the bank has already put in place. To this extent, the possibility to have a modular
solution is paramount.
7 Conclusions
In this paper, we have presented the process that was followed to elicit and analyze
requirements for a GPDR compliance platform. The complexity of the process was high
as it included the involvement of stakeholders from four different sectors, banking;
energy; health; and public administration. The process is composed of several require-
ments engineering activities that were adapted in order to specify the requirements for a
GDPR compliance platform including functional, non-functional, security, privacy,
legal and acceptance requirements. Finally, the challenges and lessons learned from this
process were summarized and presented.
Acknowledgments. This paper has received funding from the European Union’s Horizon 2020
research and innovation programme under grant agreement No. 787068.
References
Blank, S.G.: Four Steps to the Epiphany: Successful Strategies for Products that Win, Palo
(2007)
Bryman, A.: Social Research Methods, 3rd edn, p. 2008. Oxford University Press, Oxford (2008)
Cavoukian, A.: Privacy by Design. The 7 Foundational Principles, Implementation and Mapping
of Fair Information Practices (2011). https://iab.org/wp-content/IAB-uploads/2011/03/fred_
carter.pdf
Davis, A., Dieste, O., Hickey, A., Juristo, N., Moreno, A.M.: Effectiveness of requirements
elicitation techniques: empirical results derived from a systematic review. In: 14th IEEE
International Requirements Engineering Conference (RE 2006), pp. 179–188. IEEE (2006)
Deng, M., Wuyts, K., Scandariato, R., Preneel, B., Joosen, W.: A privacy threat analysis
framework: supporting the elicitation and fulfillment of privacy re-quirements. Re-quirements
Eng. 16(1), 3–32 (2011)
European Data Protection Board: First overview on the implementation of the GDPR and the
roles and means of the national supervisory authorities (2019). https://edpb.europa.eu/sites/
edpb/files/files/file1/19_2019_edpb_written_report_to_libe_en.pdf
222 A. Tsohou et al.
Faßbender, S., Heisel, M., Meis, R.: Problem-Based Security Requirements Elicitation and
Refinement with PresSuRE. In: Holzinger, A., Cardoso, J., Cordeiro, J., Libourel, T.,
Maciaszek, L.A., van Sinderen, M. (eds.) ICSOFT 2014. CCIS, vol. 555, pp. 311–330.
Springer, Cham (2015). https://doi.org/10.1007/978-3-319-25579-8_18
Gartner: Forecast Analysis: Information Security, Worldwide, 1Q17 Update, August 2017
(2017). https://www.gartner.com/en/documents/3889055
IAPP: 2018 Privacy Tech Vendor Report v.2.4e (2018). https://iapp.org/resources/article/2018-
privacy-tech-vendor-report/
ISACA: GDPR The End of the Beginning (2019). http://www.isaca.org/Knowledge-Center/
Documents/2018-GDPR-Readiness-Survey-Report.pdf
Juristo, N., Moreno, A.M., Dieste, O., Davis, A., Hickey, A.: Effectiveness of requirements
elicitation techniques: empirical results derived from a systematic review. In: 14th IEEE
International Requirements Engineering Conference (RE 2006) (RE), Minneapolis/St. Paul,
Minnesota, USA, 2006, pp. 179–188 (2006)
Kalloniatis, C., Belsis, P., Gritzalis, S.: A soft computing approach for privacy requirements
engineering: the PriS framework. Appl. Soft Comput. 11(7), 4341–4348 (2011)
Kurtz, C., Semmann, M.: Privacy by Design to Comply with GDPR: A Review on Third-Party
Data Processors (2018)
Maguire, M.: Methods to support human-centred design. Int. J. Hum.-Comput. Stud. 55(4), 587–
634 (2001)
Martin, Y.S., Kung, A.: Methods and tools for GDPR Compliance Through Privacy and Data
Protection Engineering. In: 2018 IEEE European Symposium on Security and Privacy
Workshops (EuroS&PW), pp. 108–111. IEEE (2018)
McDonald, A.M., Cranor, L.F.: The cost of reading privacy policies. ISJLP 4, 543 (2008)
Mouratidis, H., Giorgini, P.: Secure tropos: a security-oriented extension of the tropos
methodology. Int. J. Softw. Eng. Knowl. Eng. 17(02), 285–309 (2007)
Myers, M.D., Newman, M.: The qualitative interview in IS research: examining the craft. Inf.
Organ. 17(1), 2–26 (2007)
Notario, N., et al.: PRIPARE: integrating privacy best practices into a privacy engineering
methodology. In: 2015 IEEE Security and Privacy Workshop, pp. 151–158. IEEE, May 2015
Piras, L.: Agon: a gamification-based framework for acceptance requirements. Ph.D. dissertation,
University of Trento, 2018 (2018)
Piras, L., Dellagiacoma, D., Perini, A., Susi, A., Giorgini, P., Mylopoulos, J.: Design thinking
and acceptance requirements for designing gamified software. In: 13th IEEE International
Conference on Research Challenges in Information Science (RCIS), IEEE, Bruxelles (BE),
2019 (2019)
Piras, L., Giorgini, P., Mylopoulos, J.: Acceptance requirements and their gamification solutions.
In: 24th IEEE International Requirements Engineering Conference (RE), 2016. IEEE, Beijing
(2016)
Piras, L., Paja, E., Giorgini, P., Mylopoulos, J.: Goal models for acceptance requirements
analysis and gamification design. In: Mayr, H.C., Guizzardi, G., Ma, H., Pastor, O. (eds.) ER
2017. LNCS, vol. 10650, pp. 223–230. Springer, Cham (2017). https://doi.org/10.1007/978-
3-319-69904-2_18
Politou, E., Alepis, E., Patsakis, C.: Forgetting personal data and revoking consent under the
GDPR: challenges and proposed solutions. J. Cybersecurity 4(1), tyy001 (2018)
Priyadharshini, G., Shyamala, K.: Strategy and solution to comply with GDPR: guideline to
comply major articles and save penalty from non-compliance. In: 2018 2nd International
Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), pp. 190–
195. IEEE (2018)
Privacy, Security, Legal and Technology Acceptance Requirements 223
Pulse Survey: GDPR budgets top $10 million for 40% of surveyed companies, October 2017
(2017). https://www.pwc.com/us/en/services/consulting/library/general-data-protection-
regulation-gdpr-budgets.html
Reuters, T.: Study finds organizations are not ready for GDPR compliance issues (2019). https://
legal.thomsonreuters.com/en/insights/articles/study-finds-organizations-not-ready-gdpr-
compliance-issues. Accessed 5 Apr 2019
TrustArc: GDPR Compliance Status. A Comparison of US, UK and EU Companies, July 2018
(2018)
Tsohou, A., Kosta, E.: Enabling valid informed consent for location tracking through privacy
awareness of users: a process theory. Comput. Law Secur. Rev. 33(4), 434–457 (2017)
WP29 Guidelines on Data Protection Impact Assessment. Guidelines on Data Protection Impact
Assessment (DPIA) and determining whether processing is “likely to result in a high risk” for
the purposes of Regulation 2016/679 (2017). https://ec.europa.eu/newsroom/article29/item-
detail.cfm?item_id=611236
Uncertainty-Aware Authentication
Model for IoT
1 Introduction
Access control is defined as a mechanism to govern the access to the resources in a way
that such resources can be used only by authorized entities. Access control consists of
the three functions: Authentication, Authorization and Auditing [5]. In this research we
focus on uncertainty aspect of the authentication phase of the access control.
226 M. Heydari et al.
legitimate requests therefore it allows the subject to exceed their normal access rights.
This approach suffers from the same problem as BTG called the lack of scalability in
terms of policy rules. Risk-Aware access control was proposed to assess the risk of the
authentication request to determine whether the access to a resource should be granted
[17]. A number of studies suggested using RAAC. Bijon et al. [18] incorporated the
concept of risk awareness in RBAC. The role in the introduced RBAC model will be
activated only if the total risk of its active roles does not exceed a threshold. Further-
more, the threshold is determined dynamically in an adaptive manner. Baracaldo et al.
[19, 20] used trust and risk concepts on RBAC to deal with insiders. In this method, each
user is assigned a trust level and each access permission is associated with a risk value.
The risk of each role is calculated by the total risk of all direct and indirect permissions
enabled by its activation. In this method, a role is activated if the user meets the
minimum trust level required for that role. The value of the trust is determined based on
the amount of risk exposed by activating the role. Dimmock et al. [21] proposed a
method based on [22] to enhance the RBAC with trust and risk. To meet this goal trust
and cost evaluation measures are added to the OASIS policy language. This method
introduced a risk evaluation expression language to calculate the risk based on the given
values and make an access decision based on that calculation. Atlam et al. [23] devel-
oped an adaptive RAAC model for IoT. This model accepts real-time attributes
including user context, resource sensitivity, action severity and risk history as input and
estimates the overall risk value associated with each access request. The major concern
about this work is that the authors did not validate their proposed model. Nogoorani
et al. [24] proposed an access control framework for grid environment to address the
misuse of resource in virtual organizations. This method offers both risk and trust
analysis in authorization to assess the subject’s actions. The trust model uses feedback to
calculate user’s trust degree in a probabilistic approach. On the other hand, the risk
model is utility-based and uses the user’s trust degree to calculate the probability of
fulfilment of obligations. The proposed model was evaluated using simulation. The
results show that it is scalable in terms of the number of entities, the number of policy
rules and extensibility. Nurse et al. [25] argue that IoT-related characteristics such as
scalability, heterogeneity and dynamism make the current risk assessment approaches
inadequate due to the following reasons: (i) Doing periodic assessment in highly scal-
able and dynamic environments like IoT is impossible, (ii) It is difficult to gather
information about all entities participating in various IoT scenarios due to the scalability
of such an environment and (iii) Interoperability brings new challenges in terms of
dependencies among IoT entities and makes the risk analysis a challenging task.
3 Proposed Model
environment like IoT. As discussed in [28] and [7], uncertainty is caused by the lack of
information about the likelihood of an incident occurring. Therefore, we define
uncertainty in authentication as the incompleteness of information regarding the like-
lihood of whether the acceptance of an authentication request leads to an incident.
“Uncertainty” and “Risk” have similarity. In one hand, risk is measured by analysing
the likelihood of occurrence of a harmful circumstance or event besides assessing the
adverse impact that would arise if that circumstance or event occurs [29]. On the other
hand, in order to measure uncertainty, the likelihood of adverse event occurring is
considered so the “likelihood of event occurring” is common between these two
concepts.
Our uncertainty-aware data-driven model is based on the extension of ABAC. In
order to build our model, we consider three attributes from authentication request
namely as: time of the request, location in which the request comes from and the
credential provided by the user who sends the request. As shown in Fig. 1, we have
proposed an architecture based on the generic architecture of XACML [30]. According
to our proposed architecture in [28], (1) users send authentication requests to the Policy
Enforcement Point (PEP). (2) PEP as the interface between the system and the user is
responsible to forward the request to Policy Decision Point (PDP). PDP gathers policy
related to the specified resource from Policy Administration Point (PAP). (3) PDP asks
for policy from PAP. (4) PAP provides requested policy to PDP. (5) P also queries for
subject, object and environment attributes related to the request from Policy Infor-
mation Point (PIP). (6) PIP gathers requested attributes and makes it available to
PDP. (7) PDP forwards the gathered information to Indeterminacy Estimation Point
(IEP) and asks the uncertainty engine to calculate the uncertainty values associated to
the authentication request. (8) IEP queries the uncertainty engine to calculate the value
of uncertainty associated with the authentication request. (9) Uncertainty engine returns
the calculated the overall value for the uncertainty using our data-driven prediction
model which will be discussed in Sect. 4.2. (10) IEP sends the calculated value of
uncertainty to PDP. (11) PDP makes final access decision using related policy and the
value of indeterminacy which is provided by IEP. Then the decision will be forwarded
to PEP. (12) PEP fulfils the obligations based on the decision.
Fig. 1. Architecture of the proposed model based on the XACML architecture [28]
Uncertainty-Aware Authentication Model for IoT 229
4 Methodology
Fig. 3. Uncertainty areas for three defined PoIs, from PoI_1 to PoI_3 from left to right
Time. (i) The pattern for the time of authentication request depends on the business
model of the service in which the authentication process is embedded. For services are
deployed to be accessible 24 h a day, 7 days a week like Email service generally no
restriction is defined for the sake of access in terms of time. In such a case, time of the
authentication requests follows uniform distribution. In order to make the scenario for
dataset more challenging we consider a service which is mostly demanded during a
specific time period like work hours (e.g. 9 AM to 5 PM) then we should take those
time preferences into consideration and find the corresponding PDFs. (ii) According to
our assumption for this case study, the majority of users send authentication requests
during work hours (9 AM to 5 PM) and the number of requests before 9 AM and after
5 PM plummeted gradually. We also suppose that the number of requests between 12–
13 decreases due to the rest/lunch time. (iii) We have broken the time of the authen-
tication requests into 11 timeslots due to the above considerations. Based on the
likelihood of making authentication request, we assign a weight in terms of probability
for each timeslot. For example, the probability of receiving an authentication request
between 9 AM to 12 AM and 13 PM to 17 PM is higher than the other timeslots. We
have also defined an uncertainty value for each timeslot. In doing so, we determine
values in a way that authentication requests made during work hours supposed to be
less risky than any request which is made out of work hours therefore the value of
uncertainty is lower during work hours and such a value for the requests out of the
work hours increases gradually. We have also assumed the least value of uncertainty
for all authentication requests during work hours because of the potential threat of
Uncertainty-Aware Authentication Model for IoT 231
insiders. (iv) In order to generate the values for the time of authentication request, we
have applied two PDFs. Multinomial distribution was used to randomly generate the
samples for timeslots in which the request comes from. In doing so we used the
assigned weights. Then a uniform distribution was similarly applied to randomly
determine the time of the request within nominated timeslot. Finally, uncertainty values
for the generated request times were assigned based on the records of Table 1.
Location. As thoroughly discussed in [28], one of the attributes of our ABAC based
model is the location. Our approach can consider the location in which the authenti-
cation request comes from in order to make accurate access decision. It can also handle
the uncertainty of mobile users as well as fixed ones in authentication. The need for
considering mobility in authentication is on the rise the number of security and privacy
incidents caused by them is rapidly increasing [33]. We have taken the following points
into considerations to generate uncertainty values for the location of the requests: (i)
We have applied the mixture of Gaussian PDF to generate data for any specific location
of the mobile user in a two-dimensional grid, (X: longitude and Y: latitude) based on a
number of studies which suggested normally distributed locations of mobile users in
communication [34–37]. (ii) In order to make our case study more challenging, we
have defined a scenario consisting of three Point of Interests (PoI). The number of PoIs
may vary from one case study to another. According to above-mentioned assumptions
our PDF consists of three Gaussian factors (because of our three Point of Interests) in
which each of them has a weight and each PDF belongs to one PoI respectively:
We expect that most of the authentication requests to be sent from or around the first
PoI (which is generated using G1) such that the magnitude of a coefficient was chosen
in a way that reflects this fact. Next, the second PoI generates the second highest
number of requests (using G2) and the magnitude of b was chosen in a way that it is
lower than a whist the third PoI should generate the smallest number of authentication
232 M. Heydari et al.
request associated with location (using G3) and the magnitude of c was determined as
the lowest value to represent this fact that the least of the authentication requests to be
sent from or around the third PoI. So that:
a[b[c ð2Þ
(iii) We have generated our samples in terms of location (mobile and fixed) along
with a map of area 2000 m * 2000 m, which contains three PoIs namely PoI_1, PoI_2
and PoI_3. Figure 3 shows the corresponding map in detains. The assigned values as
gaussian parameters l and r for our three gaussian factors are shown in Table 2. These
values were used to generate random values in both dimensions X and Y. (iv) We have
defined five different Uncertainty Areas (UAs) for each PoI and determined uncertainty
values for each of these areas respectively (Fig. 3). In order to define UAs for each PoI,
five circles were drawn with the PoI point as the center and with (2n + 1)*r as radius
(n = 0, 1, 2, 3… and r = 200 m). The number of circles and the length of the radius
may vary from one case study and thus is considered a system parameter. (v) The
process of generating samples for the authentication request in terms of location are as
follows: First, in order to randomly choose a PoI from three PoIs a multinomial PDF
was applied using nominated weights (a, b, c) as probabilities. Then, corresponding
Gaussian PDF was applied to generate the X and Y points of the location. Third,
according to the location of the generated point like A on the map, an Uncertainty
Value (UV) was calculated using the following formula:
The value assigned by each given PoI in the above formula depends on the
uncertainty area (UA) in which the point has fallen.
Credential. The most common form of credential is username and password. We have
considered this information as the credential for this research. We have taken the
following points into considerations to generate uncertainty values for credentials: (i)
Usernames and passwords entered by users makes three possibilities: (i) both username
and password provided by the user are correct (ii) only the username is correct and
(iii) both username and password is incorrect. Data for the three possible states was
generated form a multinomial PDF as described below. (ii) Generally, most users enter
username and password correctly. Otherwise, most users enter the username correctly
but enter the password incorrectly. These were considered when assigning probability
values and associated uncertainty values listed in the Table 3 for these three states.
Uncertainty-Aware Authentication Model for IoT 233
Access Decision. After generating the uncertainty values for each attribute in the
matrix shown in Fig. 2 the final uncertainty value is calculated for each request in order
to make an authentication decision. The final value for each authentication request was
calculated by averaging the uncertainty values of time, location and credential. Gen-
erally, credential is the most important authentication attribute in comparison with time
and location. We have added weights to the generated uncertainty values to show the
priority and importance of the attributes. The magnitude of these weights may vary
based on the research priorities. Therefore, we have calculated the weighted arithmetic
mean by averaging of weighted uncertainty values (weight values: Time = 2, Loca-
tion = 3 and Credential = 5). Finally, for labelling the dataset we have used the final
uncertainty value for each request as the probability for binomial distribution to
determine the class of the result: {0: Deny and 1: Access}.
Table 3. Assigned values for credential associated PDF and corresponding uncertainty values
Username & password are Username is correct but Username & password are
correct password incorrect
Probability: 0.85 Probability: 0.10 Probability: 0.05
Uncertainty: 0.05 Uncertainty: 0.70 Uncertainty: 0.95
the class label based on the majority of the labels predicted by each individual classifier
[41]. We have applied both soft and hart vote modes in this research.
We have discussed other classifiers including decision tree, SVM, logistic regression
and Naïve Bayes in [28].
Validation. In order to validate the data model, Cross-Validation process has been
used by each of the applied classifier. Cross validation is the widely used method to
evaluate the generalizability of proposed models [42]. In doing so, 10% of dataset was
assigned to the test split (10-fold cross validation). In order to increase the chance of
finding the best fit model and improve the generalizability of the generated model we
also used the shuffling feature.
100
90
80
70
60
50
40
30
20
10
0
Logicc
Decision tree SVM Naïve Bayes AdaBoost Gradient Boost Vong
Regression
Accuracy 78.2 83.08 81.02 82.16 86.54 83.08 84.94
Precision 87 83 86 88 87 83 85
Recall 85 100 94 92 97 100 100
F1 86 91 90 90 92 91 92
In this paper we have reviewed the state-of-the-art access control models suggested for
IoT. According to our findings, applying pre-defined access policies included in tradi-
tional and emerging access control models do not result in accurate access decision for
indeterminate access scenario in IoT. We have also stressed on uncertainty as a
neglected challenge in the authentication for IoT environment which is exaggerated by
IoT characteristics like scalability, dynamism and heterogeneity. We have also reviewed
the resilient access control proposed in the literature. We found that proposed RAAC
methods in literature are inadequate for the IoT. We have also proposed an uncertainty-
aware authentication model based on ABAC. Our model was built based on the dataset
that we have synthesized based on the state-of-the-art researches. We have built
uncertainty-aware authentication models using Boosting and voting classifiers. The
results indicated that the model created using AdaBoost shows better performance in
terms of higher accuracy and precision together than the performance of the other
models. The future step of this work is to focus on ambiguity in authentication.
References
1. Zhou, W., Jia, Y., Peng, A., Zhang, Y., Liu, P.: The effect of IoT new features on security
and privacy: new threats, existing solutions, and challenges yet to be solved. IEEE Internet
Things J. 6(2), 1606–1616 (2018)
236 M. Heydari et al.
2. Bertino, E., Choo, K.-K.R., Georgakopolous, D., Nepal, S.: Internet of Things (IoT): smart
and secure service delivery. ACM Trans. Internet Technol. 16(4), 22–29 (2016)
3. Restuccia, F., D’Oro, S., Melodia, T.: Securing the Internet of Things in the age of machine
learning and software-defined networking. IEEE Internet Things 5(6), 4829–4842 (2018)
4. Zhang, C., Green, R.: Communication security in Internet of Thing: preventive measure and
avoid DDoS attack over IoT network. In: IEEE Symposium on Communications &
Networking (2015)
5. Stallings, W.: Access control. In: Computer Security, Principles and Practice. Pearson (2017)
6. Ouaddah, A., Mousannif, H., Abou, A., Abdellah, E.: Access control in the Internet of
Things: big challenges and new opportunities. Comput. Netw. 112, 237–262 (2017)
7. Heydari, M., Mylonas, A., Katos, V., Gritzalis, D.: Towards indeterminacy-tolerant access
control in IoT. In: Dehghantanha, A., Choo, K.-K.R. (eds.) Handbook of Big Data and IoT
Security, pp. 53–71. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-10543-3_4
8. Rizvi, S.Z.R., Fong, P.W.L.: Interoperability of relationship - and role-based access model.
In: Proceedings of the Sixth ACM Conference on Data and Application Security and Privacy
(2016)
9. Kaiwen, S., Lihua, Y.: Attribute-role-based hybrid access control in the Internet of Things.
In: Han, W., Huang, Z., Hu, C., Zhang, H., Guo, L. (eds.) APWeb 2014. LNCS, vol. 8710,
pp. 333–343. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-11119-3_31
10. Biswas, P., Sandhu, R., Krishnan, R.: Attribute transformation for attribute-based access
control. In: Proceedings of the 2nd ACM International Workshop on Attribute-Based Access
Control (2017)
11. Savinov, S.: A dynamic risk-based access control approach: model and implementation.
Ph.D. thesis, University of Waterloo (2017)
12. Salim, F.: Approaches to access control under uncertainty. Ph.D. thesis, Queensland
University of Technology (2012)
13. Ferreira, A., Cruz-Correia, R., Antunes, L.: How to break access control in a controlled
manner. In: 19th IEEE International Symposium on Computer-Based Medical Systems
(2006)
14. Maw, H.A., Xiao, H., Christianson, B., Malcolm, J.A.: BTG-AC: break-the-glass access
control model for medical data in wireless sensor networks. IEEE J. Biomed. Health Inform.
20(3), 763–774 (2016)
15. Schefer-Wenzl, S., Strembeck, M.: Generic support for RBAC break-glass policies in
process-aware information systems. In: 28th Annual ACM Symposium on Applied
Computing (2013)
16. Povey, D.: Optimistic security: a new access control paradigm. In: ACM Workshop on New
Security Paradigms (1999)
17. Molloy, I., Dickens, L., Morisset, C., Cheng, P.C., Lobo, J., Russo, A.: Risk-based security
decisions under uncertainty. In: Proceedings of the Second ACM Conference on Data and
Application Security and Privacy (2012)
18. Bijon, K.Z., Krishnan, R., Sandhu, R.: Risk-aware RBAC sessions. In: Venkatakrishnan, V.,
Goswami, D. (eds.) ICISS 2012. LNCS, vol. 7671, pp. 59–74. Springer, Heidelberg (2012).
https://doi.org/10.1007/978-3-642-35130-3_5
19. Baracaldo, N., Joshi, J.: A trust-and-risk aware RBAC framework: tackling insider threat. In:
ACM Proceedings of the 17th Symposium on Access Control (2012)
20. Baracaldo, N., Joshi, J.: An adaptive risk management and access control framework to
mitigate insider threats. J. Comput. Secur. 39, 237–254 (2013)
21. Dimmock, N., Belokosztolszki, A., Eyers, D., Bacon, J., Moody, K.: Using trust and risk in
role-based access control policies. In: ACM Symposium on Access Control Models and
Technologies (SACMAT) (2014)
Uncertainty-Aware Authentication Model for IoT 237
22. Bacon, J., Moody, K., Yao, W.: A model of OASIS role-based access control and its support
for active security. ACM Trans. Inf. Syst. Secur. 5(4), 492–540 (2002)
23. Atlam, H.F., Alenezi, A., Walters, R.J., Wills, G.B., Daniel, J.: Developing an adaptive risk-
based access control model for the Internet of Things. In: IEEE International Conference on
Internet of Things (2017)
24. Nogoorani, S.D., Jalili, R.: TIRIAC: a trust-driven risk-aware access control framework for
grid environments. Future Gener. Comput. Syst. 55, 238–254 (2016)
25. Nurse, J.R.C., Creese, S., De Roure, D.: Security risk assessment in Internet of Things
systems. IT Prof. 19(5), 20–26 (2017)
26. Ghorbani, H.R., Ahmadzadegan, M.H.: Security challenges in Internet of Things: survey. In:
IEEE Conference on Wireless Sensors (ICWiSe) (2017)
27. Frustaci, M., Pace, P., Aloi, G., Fortino, G.: Evaluating critical security issues of the IoT
world: present and future challenges. IEEE Internet Things J. 5(4), 2327–4662 (2017)
28. Heydari, M., Mylonas, A., Katos, V., Balaguer-Ballester, E., Tafreshi, V.H.F., Benkhelifa,
E.: Uncertainty-aware authentication model for fog computing in IoT. In: The Fourth IEEE
International Conference on Fog and Mobile Edge Computing, Rome, Italy (2019)
29. Gallagher, P.D.: NISP SP800-30 guide for conducting risk assesment. In: NIST (2012)
30. Moses, T.: Extensible access control markup language (XACML). In: OASIS (2013)
31. User-Computer Authentication Associations in Time. Los Alamos National Laboratory.
https://csr.lanl.gov/data/auth/. Accessed 13 Feb 2019
32. Lohweg, V.: Banknote authentication data set. Center for machine learning and intelligent
systems, University of California. https://archive.ics.uci.edu/ml/datasets/banknote
+authentication. Accessed 13 Feb 2019
33. Premarathne, U.S., Khalil, I., Atiquzzaman, M.: Location-dependent disclosure risk based
decision support framework for persistent authentication in pervasive computing applica-
tions. Comput. Netw. 88, 161–177 (2015)
34. Marcus, P., Linnhoff-Popien, C.: Efficient evaluation of location predicates for access control
systems. In: IEEE Sixth UKSim/AMSS European Symposium on Computer Modeling and
Simulation (2012)
35. Chandrasekaran, G., Wang, N., Hassanpour, M., Xu, M., Tafazolli, R.: Mobility as a service
(MaaS): a D2D-based information centric network architecture for edge-controlled content
distribution. IEEE Access 6, 2110–2129 (2018)
36. Ekman, F., Keranen, A., Karvo, J., Ott, J.: Working day movement model. In: ACM
Proceedings of the 1st ACM SIGMOBILE Workshop on Mobility Models (2008)
37. Keränen, A., Ott, J., Kärkkäinen, T.: The ONE simulator for DTN protocol evaluation. In:
ACM Proceedings of the 2nd International Conference on Simulation Tools and Techniques
(2009)
38. Aggarwal, C.C.: An introduction to data mining. Data Mining, pp. 1–26. Springer, Cham
(2015). https://doi.org/10.1007/978-3-319-14142-8_1
39. Hu, W., Hu, W., Maybank, S.: AdaBoost-based algorithm for network intrusion detection.
IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 28(2), 577–583 (2008)
40. Punmiya, R., Choe, S.: Energy theft detection using gradient boosting theft detector with
feature engineering-based preprocessing. IEEE Trans. Smart Grid 10(2), 2326–2329 (2019)
41. Raschka, S., Mirjalili, V.: Combining different models for ensemble learning. In: Python
Machine Learning. Packt Publishing, pp. 219–233 (2017)
42. Ian, H., Frank, E., Hall, M.A., Pal, C.J.: Data mining: practical machine learning tools and
techniques. Morgan Kaufmann Series in Data Management Systems (2016)
43. Fawcett, T.: An introduction to ROC analysis. Pattern Recogn. Lett. 27(8), 861–874 (2006)
44. Davis, J., Goadrich, M.: The relationship between precision-recall and ROC curves. In:
Proceedings of the 23rd International Conference on Machine Learning, ICML 2006 (2006)
From ISO/IEC 27002:2013 Information
Security Controls to Personal Data
Protection Controls: Guidelines
for GDPR Compliance
1 Introduction
online, exchanged on a broad scale, constitute one of the driving forces of modern
enterprises [20]. Online activities of individuals produce data that are of value
for enterprises which base their business models on such data to provide person-
alised services, exploiting targeting marketing. For this reason, the protection
of personal data has seen a major upheaval during the last decades, concen-
trating the attention of politicians, developers, public and private organisations,
legislators, authorities, as well as the general public. Personal data protection
legislation has attempted to pose restrictions to the uncontrollable use of such
data, by governments, enterprises, etc. However, different national laws had sub-
stantially different characteristics [18], allowing organisations to take advantage
of such blurred territories of legislation and proceed with the exploitation and
processing of such data. Before May 2018, European Union (EU) Member States
applied national privacy laws, following the EU Directive 95/46 [1] each Mem-
ber State of the EU had its national privacy law that the organisations had to
comply with. With the General Data Protection Regulation (hereafter, GDPR
or Regulation) [2], EU adopted a unified privacy law, aiming to protect and reg-
ulate the massive usage of personal data. The GDPR aims at the regulation and
the management of personal data, defining strict fines to the data controllers
that do not comply.
Compliance with the GDPR comprises a challenging project for organisations
for a series of reasons; the complexity of business activities and the duplication
of data (in different information flows or even entire departments within an
organisation) are the most important ones. In addition, even if organisations
need to comply with the GDPR, they lack guidelines that could help them into
complying with these requirements. There are already products being developed
that can facilitate the compliance with the GDPR, however, none of the current
technical solutions is able to capture the current personal data protection sta-
tus of an organisation, identify the gaps, assess the criticality of the processing
activities and the personal data they use, provide concrete solutions tailored to
each organisation to finally fortify its processes and guarantee the protection of
individuals’ personal data [12].
We argue that the ISO 27k standard series can form a useful baseline for
businesses to build their “towards-compliance” strategy upon, dealing with top-
ics such as risk definition and assessment, continuous evaluation and appropriate
documentation. ISO/IEC 27001:2013 [16] (hereafter, ISO 27001) and GDPR aim
both to strengthen data security and mitigate the risk of data breaches, and they
both require organisations to ensure the confidentiality, integrity and availabil-
ity of data. Recital 83 of the GDPR states In order to maintain security and to
prevent processing in infringement of this Regulation, the controller or processor
should evaluate the risks inherent in the processing and implement measures to
mitigate those risks, such as encryption. The information-risk-driven approach,
which is also described in the GDPR, consists a fundamental perspective for ISO
27001. ISO 27001 provides detailed best practices while Article 24 of the GDPR
specifies that adherence to codes of conduct and approved certifications can be
used as an element for demonstrating compliance. There are several similari-
240 V. Diamantopoulou et al.
much after the GDPR came into force. A recent Thomson Reuters article [21]
highlights that evidence show that organisations are still not fully aware of the
GDPR’s potential impact and are not ready for the GDPR compliance issues.
In a survey [13] among privacy professionals which was published in 2019 by
the International Association of Privacy Professionals (IAPP), it appears that
less than half of respondents said they are fully compliant with the GDPR.
Interestingly, nearly 20% of the privacy professionals who participated argues
that full GDPR compliance is truly impossible.
Among the reported challenges [6], it seems that organisations are battling
on the way to satisfy the data subjects’ right to erasure (“right to be forgotten”)
(GDPR, Article 17). This was cited by 53% of the survey respondents as the
biggest challenge on achieving compliance with the GDPR. Data protection-by-
design and -by-default (GDPR, Article 25) follows with 42% and “records of
processing activities” (GDPR, Article 30) with 39%. IAPP [14] has published a
Data Protection Officer’s (DPO) experience on the GDPR a year after it entered
into force also highlights that managing and addressing data subjects’ requests
was the biggest challenge.
is the level of detail they present the controls, as ISO 27001 dictates a short
description of each control, while ISO 27002 explains each control in detail,
providing good practices for their successful implementation.
The application of ISO 27001 supports organisations in creating better busi-
ness efficiency, safeguards valuable assets such as personal data or hardware,
protects staff and organisations’ reputation, and simultaneously facilitates the
attainment of compliance objectives. Organisations have always sought some
short of certification for one or more of their business activities. Regarding
information security, 39,501 ISO 27001 certificates were issued to organisations,
worldwide, in 20171 . Given its wide recognition and acceptance, and in the
absence of a GDPR compliance certification, ISO 27001 makes a good candidate
to be considered as the baseline upon which organisations can work on in order to
reach GDPR compliance. Furthermore, there is a lot of common ground between
ISO 27001 and GDPR, which additionally strengthens the previous statement.
Juxtaposing ISO 27001 and the GDPR we have identified that they are based
on common ground. Despite the fact that they have different perspectives, both
ISO 27001 and the GDPR focus on the minimisation of risk that can be realised
when a data breach occurs. ISO 27001 focuses on reducing risks to informa-
tion security by compelling organisations to produce ISMS that are continu-
ously maintained and improved. GDPR aims at the preservation of privacy of
individuals, providing them with rights against organisations that process their
personal data. GDPR also promotes accountability, by placing clear data pro-
tection responsibilities to the corresponding organisations processing such data.
The aforementioned accountability lies on the implementation of appropriate
technical and organisational measures to ensure a level of security appropriate
to the risk (GDPR, Article 32).
Both GDPR and ISO 27001 request that organisations focus on the empower-
ment of knowledge which is communicated to the leadership, and develop aware-
ness within the whole organisation regarding the protection of data, exploiting
security measures [12]. The GDPR provisions numerous personal data protec-
tion settings and controls, many of which are also recommended in ISO/IEC
27001:2013, ISO/IEC 27002:2013, and other “ISO27k” standards. Organisations
that currently have an ISMS are likely to satisfy many of the GDPR require-
ments already, needing a few adjustments to be made. In this section we anal-
yse the ISMS framework of ISO 27001 and identify synergies with the GDPR
compliance efforts. In the following subsections the fourteen control modules of
Annex A’ of ISO 27001 are presented, focusing on the level of the proposed
controls that can be implemented. At this level, we extend the information secu-
rity controls to personal data protection controls, analysing and describing the
1
https://www.iso.org/the-iso-survey.html.
From ISO/IEC 27002:2013 to Personal Data Protection Controls 243
This control module includes two categories, (i) the internal organisation, and (ii)
the mobile devices and teleworking. This control module aims at the establish-
ment of a framework for the administration on the implementation and operation
of security within the organisation, and the protection of security related with
the information accessed, processed and/or stored at teleworking sites, and the
use of portable devices.
The category Internal Organisation consists of six controls. The first refers
to information security roles and responsibilities, mentioning that all informa-
tion security responsibilities shall be defined and allocated. The second refers to
segregation of duties, where conflicting duties and areas of responsibility should
be segregated to reduce opportunities for unauthorised or unintentional modi-
fication or misuse of the organisation’s assets. The third refers to contact with
authorities, where appropriate contacts with relevant authorities should be main-
tained. The fourth refers to contact with special interest groups, where appropri-
ate contacts with special interest groups or other specialist security forums and
professional associations should be maintained. Finally, the fifth control refers to
information security in project management, where information security should
be addressed in project management, regardless of the type of the project.
Actions Towards GDPR Compliance: The organisation is responsible for
implementing an organisational framework according to which there are roles
with responsibilities for the protection of personal data. The framework should
include the role of the Data Protection Officer; in cases required. The role of
the Data Protection Officer should be designated by the senior management,
assigning this responsibility to a competent person reporting directly to the
senior management without receiving any instructions on how to perform his/her
tasks. Senior management needs to ensure that the Data Protection Officer is not
dismissed or penalised for performing his/her tasks. The organisational structure
of the organisation reflects the distinct role of the Data Protection Officer. A
Data Protection Officer should be appointed, if (i) the processing is carried
out by a public authority or body, except for courts acting in their judicial
capacity, (ii) the data controller’s main activities require regular and systematic
monitoring of the data subjects on a large scale, and (iii) the data controller’s
From ISO/IEC 27002:2013 to Personal Data Protection Controls 245
main activities are large scale processing of specific categories of personal data
(GDPR, Article 37). The organisation should appoint necessary responsibilities
to the Data Protection Officer, as described in GDPR (Article 39).
Regarding contact with authorities, the data controllers need to cooperate
with the supervisory authorities when a data breach occurs (GDPR, Article
33), informing them without undue delay, when the personal data breach affects
the rights and freedoms of the corresponding natural persons. When the data
controller realises that the data breach may pose a high risk to their rights and
freedoms, they should also inform the data subjects for the violation of their data
(GDPR, Article 34), choosing the most appropriate means for communication
(e.g., email, newsletter, press release, etc.) according the number of the affected
natural persons and the severity of the data breach.
Regarding contact with special interest groups, in order for a data controller
to be able to guarantee the protection of the personal data they process, they
need to conduct a Data Protection Impact Assessment (DPIA) when particular
types of processing is likely to result in a high risk to the rights and freedoms
of natural persons (GDPR, Article 35). The data controller carries out DPIA
in case of (i) systematic and extensive evaluation of personal aspects relating
to natural persons which is based on automated processing, including profiling,
(ii) processing on a large scale of special categories of data, (iii) systematic
monitoring of a publicly accessible area on a large scale.
Finally, regarding information security in project management, organisations
should establish a code of conduct (GDPR, Article 40). Codes of conduct can
contribute to the proper application of the GDPR, taking account of the specific
features of the various processing sectors and the specific needs of micro, small
and medium-sized enterprises. They are related to associations and other bodies
that represent data controllers or data processors. To this direction, data con-
trollers and data processors are encouraged by the GDPR to be certified with
a certification mechanism (GDPR, Article 42). Such mechanisms may be estab-
lished for the purpose of demonstrating the existence of appropriate safeguards
provided by controllers or processors. They enable the mandatory monitoring of
compliance either by the supervisory authority, or by an accredited organisation
(demonstrating independence and expertise). Codes of conduct can be drawn up
by organisations that represent data controllers or data processors and approved
either by the supervisory authority of a member state or by the European Data
Protection Board.
The human resources security module consists of three sub categories, (i) infor-
mation security prior to employment; (ii) during employment, and (iii) termi-
nation and change of employment. Information security prior to employment
contains two controls, i.e. screening and terms and conditions, during employ-
ment the controls refer to the management responsibilities, information security
246 V. Diamantopoulou et al.
awareness, education and training and disciplinary process, while in the termi-
nation and change of employment the organisation should take care of the infor-
mation security responsibilities that remain valid after termination or change of
employment.
Actions Towards GDPR Compliance: Further actions should be taken
regarding the protection of personal data that an organisation processes by
its employees. The organisation should take appropriate measures and controls
related with the management of their employees, so that they protect the per-
sonal information (of the personal data of natural persons that the organisation
keeps, e.g., personal data of customers, suppliers) that they process within the
scope of their occupation. Specifically, before the employment of their employees,
an organisation should take appropriate measures to ensure that the employees
are fit to handle personal data, e.g., by a screening process and by informing
them about possible legal consequences during the exercising of the work activi-
ties (regarding personal data misuse, etc.). During employment, the organisation
should review the already existing contracts of their employees who have access
to personal data, and make sure that they include specific clauses for confi-
dentiality, with legal bindings. Finally, after the employment, the organisation
should remove access rights to personal data the corresponding employees had
access to.
to protect users’ privacy. This means that when designing the access control safe-
guards the organisation should not take into account only the security require-
ments (e.g., identification, accountability), but also take into account privacy
requirements and principles (e.g., data minimisation).
requesting access to personal data (GDPR, Article 26). This can be extended
to the international transfers, where the organisation, before transferring the
requested personal data, should have received appropriate safeguards ensuring
an adequate level of protection of the corresponding country, the territory, or
one or more specified sectors within that third country.
Additionally, appropriate roles should be given to the corresponding employ-
ees who have access to personal data, accompanied with specific responsibilities.
This functionality promotes accountability and transparency, while it consists a
basis for the accurate response of the organisation, either to any request received
by a data subject regarding the processing of their data (GDPR, Articles 13–
22), or to the supervisory authority, when a data breach occurs (GDPR, Articles
31, 33). In this way, the organisation is able to locate and retrieve securely the
personal data it keeps.
Also, the organisation should be able to identify and assess the special cate-
gories of personal data they process. Information risks could be avoided, where
feasible, by assessing the usefulness of the personal and special categories of per-
sonal data they keep. Towards risk minimisation, the aggregation of such data
is also accepted (GDPR, Articles 9, 11).
In addition, in order to satisfy the right of data subjects to know the outcome
of requests related with the correction, completion, erasure, restriction of their
personal data (GDPR, Article 19), the organisation should inform the requestor
on the above, also providing that this process/application form is easy for insiders
and outsiders of the organisation to follow.
The module Supplier Relationships aims to manage the relationship of the organ-
isation with its suppliers, or any other third party that has access to the organ-
isation’s assets, and to set up and agree a level of information security and
service delivery. It consists of two controls, (i) information security in supplier
relationships, and (ii) supplier service delivery management.
Actions Towards GDPR Compliance: This control module sets the basis
for the establishment of a security framework among an organisation and the
external parties it collaborates with, ensuring the protection of the transferred
information. GDPR sets specific requirements regarding the management of the
relationship of the data controller with its processors. If an organisation uses
one or more third parties to process personal information (“processors”), it must
ensure they too are compliant with the GDPR (GDPR, Articles 27, 28). Towards
this direction, data controllers should conduct continuous evaluation of their
processors and suppliers, and use approved certification mechanisms in order to
demonstrate that they ensure an adequate level of protection with respect to
data protection-by-design and -by-default principles.
Moreover, organisations need to ensure the privacy and other information
security aspects of their business partners. This might contain aspects such as
jointly investigating and resolving privacy incidents, breaches or access requests,
to name a few. These requirements are applied to any relationship the organisa-
tion has with external parties, such as ISPs and CSPs, and any other third party
that the organisation has exchanged (personal) data with, for example external
payroll or marketing companies.
Finally, when data is transferred outside EU, involved organisations should
ensure the level of protection of the involved natural persons. Consequently,
organisations located outside Europe that interact with European organisations
must formally nominate privacy representatives inside Europe if they meet cer-
tain conditions (GDPR, Article 27).
From ISO/IEC 27002:2013 to Personal Data Protection Controls 251
1. The data subject has provided their consent regarding the processing of their
personal data.
2. Performance of a contract to which the data subject takes part.
3. Processing is necessary for compliance with a legal obligation of the data
controller.
4. Processing is necessary for the protection of vital interests of natural per-sons.
5. Processing is necessary for the performance of a task related with public
interest
6. Processing is necessary for the purposes of the legitimate interests pursued
by the controller or by a third party.
The following three modules are also part of the Annex A’ of the examined ISO,
however they have no direct application to the GDPR, but they can help an
From ISO/IEC 27002:2013 to Personal Data Protection Controls 253
organisation develop a culture that will assist towards reaching GDPR compli-
ance. Moreover, these modules are included in our study for the sake of com-
pleteness.
Enhancing Physical and Environmental Security for GDPR Compli-
ance: This control module concerns two controls: secure areas and equipment.
The identification of secure areas can prevent unauthorised physical access, dam-
age and interference to the organisation’s information and information processing
facilities, while the safeguarding of the equipment of the organisation prevents
loss, damage, theft or compromise of assets and interruption of organisation’s
operation.
Actions Towards GDPR Compliance: This section applies to the general
requirement of the GDPR to the organisations for implementing appropriate
technical and organisational measures to ensure the level of security appropriate
to the risk (GDPR, Articles 24, 25, 28, 32).
Enhancing Operations Security for GDPR Compliance: This control
module contains seven controls: (i) operational procedures and responsibilities,
(ii) protection from malware, (iii) back up, (iv) logging and monitoring, (v)
control of operational software, (vi) technical vulnerability management, and
(vii) information systems audit considerations. The objective of this section is
to ensure correct and secure operations of information processing facilities, pro-
tection against malware and data loss, to record events and generate evidence,
to ensure the integrity of operational systems, to prevent exploitation of techni-
cal vulnerabilities and to minimise the impact of audit activities on operational
systems.
Actions Towards GDPR Compliance: Similarly to the previous section of
“physical and environmental security”, an organisation is able to demonstrate
that they have implemented they appropriate technical and organisational mea-
sures to safeguard the personal data they keep. Additionally, the organisation
should implement procedures related with the management of the satisfaction
of the data subjects’ rights (GDPR, Articles 12–22) and for the process of the
provision of consent of the data subjects (GDPR, Articles 7).
Extending Business Continuity Management to Support GDPR Com-
pliance: This control module contains two controls: (i) information security con-
tinuity, and (ii) redundancies. The objective is the establishment of a business
continuity and disaster recovery plan. The continuity of operations is indented
to restore the operation of the organisation’s systems within a reasonable time.
In addition, staff training is required in the continuity plan, while its efficiency
must be tested and managed properly.
Actions Towards GDPR Compliance: As a general direction for the satis-
faction of the GDPR, an organisation should implement appropriate technical
and organisational measures to ensure the level of security appropriate to risk
(GDPR, Articles 24, 25, 28, 32).
254 V. Diamantopoulou et al.
6 Conclusions
The new regulation for the protection of the personal data, GDPR, provisions
numerous settings and controls focused on the management and the protection
of such data. Many of these controls are also provisioned in ISO/IEC 27001:2013,
ISO/IEC 27002:2013, and other “ISO27k” standards. Thus, organisations that
currently have developed an ISMS are likely to satisfy many of the GDPR
requirements already, needing a few adjustments to be made. Other organisa-
tions might decide to apply an ISMS as a general framework for the management
of the personal data of data subjects that they process, in the context of: (i) the
broader management of the information risks; (ii) the security of the data they
process, either in hard copy or in a digital version, as well as the relevant com-
pliance; (iii) the incident management; and (iv) addressing business continuity
issues. This work describes the necessary additional actions that an organisa-
tion is required to implement since they have already an ISMS in place to reach
compliance with the GDPR. Specifically, the fourteen control modules of Annex
A’ of ISO 27001 are presented, focusing on the lower level of analysis presented
in the ISO/IEC 27002 and providing extension of the corresponding controls, in
order to meet GDPR requirements, by focusing on data protection actions. That
means that if organisations already have an ISO 27001 framework in place, com-
pliance with GDPR requirements will not be necessitated a duplication of the
demanded effort. In addition, compliance to the GDPR is mandatory, whereas
ISO 27001 certification is not. Organisations can start from ISO 27001 certifica-
tion and reach GDPR compliance, or vice versa.
This work provides guidelines for practitioners of the domain of information
security and protection of privacy, since it presents a roadmap on how to design a
“towards GDPR compliance” project, contributing also to the awareness regard-
ing the protection of personal data of an organisation.
Future work of this study includes the validation of the proposed guidelines
to-wards GDPR compliance by a number of ISO 27001 certified organisations
that have also reached GDPR compliance. The analysis of such feedback will
further validate (or provide other perspectives to) the findings of this work.
Moreover, data protection officers could also be involved in this process, provid-
ing their experiences regarding the demanded effort to reach GDPR compliance
for an already ISO 27001 certified organisation.
References
1. European Commission: Directive 95/46/EC of the European parliament and
of the council. http://eur-lex.europa.eu/legal-content/EN/TXT/?uri=CELEX:
31995L0046. Accessed 14 May 2017
2. European Parliament: Regulation (EU) 2016/679 of the European parliament and
of the council of 27 April 2016 on the protection of natural persons with regard
to the processing of personal data and on the free movement of such data, and
repealing directive 95/46/EC (general data protection regulation) (2016)
256 V. Diamantopoulou et al.
3. Alberts, C., Dorofee, A., Stevens, J., Woody, C.: Introduction to the octave app-
roach. Technical report, Software Engineering Institute, Carnegie-Mellon Univer-
sity Pittsburgh, PA (2003)
4. Cavoukian, A., et al.: Privacy by design: the 7 foundational principles. Information
and Privacy Commissioner of Ontario, Canada 5 (2009)
5. CNIL 2018: Privacy impact assessment (PIA) - knowledge bases. Technical report
(2018)
6. CSA 2018: GDPR preparation and challenges survey report from cloud security
alliance (CSA). Technical report (2018). https://cloudsecurityalliance.org/articles/
gdpr-preparation-and-challenges-survey-report/. Accessed 09 July 2019
7. Diamantopoulou, V., Tsohou, A., Karyda, M.: General data protection regulation
and ISO/IEC 27001:2013: synergies of activities towards organisations’ compliance.
In: Gritzalis, S., Weippl, E.R., Katsikas, S.K., Anderst-Kotsis, G., Tjoa, A.M.,
Khalil, I. (eds.) TrustBus 2019. LNCS, vol. 11711, pp. 94–109. Springer, Cham
(2019). https://doi.org/10.1007/978-3-030-27813-7 7
8. ENISA: Recommended cryptographic measures - securing personal data. Technical
report (2013)
9. Ernst & Young 2018: Global forensic data analytics survey. Technical report
(2018). https://www.ey.com/Publication/vwLUAssets/ey-how-can-you-disrupt-
risk-in-an-era-of-digital-transformation/%24FILE/ey-how-can-you-disrupt-risk-
in-an-era-of-digital-transformation.pdf. Accessed 09 July 2019
10. Fredriksen, R., Kristiansen, M., Gran, B.A., Stølen, K., Opperud, T.A., Dimi-
trakos, T.: The CORAS framework for a model-based risk management process. In:
Anderson, S., Felici, M., Bologna, S. (eds.) SAFECOMP 2002. LNCS, vol. 2434, pp.
94–105. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-45732-1 11
11. Gartner 2017: Gartner says organizations are unprepared for the 2018 European
data protection regulation. Technical report (2017). https://www.gartner.com/en/
newsroom/press-releases/2017-05-03-gartner-says-organizations-are-unprepared-
for-the-2018-european-data-protection-regulation. Accessed 09 July 2019
12. IAAP: Privacy tech vendor report. Technical report (2018)
13. IAPP 2018: Annual governance report. Technical report (2018). https://iapp.org/
resources/article/iapp-ey-annual-governance-report-2018/. Accessed 09 July 2019
14. IAPP 2019: GDPR one year later: looking backward and forward. Technical
report (2019). https://iapp.org/news/a/gdpr-one-year-later-looking-backward-
and-forward/. Accessed 09 July 2019
15. ISO/IEC: ISO 27001:2013 information technology - security techniques - code of
practice for information security controls. Technical report (2013)
16. ISO/IEC: ISO 27001:2013 information technology - security techniques - informa-
tion security management systems - requirements. Technical report (2013)
17. Lambrinoudakis, C.: The general data protection regulation (GDPR) era: ten steps
for compliance of data processors and data controllers. In: Furnell, S., Mouratidis,
H., Pernul, G. (eds.) TrustBus 2018. LNCS, vol. 11033, pp. 3–8. Springer, Cham
(2018). https://doi.org/10.1007/978-3-319-98385-1 1
18. Palmieri III, N.F.: Data protection in an increasingly globalized world. Ind. LJ 94,
297 (2019)
19. Pfitzmann, A., Hansen, M.: A terminology for talking about privacy by
data minimization: Anonymity, unlinkability, undetectability, unobservability,
pseudonymity, and identity management (2010)
20. Spiekermann, S., Acquisti, A., Böhme, R., Hui, K.L.: The challenges of personal
data markets and privacy. Electron. Mark. 25(2), 161–167 (2015)
From ISO/IEC 27002:2013 to Personal Data Protection Controls 257
21. Thomson Reuters 2019: Study finds organizations are not ready for GDPR
compliance issues. Technical report (2019). https://legal.thomsonreuters.com/
en/insights/articles/study-finds-organizations-not-ready-gdpr-compliance-issues.
Accessed 09 July 2019
22. Working Party 29: Guidelines on data protection impact assessment. Technical
report (2019)
23. Yazar, Z.: A qualitative risk analysis and management tool-CRAMM. In: SANS
InfoSec Reading Room White Paper, vol. 11, pp. 12–32 (2002)
SPOSE Workshop
On the Trade-Off Between Privacy
and Utility in Mobile Services:
A Qualitative Study
1 Introduction
It is well understood that, almost every time people attempt to use mobile
services, they are making a decision to exchange their privacy for benefits. For
example, people often need to provide their location to obtain a real-time weather
forecast or share their interests to acquire accurate recommendations for goods
or activities. Indeed, the economic model behind most free mobile services is
based on such trade-offs: when using mobile services provided by Facebook or
Google, users do not directly pay the service providers money to download or
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 261–278, 2020.
https://doi.org/10.1007/978-3-030-42048-2_17
262 Y. Liu and A. Simpson
use their mobile services; rather, the mobile platforms collect a vast amount of
users’ personal data by analysing users’ in-app behaviour. The personal data is
then used to sell highly targeted ads.
People’s attitudes towards such privacy issues vary significantly. Some users
would be prepared to abandon the benefits provided by mobile services to protect
their personal data. According to a survey of 2,000 Americans conducted by the
Pew Research Center [6], 54% of mobile users have decided to not install an
app when they discovered how much personal information they would need to
share in order to use it and 30% of mobile users have uninstalled an app after
discovering it was collecting personal information that they did not wish to share.
Some others would enjoy the benefits without considering the potential privacy
risks. A field experiment [5] showed that about 93% of participants are willing
to provide personal data about their date of birth and monthly income for a 1
Euro discount for purchasing DVDs.
Various privacy-preserving technologies, with examples including permission
management tools [10,13], app analysing tools [2,26] and privacy-preserving
frameworks in specific fields [17,27], have been proposed to help people decrease
their privacy risks when using mobile devices. While the theoretical advantages of
such privacy-preserving technologies are clear, little adoption has been observed
in practice [4].
Online users’ privacy concerns may vary depending on the situation [25]. In
terms of mobile services, factors such as perceived convenience [20], perceived
financial cost [18], expected degree of information disclosure and potential pri-
vacy risks [12] all play important roles and lead people to varying decisions.
However, many questions still remain unanswered in this context. Such ques-
tions include: what factors influence people’s decision-making processes?; what
are the benefits of, and barriers to, using privacy-preserving technologies in this
context?; what direction might be followed to foster adoption of such technolo-
gies?; and do cultural differences play a role?
To start to address these questions, we conducted a qualitative study by
combining the methods of focus group sessions (two groups with five partici-
pants each), individual interviews (16 participants), and a questionnaire survey
(60 participants). The objectives of this study were: to investigate the human
factors influencing decision-making pertaining to the trade-off between privacy
and utility in mobile services; to explore the direction that might be followed to
foster adoption of privacy-preserving technologies that could help mobile users
to balance privacy and utility in more effective ways; and to study the impact
of cultural differences regarding the protection of privacy.
The paper contributes to the debate surrounding mobile privacy issues and
to research into the adoption of privacy-preserving technology. It does so by pre-
senting a detailed description of human factors influencing mobile users’ decision
making, and by presenting theoretical and practical guidelines for developers who
are responsible for designing privacy-aware mobile services and for researchers
who are interested in designing privacy-preserving mobile tools.
On the Trade-Off Between Privacy and Utility in Mobile Services 263
1. Privacy paradox. Despite the various levels of privacy awareness and knowl-
edge of participants in this study, people’s stated attitudes may be influenced
by the so-called privacy paradox. As such, users’ observed actions should also
be used to infer the level of privacy concern [23].
2. Cultural differences. Privacy concerns and related approaches of protect-
ing privacy vary between cultures [7]. The interview subjects of our study
(26 participants in total) were drawn from diverse cultures and backgrounds
(38% were British, 38% were Chinese, and 24% were participants from other
countries).
3. A complex array of factors. People’s decision-making can be affected by
a complex array of factors. It is important to declare that the human factors
we discuss in this study cannot fully explain the decision-making process. In
addition, these factors and their importance to the participants are relatively
subjective.
3 Methodology
We now introduce the qualitative approaches adopted in our study and present
details of the scenarios and the design of questions.
This study was conducted in Oxford, UK, in February 2017. Two primary qual-
itative approaches were adopted: focus group interviews (10 participants) and
individual interviews (16 participants). In addition, a questionnaire survey (60
participants) also provided auxiliary data.
On the Trade-Off Between Privacy and Utility in Mobile Services 265
Focus groups are useful for gathering detailed information about both per-
sonal and group feelings [14]. A broader range of information can also be collected
from the interaction and discussion among the participants. By contrast, indi-
vidual interviews are helpful for researchers to probe more deeply on each topic
and get information from non-verbal responses [9].
Two groups based on participants’ technology backgrounds were selected for
the current study. Members of the first group were doctoral students in the
field of system and software security. Instead of the strong knowledge of security
technology, the second group was internationally diverse: the members are adults
from five different countries with different occupational backgrounds. The focus
groups were set up in such a way to ensure appropriate interaction within each
group and to provide a clear contrast between different groups, enabling us to
better discuss how awareness and knowledge of privacy risks might affect the
decision-making process.
The participants of individual interviews were also categorised into two
groups to investigate the similarities and differences between attitudes between
young British citizens and young Chinese citizens. The interviews with Chinese
residents were conducted via Skype in the language of Chinese. The individual
interview groups were set up in such a way to collect comments from participants
from different cultures.
The group size of each focus group session was 5 and the size of each indi-
vidual interview category was 8. The total number of interview subjects was
26. All participants were asked to complete a questionnaire1 , which was used to
deduce participants’ privacy attitudes, as a pre-survey before they attended the
interview.
Participant recruitment was conducted by posting posters on Facebook and
Wechat Moments (social networking), and by circulating advertisements through
the mail lists of different departments, colleges, and clubs of Oxford University.
All of the subjects were offered a £10 Amazon gift voucher for their time. The
distribution of participants is shown in Table 1.
Wash et al. [29] argued that the result reflected by users’ actual behaviours is
more accurate than that collected by their self-reports when studying certain user
decisions. To better measure participants’ actual behaviours, a card-exchanging
game was applied in the focus group sessions to start off the discussion.
Participants were asked to play the roles of “mobile users” and “service
providers” in a two-round card-exchanging game. Five cards representing per-
sonal information (including gender, location and contacts list) were issued to
“mobile users” and five cards representing mobile services (e.g. real-time weather
forecast, recommendations of interested activities and a £5 Amazon voucher)
1
Full survey text can be accessed here:
https://yangliu.typeform.com/to/OEPM6f (English Version),
https://yang46.typeform.com/to/PR3oWD (Chinese Version).
266 Y. Liu and A. Simpson
were issued to “service providers”. In the first round, “mobile users” were asked
to select between zero and five information cards they can afford to disclose,
then exchange them with the corresponding number of service cards decided by
“service providers”. In this round, “mobile users” were not able to figure out the
service cards they could obtain before the trade was finally made. In the second
round, “service providers” were asked to show all their service cards before per-
forming the trade. Therefore, “mobile users” were able to choose the particular
service cards, and then exchange them with self-selected information cards.
Compared to traditional approaches (e.g. interview or focus group), the card
game has an advantage in collecting users’ actual behaviours. In the first round,
participants are blind to the exact services they might receive, whereas, in the
second round, participants are able to exchange specific service with self-selected
personal information. Participants’ actual behaviours under different conditions
are observed and recorded during the game. In addition, the card game served as
an icebreaker and helped participants to build mental associations with privacy
and utility exchange.
Apart from the card game, the same scenarios were applied in both the
focus group sessions and the individual interview sessions to collect participants’
comments. At the beginning of each session, five scenarios that mobile users
typically encountered in their everyday lives were presented. The aim was to
explore participants’ experience with balancing privacy and utility related to
mobile services. The scenarios were:
For each scenario, participants were requested to think about and comment
on the following topics.
4. In what ways do you think your personal information will be used? Do you
see any potential risks? (To explore participants’ understanding of privacy
risks.)
5. What utility do you think is obtained in this scenario? Do you think it is
worth the cost? (To explore participants’ feelings of, and attitudes toward,
the trade-off between privacy and utility.)
They were then presented with a brief introduction to the Privacy-Preserving
Targeted Mobile Advertising (PPTMA) framework [17] as an example technol-
ogy. The framework was applied to the five scenarios to showed how privacy-
preserving technologies could help mobile users to take control of their sensitive
information. The example also introduced the possibility of taking advantage of
mobile services without compromising users’ privacy.
After showing the examples, participants were requested to think about and
comment on the following topics.
1. What do you see as the benefits of using such privacy-preserving technologies?
2. What do you see as the barriers to using such privacy-preserving technologies?
3. Do you think such privacy-preserving technologies could influence your
decision-making process with regards to making trade-offs between privacy
and utility in mobile services?
4. Who do you think should be responsible for certifying that such privacy-
preserving technologies do what they say?
4 Results
The initial results indicate that the focus group members, individual interview
members and questionnaire survey participants demonstrated a degree of con-
sistency with regards to their mobile service use. For example, about 80% of
them spend more than 60 min on mobile services per day, with more than half of
the total participants spending more than 120 min. The devices are mostly used
for services such as social networking (90%), searching for information (88%),
creating and checking emails (63%), and online shopping (53%). Furthermore,
most of the participants had some experience in making decisions involving the
trade-off between privacy and utility in mobile services.
Questionnaire: To the best of your knowledge, which of the following information is technically
possible to be collected/deduced by an app provider when you use an app? (Multiple choices)
Item China (43 Western Frequency
participants) Countries (17) in total (60)
Your geographical location 38 16 54
Your phone number 29 15 44
Your contacts 26 12 38
Time of usage 20 17 37
Your email address 23 13 36
Your network operator 20 13 33
The list of apps installed on your 17 14 31
phone
. . . .
. . . .
. . . .
“If I know my privacy is compromised I would not make the trade-off. The
key problem is that in many cases I didn’t realise my personal information
is being collected.” (I2P6, Chinese, Female, Individual interview, Semi-
skilled worker)
“I personally feel safe. Maybe because that I don’t consider myself that
important or special. I can’t see why they want my information. So if
Google requires my location details to provide services I’ll give it without
thinking about the trade-off” (G2P3, Finnish, Female, Focus group, House
person)
On the other hand, many interviewees stated that they refused to make such
trade-offs when sensitive information is required for exchanging services.
“Yes, I have refused to install apps that ask for very specific information.
I tried to find substitutes in that case.” (G2P1, British, Female, Focus
group, Professional and managerial occupation)
“I didn’t pay for using the apps so I understand they need my information
to make money by some means. But for most apps, I would only use them if
they promise that the data would only be used in aggregate form.” (G1P4,
American, Male, Focus group, Cyber security doctoral student)
The findings suggests that limited awareness and knowledge of privacy leak-
age could positively affect the user’s acceptance of mobile services with risks of
disclosing personal information. It is noteworthy that some interviewees acknowl-
edged that they normally consider themselves to have serious privacy concerns
in a pre-survey; however, in the subsequent card-exchanging game (personal
information cards versus mobile services cards), they unconsciously ignored the
privacy risks and decided to make the trade-offs.
On the Trade-Off Between Privacy and Utility in Mobile Services 269
In addition, our findings show that, while most participants could identify
the risks of privacy leakage about the information with a high exposure rate from
public media (e.g. geographical location, phone number and contacts), certain
personal information that seems difficult to collect (but, in fact, is not) is rarely
considered by the users. Take “the retail price of your mobile” in Table 2 as an
example: it is straightforward to obtain the information of the phone model.
The phone model then leads to a precise retail price, which can then be used
to deduce the spending power of the phone owner. Overall, participants showed
relatively low awareness of such privacy leakage.
Trust in Service Providers. Earp and Baumer [8] found that, if a site or a
company is well-known, consumers would be more likely to disclose information
to it. Similar comments received in our study suggests that trusting the repu-
tation of a company would weaken users’ awareness of privacy risks from such
a company and positively impact the adoption of the trade-off. The following
representative comment indicates how trust can affect users’ privacy awareness
and subsequently influences decision-making:
“I feel safe to use some mobile services because the companies like Google
or Facebook are famous and so many people are using their services. Fur-
thermore, the government can help to monitor them.” (I2P4, Chinese,
Female, Individual interview, House person)
However, the extent that trust affects privacy awareness may vary accord-
ing to users’ knowledge of privacy risks. Although a good reputation is viewed
positively by participants with more privacy knowledge, such participants show
stronger concerns under the same situation: We also know no
leakage to bear the risks. For example, a statement made by a participant shows
the ineluctability of using mobile service such as Facebook:
“I rarely post pictures or comments to Facebook and I’m careful with
what I post and where from. However, as a small business, (to advertise my
business) I have no choice about being on Facebook. If I didn’t have to be I
would avoid it.” (I1P8, British, Female, Individual interview, Professional
and managerial occupation)
These statements suggest that, in some situations, the strong desire for
mobile services may overcome the awareness of privacy risks and encourage users
On the Trade-Off Between Privacy and Utility in Mobile Services 271
to disclose their information. For example, there is no substitute for the Face-
book app, and the Amazon app may provide significant benefits—such as useful
recommendations. The auxiliary data of Table 3 shows some factors that might
encourage users to give personal information to an app.
“Our personal information is not only exposed by mobile. From this aspect,
we don’t need to feel so sensitive. We are surrounding by internet, who
can escape?” (I2P7, Chinese, Female, Individual interview, Semi-skilled
worker)
“There are many ways to use our information. Maybe to improve the ser-
vice to make you use it more or maybe they just want to sell it or use with
advertisers. If we think too much we’ll have no app to use, because every
app is doing this.” (I1P6, British, Male, Individual interview, Student)
“It’s useful. But I think it’s difficult for my parents to learn how to use
it. They are not good at IT stuff.” (I2P6, Chinese, Female, Individual
interview, Semi-skilled worker)
“I get a bit annoyed once you use some apps and then you start to get
notifications (from the privacy-preserving tools) reminding you to protect
your data.” (G2P5, Australian, Male, Focus group, white collar worker)
On the Trade-Off Between Privacy and Utility in Mobile Services 273
“Maybe the government but can they be trusted? I hope that third party
standards exist but I’m not sure if they do and how or who could regulate
them.” (G2P1, British, Female, Focus group, Professional and managerial
occupation)
“To some extent the person downloading the app is responsible for check-
ing what data will be taken from them, because when he accesses the
(Android) app store, the app to be downloaded has already told the him
what permission it needs and what information it will collect.” (I1P4,
British, Female, Individual interview, Student)
play a role in monitoring technology providers, the corresponding figure for par-
ticipants from western countries was about 24% (4 out of 17). From a different
aspect, our findings suggest that users from collectivistic countries are more
likely to be comfortable with a government department protecting their individ-
ual privacy, while users from individualistic countries are more likely to see the
responsibility reside with some combination of individuals and non-governmental
organisations. Representative comments from each groups are as follows:
5 Discussion
We have reported the results of a user study that investigates the trade-off
between privacy and utility in mobile services. The research explored the human
factors that influence the decision-making process pertaining to the trade-off
and reported the major concerns of adopting privacy-preserving technologies. In
addition, the research explored the cultural differences regarding the protection
of privacy with participants from the UK and China as an example. Our findings
On the Trade-Off Between Privacy and Utility in Mobile Services 275
Acknowledgments. The authors thank the participants of the survey for their
valuable comments. We are grateful to the reviewers for their constructive and
helpful comments. We also wish to thank Norbert Nthala, Emma Osborn and
Aaron Ceross for discussions that helped to improve this work. This work is
partly supported by the National Key Research and Development Program of
China (2017YFB0802204), Key Research and Development Program for Guangdong
Province, China (2019B010136001), and Basic Research Project of Shenzhen, China
(JCYJ20180507183624136).
References
1. AppBrain. Google Play Stats (2017). http://www.appbrain.com/stats. Accessed
Dec 2017
2. Bal, G., Rannenberg, K., Hong, J.: Styx: design and evaluation of a new privacy
risk communication method for smartphones. In: Cuppens-Boulahia, N., Cuppens,
F., Jajodia, S., Abou El Kalam, A., Sans, T. (eds.) SEC 2014. IAICT, vol. 428, pp.
113–126. Springer, Heidelberg (2014). https://doi.org/10.1007/978-3-642-55415-
5 10
3. Barnes, S.B.: A privacy paradox: social networking in the united states. First Mon-
day 11(9) (2006). https://doi.org/10.5210/fm.v11i9.1394
4. Benenson, Z., Girard, A., Krontiris, I.: User acceptance factors for anonymous cre-
dentials: an empirical investigation. In: Workshop on the Economics of Information
Security (WEIS 2015) (2015)
5. Beresford, A.R., Kübler, D., Preibusch, S.: Unwillingness to pay for privacy: a field
experiment. Econ. Lett. 117(1), 25–27 (2012)
6. Boyles, J.L., Smith, A., Madden, M.: Privacy and data management on mobile
devices, vol. 4. Pew Internet & American Life Project (2012)
7. Cooper, R., Assal, H., Chiasson, S.: Cross-national privacy concerns on data collec-
tion by government agencies. In: Proceedings of the 15th International Conference
on Privacy, Security and Trust (PST), pp. 28–30 (2017)
On the Trade-Off Between Privacy and Utility in Mobile Services 277
8. Earp, J.B., Baumer, D.: Innovative web use to learn about consumer behavior and
online privacy. Commun. ACM 46(4), 81–83 (2003)
9. Gorden, R.L.: Interviewing: Strategy, Techniques, and Tactics. Dorsey Press,
Homewood (1969)
10. Holavanalli, S., et al.: Flow permissions for Android. In: Proceedings of the 28th
IEEE/ACM International Conference on Automated Software Engineering (ASE
2013), pp. 652–657. IEEE, Palo Alto (2013)
11. Kayes, I., Kourtellis, N., Quercia, D., Iamnitchi, A., Bonchi, F.: Cultures in commu-
nity question answering. In: Proceedings of the 26th ACM Conference on Hypertext
& Social Media (HT 2015), pp. 175–184. ACM (2015)
12. Keith, M.J., Thompson, S.C., Hale, J., Lowry, P.B., Greer, C.: Information disclo-
sure on mobile devices: re-examining privacy calculus with actual user behavior.
Int. J. Hum Comput Stud. 71(12), 1163–1173 (2013)
13. Kern, M., Sametinger, J.: Permission tracking in Android. In: Proceedings of the
6th International Conference on Mobile Ubiquitous Computing, Systems, Services
and Technologies (UBICOMM 2012), Barcelona, Spain, pp. 148–155 (2012)
14. Kitzinger, J.: Qualitative research: introducing focus groups. BMJ: Br. Med. J.
311(7000), 299 (1995)
15. Liu, F., Zhao, X., Chau, P.Y., Tang, Q.: Roles of perceived value and individual
differences in the acceptance of mobile coupon applications. Internet Res. 25(3),
471–495 (2015)
16. Liu, Y., Simpson, A.: Privacy-preserving targeted mobile advertising: formal mod-
els and analysis. In: Livraga, G., Torra, V., Aldini, A., Martinelli, F., Suri, N. (eds.)
DPM/QASA -2016. LNCS, vol. 9963, pp. 94–110. Springer, Cham (2016). https://
doi.org/10.1007/978-3-319-47072-6 7
17. Liu, Y., Simpson, A.C.: Privacy-preserving targeted mobile advertising: require-
ments, design and a prototype implementation. Softw.: Pract. Exp. 46(12), 1657–
1684 (2016)
18. Luo, X., Li, H., Zhang, J., Shim, J.P.: Examining multi-dimensional trust and
multi-faceted risk in initial acceptance of emerging technologies: an empirical study
of mobile banking services. Decis. Support Syst. 49(2), 222–234 (2010)
19. Marvin, G.: Survey: 3 out of 4 consumers now notice retargeted ads (2016).
http://marketingland.com/3-out-4-consumers-notice-retargeted-ads-67813.
Accessed July 2017
20. Okazaki, S., Mendez, F.: Exploring convenience in mobile commerce: moderating
effects of gender. Comput. Hum. Behav. 29(3), 1234–1242 (2013)
21. Phelan, C., Lampe, C., Resnick, P.: It’s creepy, but it doesn’t bother me. In:
Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems
(CHI 2016), pp. 5240–5251. ACM (2016)
22. Pötzsch, S.: Privacy awareness: a means to solve the privacy paradox? In: Matyáš,
V., Fischer-Hübner, S., Cvrček, D., Švenda, P. (eds.) Privacy and Identity 2008.
IAICT, vol. 298, pp. 226–236. Springer, Heidelberg (2009). https://doi.org/10.
1007/978-3-642-03315-5 17
23. Preibusch, S.: Guide to measuring privacy concern: review of survey and observa-
tional instruments. Int. J. Hum. Comput. Stud. 71(12), 1133–1143 (2013)
24. Sabouri, A.: On the user acceptance of privacy-preserving attribute-based creden-
tials – a qualitative study. In: Livraga, G., Torra, V., Aldini, A., Martinelli, F.,
Suri, N. (eds.) DPM/QASA -2016. LNCS, vol. 9963, pp. 130–145. Springer, Cham
(2016). https://doi.org/10.1007/978-3-319-47072-6 9
25. Sheehan, K.B.: Toward a typology of internet users and online privacy concerns.
Inf. Soc. 18(1), 21–32 (2002)
278 Y. Liu and A. Simpson
26. Taylor, V.F., Martinovic, I.: Securank: starving permission-hungry apps using con-
textual permission analysis. In: Proceedings of the 6th Workshop on Security and
Privacy in Smartphones and Mobile Devices, SPSM 2016, pp. 43–52. ACM, New
York (2016)
27. Toubiana, V., Narayanan, A., Boneh, D., Nissenbaum, H., Barocas, S.: Adnostic:
privacy preserving targeted advertising. In: Proceedings of the 17th Annual Net-
work and Distributed System Security Symposium (NDSS), San Diego, CA, USA,
p. 2010 (2010)
28. Virkki, J., Chen, L.: Personal perspectives: individual privacy in the IoT. Adv.
Internet Things 3(02), 21 (2013)
29. Wash, R., Rader, E., Fennell, C.: Can people self-report security accurately?: agree-
ment between self-report and behavioral measures. In: Proceedings of the 2017 CHI
Conference on Human Factors in Computing Systems (CHI 2017), pp. 2228–2232.
ACM (2017)
Analysis of Automation Potentials
in Privacy Impact Assessment Processes
Jan Zibuschka(B)
Abstract. With the recent introduction of the EU’s General Data Pro-
tection Regulation (GDPR), privacy impact assessments (PIA) have
become mandatory in many cases. To support organisations in correctly
implementing those, researchers and practitioners have provided refer-
ence processes and tooling. Integrating automation features into PIA
tools can streamline the implementation of compliant privacy impact
assessments in organizations. Based on a general reference architecture
and reference process based on guidance by authorities, this contribu-
tion offers a systematic analysis of which process steps show the most
promise with regard to this, and discusses impediments to this approach
and directions for future research.
1 Introduction
In its Article 35, the recent European General Data Protection Regulation man-
dates the controller performs a privacy impact assessment where processing of
personal information is likely to result in a high risk to the rights and freedoms
of individuals [15]. Specifically, this applies to cases of profiling, surveillance
of public places, or large-scale processing of personal information [3]. Similar
requirements are set forth by other international regulators, e.g. for governmen-
tal information processing in Canada [18]. Organizations across Europe have
established processes to perform such privacy impact assessments, across all
business sectors, including industry, telecommunications, and electronics [19].
Whether those were implemented in response to the new legal requirement or
existed previously, in any case the processes need to be reviewed in light of new
regulation.
Enterprise information systems are proven tool to integrate and extend busi-
ness processes [21], and also offer the chance to prescribe specific processes. While
this effect may be detrimental to organisation if it is not employed consciously
[7], implementing enterprise systems supporting privacy impact analysis holds
the promise to ensure compliance. Additionally, offering tool support for PIA1
processes reduces time and effort organizations need to spend, which have been
cited as main arguments against mandatory PIAs [18].
This contribution reviews a process [4] and tooling [5] proposed by some
European data protection authorities, identifying subprocesses that are not cur-
rently supported by tools, but offer a significant potential to be supported by
automation features in next generation tools.
The rest of the paper is structured as follows: In the following section, we pro-
vide related work, including processes and tooling for privacy impact assessments
as well as related disciplines, specifically security engineering and model-based
engineering. After this, we introduce some assumptions about tool architecture
and capabilities, which form the basis for the following section, which holds
a step-by-step analysis of automation potentials in a specific PIA process. We
conclude by discussing our findings and next steps.
2 Related Work
1
we use the terms privacy impact assessment and data protection impact assessment
interchangeably.
PIA Automation Potentials 281
Privacy Engineer
using PIA tool
Fig. 1. System overview (solid lines: current-gen (metainformation), dotted lines: next-
gen (system access).)
4 Process Analysis
This section provides a step-by-step walk-through of PIA subprocesses based on
the PIA process proposed by [4]. For each subprocess, it gives an analysis of
existing tooling and automation potential.
A3
Target of evaluation:
description of the sys-
A2 tem, identification of
Projecting the data and data flows A5
A1
assessment: definition Identification of Documentation of
Relevance threshold:
of DPIA’s scope relevant legal tasks and issues
is a DPIA necessary?
and identification requirements
of DPIA team A4
Identification of ac-
tors involved/persons
concerned
Fig. 2. Tooling in preparation stage [4] (dotted dark: current-gen tooling; dashed light:
next-gen automation potentials.)
B2
B3
B1 Identification
Identification of B4
Identification of of potential at-
evaluation criteria Evaluation of risk
protection goals tackers, motives
and benchmarks
and objectives
Catalogue of typical
objectives, attackers
and consequences
Fig. 3. Tooling in evaluation stage [4] (dotted dark: current-gen tooling; dashed light:
next-gen automation potentials.)
Overall, tooling during the evaluation stage, especially threat modelling tools,
are already common in the security space, and the most advanced functions
284 J. Zibuschka
provided by current PIA tooling also fall into this subprocess. There is addi-
tional potential for extending the underlying knowledge bases into the privacy
space and offering specialized knowledge bases defining uniform attacker models
and evaluation benchmarks for various domains and/or organizations, offering
a standardized basis for the evaluation of risk (step B4, see Fig. 3). We do not
envision the automation of the evaluation of risk itself, as assessment of e.g. the
proportionality of processing typically performed in this step [4] can hardly be
performed in an unsupervised fashion.
C2
C3 C5
Documentation of
DPIA Report Publication of Auditing of Eval-
evaluation results in
DPIA Report uation Results
standardized form
C1
Identification of ap-
propriate safeguards
C4
Implementation
of safeguards
Catalogue of
typical safeguards
Fig. 4. Tooling in report and safeguards stage [4] (dotted dark: current-gen tooling;
dashed light: next-gen automation potentials.)
5 Discussion
There are clear directions for future research. We do not offer a complete system
architecture implementing automation in privacy impact assessments. Clearly,
integrating existing project management, privacy impact assessment, and threat
modelling tools, and providing integration with the target system has the promise
of significantly lowering efforts for PIA implementation, addressing one of the
core arguments of its opponents [18]. To address additional effort caused by
the need to model the target system, we propose automatic generation of the
model, with the long term goal of round-trip engineering, a proven approach from
other model-driven engineering domains [13]. Experience from security response
centres also suggests that tooling giving access to target systems can address
many pain points in day-to-day operations [14].
The less technical nature of determining the privacy impact, compared to
more technical security engineering approaches, may pose additional problems
for concrete implementations due to the interdisciplinary nature of privacy anal-
yses, covering legal, economic, and societal in addition to technical aspects [11].
We need to avoid codifying in technology [7], and specifically automating, things
that do not necessarily apply in all cases from a legal perspective [10]. In addition,
adding automation in privacy impact assessments may reduce the awareness for
privacy issues generated by a manual privacy impact assessment [19]. This con-
cern may already apply to the current-gen CNIL tool, which we do not envision
to replace but merely extend.
We offer this contribution as a basis for discussion of the merits and flaws of
such systems. To evaluate the real-world applicability of our approach, it should
be evaluated what does and does not work for various target systems, and which
part of typical target systems would remain auxiliary and inaccessible by PIA
tools.
References
1. Ahmadian, A.S., Strüber, D., Riediger, V., Jürjens, J.: Supporting privacy impact
assessment by model-based privacy analysis. In: Proceedings of the 33rd Annual
ACM Symposium on Applied Computing, pp. 1467–1474. ACM Press (2018)
2. Baskerville, R.: Information systems security design methods: implications for
information systems development. ACM Comput. Surv. 25(4), 375–414 (1993)
3. Bieker, F., Bremert, B., Hansen, M.: Die Risikobeurteilung nach der DSGVO.
Datenschutz Datensicherheit DuD 42(8), 492–496 (2018)
4. Bieker, F., Friedewald, M., Hansen, M., Obersteller, H., Rost, M.: A process for
data protection impact assessment under the European general data protection
regulation. In: Schiffner, S., Serna, J., Ikonomou, D., Rannenberg, K. (eds.) APF
2016. LNCS, vol. 9857, pp. 21–37. Springer, Cham (2016). https://doi.org/10.1007/
978-3-319-44760-5 2
5. CNIL: PIA Software (2019). https://www.cnil.fr/en/open-source-pia-software-
helps-carry-out-data-protection-impact-assesment/. Accessed 26 June 2019
286 J. Zibuschka
6. Dewitte, P., et al.: A comparison of system description models for data protection
by design. In: Proceedings of the 34th ACM/SIGAPP Symposium on Applied
Computing, pp. 1512–1515. ACM Press (2019)
7. Gosain, S.: Enterprise information systems as objects and carriers of institutional
forces: the new iron cage? J. AIS 5(4), 6 (2004)
8. Kokolakis, S., Demopoulos, A., Kiountouzis, E.: The use of business process mod-
elling in information systems security analysis and design. Inf. Manag. Comput.
Secur. 8(3), 107–116 (2000)
9. Montesino, R., Fenz, S.: Information security automation: how far can we go? In:
Sixth International Conference on Availability, Reliability and Security, pp. 280–
285. IEEE (2011)
10. Pagallo, U., Durante, M.: The pros and cons of legal automation and its governance.
Eur. J. Risk Regul. 7(2), 323–334 (2016)
11. Radmacher, M., Zibuschka, J., Scherner, T., Fritsch, L., Rannenberg, K.: Pri-
vatsphärenfreundliche topozentrische Dienste unter Berücksichtigung rechtlicher,
technischer und wirtschaftlicher Restriktionen. In: 8 Internationale Tagung
Wirtschaftsinformatik 2007 - Band 1, pp. 237–254 (2007)
12. Sion, L., Van Landuyt, D., Yskout, K., Joosen, W.: SPARTA: security & privacy
architecture through risk-driven threat assessment. In: 2018 International Confer-
ence on Software Architecture Companion, pp. 89–92. IEEE (2018)
13. Streitferdt, D., Wendt, G., Nenninger, P., Nyßen, A., Lichter, H.: Model driven
development challenges in the automation domain. In: 32nd Annual Computer
Software and Applications Conference, pp. 1372–1375. IEEE (2008)
14. Sundaramurthy, S.C., McHugh, J., Ou, X., Wesch, M., Bardas, A.G., Rajagopalan,
S.R.: Turning contradictions into innovations or: how we learned to stop whining
and improve security operations. In: SOUPS 2016, pp. 237–251. USENIX Associ-
ation, Denver (2016)
15. Tikkinen-Piri, C., Rohunen, A., Markkula, J.: EU general data protection regu-
lation: changes and implications for personal data collecting companies. Comput.
Law Secur. Rev. 34(1), 134–153 (2018)
16. Tsohou, A., et al.: Privacy, security, legal and technology acceptance requirements
for a GDPR compliance platform. In: SECPRE Workshop at ESORICS 2019.
Springer, Luxembourg (2019)
17. Vermeulen, C., Von Solms, R.: The information security management toolbox -
taking the pain out of security management. Inf. Manag. Comput. Secur. 10(3),
119–125 (2002)
18. Wright, D.: Should privacy impact assessments be mandatory? Commun. ACM
54(8), 121 (2011)
19. Wright, D.: The state of the art in privacy impact assessment. Comput. Law Secur.
Rev. 28(1), 54–61 (2012)
20. Wuyts, K., Scandariato, R., Joosen, W.: Empirical evaluation of a privacy-focused
threat modeling methodology. J. Syst. Softw. 96, 122–138 (2014)
21. Xu, L.D.: Enterprise systems: state-of-the-art and future trends. IEEE Trans. Ind.
Inform. 7(4), 630–640 (2011)
22. Zibuschka, J., Nofer, M., Zimmermann, C., Hinz, O.: Users’ preferences concerning
privacy properties of assistant systems on the Internet of Things. In: Proceedings
of the Twenty-fifth Americas Conference on Information Systems, AIS (2019)
An Insight into Decisive Factors in Cloud
Provider Selection with a Focus
on Security
Abstract. In the last ten years cloud computing has developed from a
buzz word to the new computing paradigm on a global scale. Comput-
ing power or storage capacity can be bought and consumed flexibly and
on-demand, which opens up new opportunities for cost-saving and data
processing. However, it also goes with security concerns as it represents
a form of IT outsourcing. We investigate how these concerns manifest
as a decisive factor in cloud provider selection by interviews with eight
practitioners from German companies. As only a moderate interest is
discovered, it is further examined why this is the case. Additionally,
we compared the results from a systematic literature survey on cloud
security assurance to cloud customers’ verification of their providers’
security measures. This paper provides a qualitative in-depth examina-
tion of companies’ attitudes towards security in the cloud. The results
of the analysed sample show that security is not necessarily decisive in
cloud provider selection. Nevertheless, providers are required to guaran-
tee security and comply. Traditional forms of assurance techniques play
a role in assessing cloud providers and verifying their security measures.
Moreover, compliance is identified as a strong driver to pursue security
and assurance.
1 Introduction
Cloud Computing has been emerging as the new computing paradigm in the last
ten years, enabling consumers to purchase computing power and storage capacity
on-demand, conveniently and cost efficiently from specialized providers. Recent
studies claim that cloud computing has left the hype phase behind and can
already be considered the norm for IT [10].
Besides the potential economic benefits of cloud adoption, it also goes with
security concerns as it represents a form of IT outsourcing and exhibits tech-
nological peculiarities concerning size, structure and geographical dispersion
[35]. With rising adoption rates of cloud services, security concerns remained
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 287–306, 2020.
https://doi.org/10.1007/978-3-030-42048-2_19
288 S. Pape and J. Stankovic
unchanged or even rose as well. On the other hand, many technical reports also
reveal benefits to security in the cloud. It is argued that a cloud provider (CP)
enjoys economies of scale in terms of security as well, being able to invest more
and thereby achieve a higher security level on a much larger scale than most
client companies would with an in-house data centre [24,29]. Thus, in either
case, one would expect companies to incorporate security into their provider
selection and cloud use.
We investigate organizations’ practises when selecting a secure CP: “What
role does security play in CP selection?”. Despite expected “inherent differences
in such things as the intended purpose, assets held, legal obligations, exposure
to the public, threats faced, and tolerance to risk” between different companies
or organizations [29], we expected to verify the importance of security. Under
that assumption there would be an incentive for providers to invest in secu-
rity measures, as potential customers might make their choice based on this
characteristic [24]. Moreover, in order to prevent a market for lemons in cloud
computing [1], we expected cloud service providers and customers to come up
with quality/security assurance methods. Thus, we intended the follow-up ques-
tion: How are the providers’ security measures verified? – if security is a selection
criteria. Or respectively: Why is security not considered in CP selection?
In order to find answers for the underlying research questions a qualitative
approach is taken. Practitioners from eight German companies who are associ-
ated with CP selection are interviewed and questioned about their companies’
provider selection and ways to establish assurance.
2 Related Work
Our research questions can be related to contributions on provider selection,
the role of security and security assurance. Security concerns, which are seen as
the inhibiting factor of cloud adoption, can be easily related to well researched
issues. A bunch of issues is related to technical properties of cloud computing,
i.e. the complex architecture [29], multi-tenancy in connection with isolation
failures [24,29], and network vulnerabilities The list of risks also includes the
threat of a malicious insider on the CP’s side [9], who may abuse his privileges.
However, this is a general outsourcing issues due to a loss of governance which
can bear dangers for the cloud customers [24]. Therefore, focus in this section is
on measures for the CP to assure the security level of its service (corresponding
to our extended research question). Assurance is also often necessary from a
legal and compliance perspective since most companies underlie a variety of
legal obligations, depending on the sector and the type of data they handle.
Since we follow the qualitative content analysis method which is considered
hermeneutic and uses deductive examination (cf. Sect. 3.2), an inherent under-
standing of the topic was necessary in order to interpret the material. Therefore,
we conducted a systematic literature survey on security assurance measures.
Decisive Factors in Cloud Provider Selection with a Focus on Security 289
We rely on a survey from Ardagna et al. [7] which covers contributions on security
measures and assurance techniques until 2014 and followed their methods and
definitions as close as possible to update it for our recent research. Due to space
limitations, we can not show the results in detail, but only give a brief summary
and list them in Table 1.
Almost all contributions reasoned with customers’ security concerns as the
main inhibiting factor of cloud adoption and that a contribution might provide
the needed transparency to resolve that issue. A further justification for new
contributions on security assurance were the “special properties” of the cloud
which raise new requirements for that topic. Clearly each contribution presented
the benefits of its solution, some also covered the challenges, but the drawbacks
of certain assurance techniques could only be found in a few contributions from
adjacent categories. Certification and security SLAs were presented as the more
accessible and convenient measures. In these contributions the customer is clearly
involved in the negotiation and provider choice. On the contrary, contributions
on auditing, monitoring and testing are mostly technical models or frameworks.
290 S. Pape and J. Stankovic
It might be difficult to apply these technical models and is it not clear if they
are practical in reality and who would implement them.
2.2 CP Selection
In this section qualitative research which determined relevant criteria for CP
selection will be discussed. The presented contributions suggest a formal and
systematic selection process of a CP and identify security as a relevant criterion.
They pursue similar research questions and use a qualitative approach like we do.
Nevertheless, their results are narrowed down into compact lists, where security
is identified as a requirement but not further discussed. We aim to close this
gap, by giving further insight into experts’ answers and the role of security.
Repschläger et al. [60] develop a CP classification model with a focus on
infrastructure as a service (IaaS). The relevant target dimensions are determined
as a result of expert interviews and validated and expanded through a literature
review. The authors conduct five interviews with experts providing different
perspectives on common objectives in cloud computing.
Similarly, Hetzenecker et al. [26] derive a model of requirements to support
the user in evaluating CPs. Their model consists of six categories with in total
41 requirements. “Information security” is derived as a category with 15 require-
ments, such as integrity, availability, data disposal, encryption or scalability. All
requirements are only presented by a title but not further elaborated.
Lang et al. [39] conduct a Delphi study with 19 decision makers in order to
determine relevant selection criteria with a high abstraction level. Security is only
identified as a component of the highest rated criterion “functionality” which
does not permit to make any statements about the importance of security at all.
The authors call for further research to investigate their identified requirements
on a lower abstraction level.
3 Methodology
In this section we briefly describe how the interviews were conducted and how
the data was analysed.
Decisive Factors in Cloud Provider Selection with a Focus on Security 291
the material, in the first phase of analysis each interview was summarized and
the peculiarities of the given answers were noted. Next, master-codes were devel-
oped and tested on the first three interviews before coding the whole material.
These codes were generated mostly deductively out of the interview questions.
For instance, the codes “Provider Selection” and “Assurance Techniques” were
rather straight forward, as these where the main research questions. The result
of this phase was a list of master-codes. After coding the whole material with the
master-codes, all passages coded with the same master-code were grouped and
reread. At this point the aim was to differentiate the master-codes by inductively
deriving sub-codes for each master-code. While proceeding from one interview
to the next, the generated sub-codes were revised and sorted. The final product
was a list of sub-codes which differentiated the master-codes. A sample of the
derived coding can be found in Table 4.
294 S. Pape and J. Stankovic
4 Interview Results
The interviews and the data analysis were conducted with regard to the initial
research questions. This resulted in a coding frame of five master-codes from
which three address our research questions directly. In the next subsections,
we briefly show the results of the role of security in CP selection, reasons for a
moderate interest in security, and the verification of providers’ security measures.
Since in most of the interviews compliance was strongly connected with security,
we also investigated the role of the General Data Protection Regulation (GDPR).
R5 stressed the importance of creating awareness in-house for the new technology
and its specific risks.
Encryption: Four respondents reported encryption as a mean to secure the
cloud. R6 and R8 attached great importance on encrypting their outsourced data
and R1 and R2 reported on means of encryption implemented by their clients.
Additionally, R2 pointed out the potential drawbacks for the cloud customer. R2:
When we provide the infrastructure only, encryption is mostly in the hands of the
customer. But then he has to manage the keys, which represents an additional
complexity he has to handle.
Data Criticality: In addition, some users saw security relatively to the critical-
ity of data they placed into the cloud. R1 and R6 stated that business critical-
data was preferably not outsourced at all. R1: In my opinion, it will always be
the case that for a certain part the companies say: “These are my crown jewels,
which I don’t give away. No matter how much I trust a provider, I want to have
these with me”.
Trust: As the opposite side of mitigation, ideas were raised resonating with trust
towards the provider. Maybe the most prominent statement to this topic was
given by R1: I believe that many give their providers a few laurels in advance.
“Okay they do this on such a large scale and I either I do not trust them per
se. In this case I address encryption and other topics. Or as I said, I give them
laurels in advance and say, yes this is going to work out”, assuming that many
users trust their providers without any proof. R2, R4, R5 and R8 expressed their
belief that the incentives for providers were set in such a way that they cannot
afford to make mistakes with customers’ data.
Personal Responsibility: R2 tried to explain the popularity of Amazon with
the “IBM Effect”. R2: Well I can rely on them (AWS), at least at most times.
And when there is a service failure, it applies to everyone and one can say: “Yes,
you know it, AWS just had an outage”. So it’s the IBM effect: “No one ever got
fired for buying IBM”, applies to AWS nowadays. R3 agreed with this idea.
Finally, independently of mitigation or trust one question had to be included
in light of the given answers concerning the importance of security. Throughout
some discussions one could have gotten the impression that some companies
simply avoided being held accountable in case of a data breach. Therefore the
respondents were asked whether there was a personal responsibility or even an
intrinsic motivation to pursue security conscientiously. Consequently, the code
“Personal Responsibility” was covered with six respondents.
Compliance: The resulting discussions with R1 and R2 were leaned on the
fulfillment of GDPR and compliance requirements and both respondents revealed
the belief that the choice of a secure provider is rather extrinsically motivated
by the need to comply. They also agreed that the regulating authorities still
have not drawn any consequences but most likely would do so in the future
in order to set an example. R1: [...] I believe that many (companies) still wait
until the first penalties are issued, as surprisingly it (GDPR) did not have that
Decisive Factors in Cloud Provider Selection with a Focus on Security 297
many impact yet. [...] I think the first time something happens and jurisdiction
is drawn, and a company really has to pay for it, many others will have a second
awakening. R4 and R6 agreed that compliance is decisive for the final choice.
However, according to R4 intrinsic motivation is individual and depends on the
employee’s training. R4: Well it depends on who is dealing with the topic. As I
already said, the energy sector has very high security requirements, so if a classic
energy economist deals with it, then security and compliance are in his blood. [...]
If it is a developer, he may not care. He only asks where to put the data, but
does not really think about it himself. However, R4 adds that in recent years the
awareness has risen among all the employees.
benefits of a third party audit, namely that an expert was checking the status of
a system and giving advice on how to improve it, which was according to R8 an
advantage compared to certificates. While R4 doubted the competence of some
auditors, R8 pointed out the conflict of interest. R8: Exactly, it depends on what
kind of auditor you get. You can entrust someone who issues an affirmation for
you: “Audit accomplished”, or you can entrust someone who works conscien-
tiously. The only problem is that the ones who work conscientiously, are often
those who are not well received and afterwards have trouble reselling. There is a
slight conflict of interest.
Contracts: It was often discussed in connection to assurance that respondents
had contractual agreements with their providers (R4 and R6). R6 added the
possibility to contractually seal where data is located and processed. R2 pointed
out that contractual agreements were often not only an option but a require-
ment in light of GDPR, while R4 and R6 gave the important reason for having
a contractual agreement, namely that in case of non-fulfillment a compensation
was ensured. R1, R2 and R4 mentioned the possibility to contractually include
the users’ right to visit the data center in person. According to R2 such a clause
may be necessary or important to a client, who handles personal data. Neverthe-
less, the respondents admitted that in reality such a visit hardly ever happened.
Additionally, R2 doubted the sense of sending company representatives to visit
a data center. R2: If someone like you or me went there, what would we be sup-
posed to see? If the door is not open somewhere or a cable hanging loosely, we
would have no idea how secure this is and whether it is in accordance to the
norm. R1 added that the providers tried to avoid such visits as they considered
the interior of their data centre as a company secret. Additionally, checking tech-
nical documentation or documentation of processes was found in the interviews
(R4, R6, R7).
Tests: Additionally, R4 and R6 talked about security tests as a mean of assur-
ance. R6: That means that for a cloud service we will not check whether it is
externally attackable, as most data centres must have tested this already for about
five-, six-, seven-, eight hundred times. What we check is whether the access point
we have to the data centre is secure enough. R4 also stressed that the tests were
not done on the CPs’ side but on the final application, which was supposed to
run in the cloud or as a hybrid application. Both respondents pointed out some
drawbacks of penetration-testing, first the costliness and second that such tests
could only be run for known cases.
Two respondents stood out with their companies’ specific assurance tech-
niques. R5 reported of his companies’ own cloud risk process which helped eval-
uating a provider with regard to the risk he poses to the company and its data.
The process incorporated some of the already presented techniques, like demand-
ing a certification and contractually sealing requirements, but more than this,
it was a spreadsheet for assessing the likeliness of scenarios and finally present-
ing the risk imposed by a provider. Finally, the management was in charge of
deciding whether this risk was acceptable or not. The other individual measure
Decisive Factors in Cloud Provider Selection with a Focus on Security 299
was taken by R4’s company, which had designed their own questionnaire for CPs
comparable to the CAIQ by the CSA.
Finally, besides all the collected assurance techniques it has to be mentioned
that several respondents also expressed scepticism when talking about assurance.
According to R3 there was no gain from SLAs and contracts, as even if there
was a written agreement one had to suffer in case of a data breach in terms of
data loss. R4 pointed out the drawback of a third party audit, by telling his
own experience with auditors who believed him anything he told them. R7 had
doubts about assurance in general and pointed out how the need to control or
verify everything although one had outsourced brought unnecessary costliness.
Similarly, R8 criticized that certificates do not show real security.
Due to the previous answers, we also elaborate how the GDPR influenced the
decisions and to what extent interviewees reported about German and European
cloud services which do not transfer data outside of the European Union.
GDPR: According to R2 and R6, a result of the GDPR is that more attention is
turned to data protection. R2 claims that the GDPR allows to ensure technical
and organisational measures by SLAs more easily.
R1 and R2 agree that since so far data protection authorities have not pun-
ished companies by a fine, most companies will assume the first cases will hit
large companies and wait for that. R2 was more concerned about written warn-
ings from competitors. R7 reported that his company’s data security officer
answered to a request about using cloud services that an agreement of the par-
ent company (in Great Britain) with the cloud provider is seen as valid for all
subsidiary companies. In contrast, R4 reported that the regulation requires data
centres in the EU, which still did not work out for them, because of US employ-
ees with access to the stored data. However, they use a CP in Switzerland for
non business critical data.
Localisation of CPs: Statements on the localisation of CPs were ambivalent.
On the one hand, R3 was concerned about US industrial espionage facilitated
by war on terror laws and thus demands a German/European solution with all
components (software, hardware) built and run in Germany/EU. This is in line
with the report of a “Robin Hood” bonus for a localised offer (R2).
On the other hand R1 and R2 report that at the beginning localisation
seemed important, but then lost importance due to data centres in Germany
(from the large CPs) and due to observations of other companies seemingly
running their cloud services GDPR-compliant with non-EU CPs. An additional
argument was that the advantages of localisation can not compensate higher
costs (R3, R4, R7), missing features (R1, R2) or development tools (R3) for the
German version, customers in the US (R1), and missing trust in the continuity
of the service (R4). Many interviewees (R1, R2, R3, R4, R7) were referring to
300 S. Pape and J. Stankovic
the “German cloud”, a cooperation between Telekom and Microsoft which was
ended last year1 .
5 Discussion
Role of Security: With regard to the original question on the role of security in
cloud provider selection the collected findings are ambiguous. Selection criteria
like usability and costs were expressed straightforwardly and matched the find-
ings of the related work [26,60]. Security however, was never the first answer the
respondents extensively engaged in. Neither could they provide concrete secu-
rity requirements comparable to those found in the related contributions. On the
other hand, security as a requirement was present in all the discussions. More-
over, availability and in rare cases confidentiality could be extracted as goals.
Two respondents revealed that although security had not been a selection crite-
rion, it was considered in retrospect in some cases, where the companies analysed
the services after having tested them first. Moreover, the findings from this sam-
ple challenge the idea of a systematic provider selection suggested in related
works. In this sample it was rarely the case that providers were compared and
evaluated in advance with regard to certain criteria.
Moderate Interest in Security: Some respondents assessed the situation and
acted in accordance to the mitigation measures proposed in cloud organizations’
technical reports. For instance, one could identify the awareness of the separation
of duties and the willingness to employ encryption on the user side. These users
were aware that security in the cloud was not only the cloud provider’s duty and
took own responsibility. On the other hand, namely the capability of a provider
to grant compensations speaks however again for a financial interest rather than
an intrinsic motivation to establish security. The initial assumption that the
requirement on security is extrinsically motivated by compliance was clearly
supported by the respondents’ answers on personal responsibility. The answers
revealed as well a different side to the client provider relationship, which was a
great amount of trust towards the cloud provider and the acceptance of risk to
a certain extent. The idea that an “IBM effect” exists when choosing Amazon’s
services indicates that this could be a way for decision makers to be exonerated
from responsibility.
Security Assurance: Overall, the respondents revealed to rely on certifica-
tion, audits, contractual agreements and testing as common means of assurance.
Besides those assurance techniques, two respondents presented own company-
specific methods. The results from this sample show that except for C5 which
is a cloud-specific certificate and audit, the companies rather rely on traditional
forms of assurance than cloud-specific ones. Especially contractual agreements
are considered a convenient method in order to establish compliance and guar-
antee for a compensation in case of non-fulfillment. Surprisingly, contractually
agreed measures like data center visits are not often undertaken. These findings
1
https://heise.de/-4152650.
Decisive Factors in Cloud Provider Selection with a Focus on Security 301
are one more indicator that security and also assurance are overshadowed by
compliance, but that at the same time regulation may miss out on establishing
real and not only paper-based assurance.
In comparison to the findings from academic literature cloud-specific assur-
ance techniques seemed to have not really thrived in practice. Certification which
was most present in the literature review was similarly well accepted among the
practitioners as a convenient assurance technique. Testing in terms of applica-
tion security was also present in both, literature and interviews. However, it is
striking but not surprising that neither monitoring nor auditing, which offered
many cloud-specific frameworks in literature, were present among the respon-
dents. Contractual agreements could be compared to security SLAs with regard
to how they work, except that there are no actual metrics agreed upon but rules.
6 Conclusion
Previous research identified security as a requirement considered by CP cus-
tomers. Our sample indicates that security may not always be a selection crite-
rion and neither the most decisive one. If considered in the CP selection, then
mostly in terms of availability and for the sake of compliance. Especially the
focus on compliance it not surprising as it has been observed in other sectors as
well [16,54]. Nevertheless, it is certainly a requirement companies have, which
manifests itself in cloud use. This is indicated by retrospective analysis and con-
siderations of multiple providers.
CP Selection Process: In our sample we could rarely find any elaborated
process of eliciting requirements and then coming to a rational decision which
CP to select. Instead, CP were chosen based on vouchers, by chance (just pick
on CP for ’testing’, but then stick with it), by the management because of
established relationships, or because of previous experience from a developer.
Even more, some companies make use of many CPs in an unstructured way,
302 S. Pape and J. Stankovic
e.g. each department decides by its own. Another pattern we could identify was
that companies often try to ’first get into the cloud’ and then optimise costs and
sometimes security (lift and shift) or try to sort out the collection of different
CPs. Further research would be desired to investigate why the methodology
proposed by research seems to be rarely used in practise.
For that purpose the different roles in the requirements/decision making
process should be investigated in detail and elaborated at which step the relevant
methodologies from research were not considered and why.
Assurance: The respondents reported on using more than one assurance tech-
nique, combined models from the literature were not present at all. Addition-
ally, they saw flaws in the existing assurance techniques and may not even be
acquainted with possible cloud-specific assurance. Thus, the noteworthy find-
ing of this comparison is a divergence between the assurance methods adopted
in practice and the cloud-specific ones proposed in literature. It can be specu-
lated whether some academic approaches to assurance have never exceeded their
theoretical approach or if they were not able to gain ground in practice yet.
Company Size: Although the results uncover many dimensions and patterns
of cloud security, they are not complete. As mentioned earlier, no saturation
of interviews could be reached among small and unregulated companies. In
contrast, large regulated companies were well represented and most likely con-
tributed to a strong focus on compliance in this analysis. Future work could
examine on a larger scale whether and how companies have incorporated secu-
rity into their provider selection and in particular investigate commonalities and
differences between smaller and larger companies.
Big CPs vs. Localisation: It seems that the big CPs are in general trusted
by the companies and the idea of a German cloud failed. Companies are trying
to setup a compliant way to work with the big CPs. However, one interviewee
was concerned about industrial espionage and strongly voted for a European
or German CP with all components made in the EU. Further research should
unfold the different dimensions of trust, and also investigate to which extent
regulations or agreements as the EU–US Privacy Shield influence it.
Gaps Between Research and Practise: In the requirement elicitation and
decision making process and in the use of assurance technologies there seems to
be a gap between research and practise. This gap is something which seems to
be quite common in a lot of areas [52]. Further work should investigate whether
this is just a typical finding and already existing ideas can be applied to bridge
it [20] or if it is a context specific problem and new ideas are needed.
References
1. Akerlof, G.A.: The market for “lemons”: quality uncertainty and the market mech-
anism. In: Uncertainty in Economics, pp. 235–251. Elsevier (1978)
2. Alhenaki, L., Alwatban, A., Alahmri, B., Alarifi, N.: Security in cloud computing:
a survey. Int. J. Comput. Sci. Inf. Secur. (IJCSIS) 17(4), 67–90 (2019)
Decisive Factors in Cloud Provider Selection with a Focus on Security 303
3. Anisetti, M., Ardagna, C.A., Damiani, E.: A certification-based trust model for
autonomic cloud computing systems. In: 2014 International Conference on Cloud
and Autonomic Computing, pp. 212–219 (2014)
4. Anisetti, M., Ardagna, C.A., Damiani, E.: A test-based incremental security certi-
fication scheme for cloud-based systems. In: 2015 IEEE International Conference
on Services Computing, pp. 736–741 (2015)
5. Anisetti, M., Ardagna, C.A., Damiani, E., Gaudenzi, F., Veca, R.: Toward secu-
rity and performance certification of open stack. In: 2015 IEEE 8th International
Conference on Cloud Computing, pp. 564–571 (2015)
6. Anisetti, M., Ardagna, C.A., Gaudenzi, F., Damiani, E.: A certification framework
for cloud-based services. In: Proceedings of the 31st Annual ACM Symposium on
Applied Computing, SAC 2016, pp. 440–447. ACM (2016)
7. Ardagna, C.A., Asal, R., Damiani, E., Vu, Q.H.: From security to assurance in the
cloud: a survey. ACM Comput. Surv. 48(1), 2:1–2:50 (2015)
8. Ba, H., Zhou, H., Bai, S., Ren, J., Wang, Z., Ci, L.: jMonAtt: integrity monitoring
and attestation of JVM-based applications in cloud computing. In: ICISCE, pp.
419–423 (2017)
9. Bleikertz, S., Mastelic, T., Pape, S., Pieters, W., Dimkov, T.: Defining the cloud
battlefield - supporting security assessments by cloud customers. In: IC2E, pp.
78–87 (2013)
10. Briggs, B., Lamar, K., Kark, K., Shaikh, A.: Manifesting legacy: looking beyond
the digital era. Technical report, 2018 Global CIO Survey, Deloitte (2018)
11. Casola, V., Benedictis, A.D., Rak, M., Villano, U.: SLA-based secure cloud appli-
cation development: the SPECS framework. In: SYNASC, pp. 337–344 (2015)
12. CSA: Top threats to cloud computing v1.0. Technical report, Cloud Security
Alliance (2010). https://cloudsecurityalliance.org/topthreats/csathreats.v1.0.pdf
13. CSA: The notorious nine: cloud computing top threats in 2013. Techni-
cal report, Cloud Security Alliance (2013). https://cloudsecurityalliance.org/
download/artifacts/the-notorious-nine-cloud-computing-top-threats-in-2013/
14. CSA: The treacherous 12 - cloud computing top threats in 2016. Technical report,
Cloud Security Alliance (2016). https://downloads.cloudsecurityalliance.org/
assets/research/top-threats/Treacherous-12 Cloud-Computing Top-Threats.pdf
15. CSA: Top threats to cloud computing the egregious 11. Technical report, Cloud
Security Alliance (2019). https://cloudsecurityalliance.org/download/artifacts/
top-threats-to-cloud-computing-egregious-eleven/
16. Dax, J., et al.: IT security status of German energy providers (2017). https://arxiv.
org/abs/1709.01254
17. Deng, L., Liu, P., Xu, J., Chen, P., Zeng, Q.: Dancing with wolves: towards
practical event-driven VMM monitoring. In: Proceedings of the 13th ACM SIG-
PLAN/SIGOPS International Conference on VEE, pp. 83–96. ACM (2017)
18. Di Giulio, C., Kamhoua, C., Campbell, R.H., Sprabery, R., Kwiat, K., Bashir,
M.N.: IT security and privacy standards in comparison: improving FedRAMP
authorization for cloud service providers. In: CCGrid, pp. 1090–1099 (2017)
19. Di Giulio, C., Sprabery, R., Kamhoua, C., Kwiat, K., Campbell, R.H., Bashir,
M.N.: Cloud standards in comparison: are new security frameworks improving
cloud security? In: CLOUD, pp. 50–57 (2017)
20. Ferguson, J.: Bridging the gap between research and practice. Knowl. Manag. Dev.
J. 1(3), 46–54 (2005)
21. Fernando, R., Ranchal, R., Bhargava, B., Angin, P.: A monitoring approach for
policy enforcement in cloud services. In: CLOUD, pp. 600–607 (2017)
304 S. Pape and J. Stankovic
22. Ghutugade, K.B., Patil, G.A.: Privacy preserving auditing for shared data in cloud.
In: CAST, pp. 300–305 (2016)
23. Gupta, P., Seetharaman, A., Raj, J.R.: The usage and adoption of cloud computing
by small and medium businesses. Int. J. Inf. Manag. 33(5), 861–874 (2013)
24. Haeberlen, T., Dupré, L.: Cloud computing - benefits, risks and recommendations
for information security. Technical report, ENISA (2012)
25. Henze, M., et al.: Practical data compliance for cloud storage. In: 2017 IEEE
International Conference on Cloud Engineering (IC2E), pp. 252–258 (2017)
26. Hetzenecker, J., Kammerer, S., Amberg, M., Zeiler, V.: Anforderungen an cloud
computing Anbieter. In: MKWI (2012)
27. Ismail, U.M., Islam, S., Islam, S.: Towards cloud security monitoring: a case study.
In: Cybersecurity and Cyberforensics Conference (CCC), pp. 8–14 (2016)
28. Jakhotia, K., Bhosale, R., Lingam, C.: Novel architecture for enabling proof of
retrievability using AES algorithm. In: ICCMC, pp. 388–393 (2017)
29. Jansen, W., Grance, T.: SP 800-144. Guidelines on security and privacy in public
cloud computing. Technical report, NIST (2011)
30. Jiang, T., Chen, X., Ma, J.: Public integrity auditing for shared dynamic cloud
data with group user revocation. IEEE Trans. Comput. 65(8), 2363–2373 (2016)
31. Kaaniche, N., Mohamed, M., Laurent, M., Ludwig, H.: Security SLA based moni-
toring in clouds. In: IEEE EDGE, pp. 90–97 (2017)
32. Kanstrén, T., Lehtonen, S., Savola, R., Kukkohovi, H., Hätönen, K.: Architecture
for high confidence cloud security monitoring. In: IC2E, pp. 195–200 (2015)
33. Katopodis, S., Spanoudakis, G., Mahbub, K.: Towards hybrid cloud service cer-
tification models. In: IEEE International Conference on Services Computing, pp.
394–399 (2014)
34. Krotsiani, M., Spanoudakis, G.: Continuous certification of non-repudiation in
cloud storage services. In: 2014 IEEE 13th International Conference on Trust,
Security and Privacy in Computing and Communications, pp. 921–928 (2014)
35. Krutz, R.L., Vines, R.D.: Cloud Security: A Comprehensive Guide to Secure Cloud
Computing. Wiley, Hoboken (2010)
36. Kuckartz, U.: Qualitative Inhaltsanalyse: Methoden, Praxis, Computerun-
terstützung. Beltz Juventa (2016)
37. Kumar, R., Goyal, R.: On cloud security requirements, threats, vulnerabilities and
countermeasures: a survey. Comput. Sci. Rev. 33, 1–48 (2019)
38. Lacity, M.C., Reynolds, P.: Cloud services practices for small and medium-sized
enterprises. MIS Q. Exec. 13(1), 31–44 (2014)
39. Lang, M., Wiesche, M., Krcmar, H.: What are the most important criteria for
cloud service provider selection? A Delphi study. In: ECIS (2016)
40. Lee, C., Kavi, K.M., Paul, R.A., Gomathisankaran, M.: Ontology of secure service
level agreement. In: 2015 IEEE 16th International Symposium on High Assurance
Systems Engineering, pp. 166–172 (2015)
41. Lins, S., Grochol, P., Schneider, S., Sunyaev, A.: Dynamic certification of cloud
services: trust, but verify!. IEEE Secur. Priv. 14(2), 66–71 (2016)
42. Lins, S., Schneider, S., Sunyaev, A.: Trust is good, control is better: creating secure
clouds by continuous auditing. IEEE Trans. Cloud Comput. 6(3), 890–903 (2018)
43. Lins, S., Thiebes, S., Schneider, S., Sunyaev, A.: What is really going on at your
cloud service provider? Creating trustworthy certifications by continuous auditing.
In: 48th HICSS, pp. 5352–5361 (2015)
44. Luna, J., Suri, N., Iorga, M., Karmel, A.: Leveraging the potential of cloud secu-
rity service-level agreements through standards. IEEE Cloud Comput. 2(3), 32–40
(2015)
Decisive Factors in Cloud Provider Selection with a Focus on Security 305
45. Ma, M., Weber, J., van den Berg, J.: Secure public-auditing cloud storage enabling
data dynamics in the standard model. In: DIPDMWC, pp. 170–175 (2016)
46. Mahesh, A., Suresh, N., Gupta, M., Sharman, R.: Cloud risk resilience: investi-
gation of audit practices and technology advances-a technical report. Int. J. Risk
Conting. Manag. (IJRCM) 8(2), 66–92 (2019)
47. Majumdar, S., Madi, T., Wang, Y., Jarraya, Y., Pourzandi, M., Wang, L., Debbabi,
M.: User-level runtime security auditing for the cloud. IEEE Trans. Inf. Forensics
Secur. 13(5), 1185–1199 (2018)
48. Meera, G., Geethakumari, G.: A provenance auditing framework for cloud com-
puting systems. In: SPICES, pp. 1–5 (2015)
49. Mohammed, M.M.Z.E., Pathan, A.K.: International center for monitoring cloud
computing providers (ICMCCP) for ensuring trusted clouds. In: IEEE 11th Inter-
national Conference on Ubiquitous Intelligence and Its Associated Workshops, pp.
571–576 (2014)
50. More, S.S., Chaudhari, S.S.: Secure and efficient public auditing scheme for cloud
storage. In: CAST, pp. 439–444 (2016)
51. Munoz, A., Mafia, A.: Software and hardware certification techniques in a combined
certification model. In: SECRYPT, pp. 1–6 (2014)
52. Norman, D.A.: The research-practice gap: the need for translational developers.
Interactions 17(4), 9–12 (2010)
53. Nugraha, Y., Martin, A.: Towards the classification of confidentiality capabilities
in trustworthy service level agreements. In: IC2E, pp. 304–310 (2017)
54. Pape, S., Pipek, V., Rannenberg, K., Schmitz, C., Sekulla, A., Terhaag, F.: Stand
zur IT-Sicherheit deutscher Stromnetzbetreiber (2018). http://dokumentix.ub.uni-
siegen.de/opus/volltexte/2018/1394/
55. Parasuraman, K., Srinivasababu, P., Angelin, S.R., Devi, T.A.M.: Secured docu-
ment management through a third party auditor scheme in cloud computing. In:
ICECCE, pp. 109–118 (2014)
56. Pasquier, T.F.J., Singh, J., Bacon, J., Eyers, D.: Information flow audit for PaaS
clouds. In: IEEE IC2E, pp. 42–51 (2016)
57. Polash, F., Shiva, S.: Building trust in cloud: service certification challenges and
approaches. In: 9th International Conference on Complex, Intelligent, and Software
Intensive Systems, pp. 187–191 (2015)
58. Ramokapane, K.M., Rashid, A., Such, J.M.: Assured deletion in the cloud: require-
ments, challenges and future directions. In: CCSW, pp. 97–108. ACM (2016)
59. Rashmi, R.P., Sangve, S.M.: Public auditing system: improved remote data pos-
session checking protocol for secure cloud storage. In: iCATccT, pp. 75–80 (2015)
60. Repschläger, J., Wind, S., Zarnekow, R., Turowski, K.: Developing a cloud provider
selection model. In: EMISA (2011)
61. Rewadkar, D.N., Ghatage, S.Y.: Cloud storage system enabling secure privacy
preserving third party audit. In: ICCICCT, pp. 695–699 (2014)
62. Rios, E., Mallouli, W., Rak, M., Casola, V., Ortiz, A.M.: SLA-driven monitoring
of multi-cloud application components using the MUSA framework. In: IEEE 36th
ICDCSW, pp. 55–60 (2016)
63. Rizvi, S.S., Bolish, T.A., Pfeffer III, J.R.: Security evaluation of cloud service
providers using third party auditors. In: Second International Conference on Inter-
net of Things, Data and Cloud Computing, pp. 106:1–106:6 (2017)
64. Ryoo, J., Rizvi, S., Aiken, W., Kissell, J.: Cloud security auditing: challenges and
emerging approaches. IEEE Secur. Priv. 12(6), 68–74 (2014)
306 S. Pape and J. Stankovic
65. Schneider, S., Lansing, J., Gao, F., Sunyaev, A.: A taxonomic perspective on certi-
fication schemes: development of a taxonomy for cloud service certification criteria.
In: HICSS, pp. 4998–5007 (2014)
66. Sen, A., Madria, S.: Data analysis of cloud security alliance’s security, trust &
assurance registry. In: ICDCN, pp. 42:1–42:10. ACM (2018)
67. Sotiriadis, S., Lehmets, A., Petrakis, E.G.M., Bessis, N.: Unit and integration test-
ing of modular cloud services. In: AINA, pp. 1116–1123 (2017)
68. Stephanow, P., Khajehmoogahi, K.: Towards continuous security certification of
software-as-a-service applications using web application testing techniques. In:
AINA, pp. 931–938 (2017)
69. Thendral, G., Valliyammai, C.: Dynamic auditing and updating services in cloud
storage. In: International Conference on Recent Trends in Information Technology,
pp. 1–6 (2014)
70. Tung, Y., Lin, C., Shan, H.: Test as a service: a framework for web security TaaS
service in cloud environment. In: 2014 IEEE 8th International Symposium on Ser-
vice Oriented System Engineering, pp. 212–217 (2014)
71. Zhang, H., Manzoor, S., Suri, N.: Monitoring path discovery for supporting indirect
monitoring of cloud services. In: IEEE IC2E, pp. 274–277 (2018)
72. Zhang, H., Trapero, R., Luna, J., Suri, N.: deQAM: a dependency based indirect
monitoring approach for cloud services. In: IEEE SCC, pp. 27–34 (2017)
Discrete Event Simulation of Jail Operations
in Pursuit of Organizational Culture Change
Abstract. Justice facilities such as jails are complex adaptive systems. They are
people-driven, whether by the organizational culture of those that operate them,
or by inmate culture. The development of organizational culture is organic and
based on buy-in at all levels, or the lack thereof. Organizational culture evolves,
including in response to attempted interventions from within or without.
Physical and electronic security’s relationship to detention operations and
detainee supervision involves similar dynamics.
In this paper, we explore jail operations. We model human use of a housing
unit and associated support spaces via discrete-event simulation. We simulate
this system to understand the capacity and limits on human use of building
spaces. We explore how this sociotechnical system responds when stressed. We
thereby validate the design within limits that correspond to planned operational
capacity. The goal of the research is to design spaces and environments that
support improved outcomes via improvements in organizational culture.
We introduce the specifics of jail operations via this model while exploring
the full range of applications for this type of simulation in the built environment.
1 Introduction
classifications are well known. Jail detainees’ security risk and levels of needs are less
well known. Prison stays are lengthy. Jail stays are short. For these reasons, and others,
jails are more difficult to design and operate than prisons.
simulation advances forward in time. Eventually, the sallyport becomes available and
the detainee leaves the queue, seizes the sallyport and continues moving through the
process. Meanwhile, the simulation tracks ‘time in system,’ resource utilization, and
many more performance metrics. Simulation is a technique for the explicit discovery
and analysis of process bottlenecks. It incorporates real-world situations, such as
variations in activity times, resource constraints and downtimes. Detainee goal directed
behavior is defined algorithmically. The simulation establishes parameters for move-
ment, tracks the state of spaces (number of occupants) and controlled openings (open or
closed), while defining schedules and operational norms.
2 Conceptual Design
2.1 Housing Unit Layout (See Appendix A, Figs. 3 and 4)
The housing unit has a capacity of 60 beds, which the designer arrayed in three tiers.
Each tier hosts 20 cells (ten on a side). Between the cells are the dayroom and the
adjacent indoor recreation area. The size of the dayroom is slightly larger than the
60 35 square feet per detainee required by the American Correctional Association
Local Adult Detention Facility Standards. The size of indoor recreation is based on
accommodating half-court basketball, providing enough space beyond the three-point
line for it to function for pickup games or shooting baskets. The layout barely
accommodates this use. Half-court basketball is a superior form of recreation because it
involves vigorous large muscle exercise, which significantly improves brain chemistry
and lowers the propensity for aggressiveness and violence, compared to lower intensity
or passive recreation.
2.2 Glazed End Wall (See Appendix A, Figs. 10, 11, 12 and 13)
The end of the dayroom/indoor recreation footprint is a glazed wall. The triple tier
design has the benefit of allowing natural light to penetrate deeply into the building
interior. This has both health and energy conservation benefits. Access to natural light
is important for human health, whether related to the production of Vitamin D or the
regulation of circadian rhythms, and has the added benefit of limiting the use of
artificial light, which both uses electricity and produces heat which must be removed by
air conditioning systems which also use electricity. The feeling of openness and the
expansive views provided by the glazed wall are psychologically relevant, as they
mitigate the institutional nature of jail.
310 H. D. Lester and M. J. Miller
3 A Different Approach
3.1 Detainee Movement (See Appendix A, Figs. 6, 7 and 8)
The point that detainee movement is inherently safe when limited to a single detainee
at a time is a critical driver of both the design—and the desired outcomes—of this
concept. Conventional correctional design involves escorted movement, meaning that a
correctional officer escorts a detainee or detainees from point A to point B [5]. On
Rikers Island, this means two things: One, the detainee is placed in restraints (handcuffs
at a minimum, but often additional restraints; typical when groups are escorted) which
is both time consuming (increasing staffing levels and cost) and something that rein-
forces detainee ‘dangerousness’ and (feelings of) confinement. Two, the opportunity
for violence exists during escorted movement. Many use-of-force incidents in the NYC
jail system are associated with escorted movement. Furthermore, use-of-force incidents
and other security related issues result in housing units being locked down, meaning
that detainees are confined to their cells, preventing beneficial use of day space,
movement to programs or visitation, and severely impacting operations [6]. Any design
that eliminates routine escorted movement, restricts detainee transitions through spaces
to a single detainee at a time (unescorted movement), and ensures that detainees come
in contact with other detainees, non-custodial staff or volunteers, visitors, or custodial
staff only in direct supervision settings—will increase safety and reduce levels of stress
experienced by everyone. Reduced stress is a foundation of organizational culture
change.
Triggers of Aggression. Bad news, on the other hand, increases stress, and resulting
aggression. Most bad news arrives during visitation (personal or professional,) and
especially during phone calls. For this reason, inmate phones should be placed prox-
imate to staff stations, so custody staff can better monitor body language and behaviors
that result from such calls. Early intervention on the staff member’s part can break the
cycle that otherwise begins with the bad news received, continues with transferred
aggression, and ends with a violent outburst or an assault.
A Call for Action. Jail is an extreme situation that places detainees and staff under
stress. Extreme situations demand extreme solutions, even in the no-beta-testing culture
of detention and corrections. Under such circumstances, drivers that otherwise might
not be salient are revealed, allowing design and operational mitigation strategies to
manage inmate behavior [13]. The contribution of psychological drivers to the suc-
cessful outcomes demonstrated by direct supervision [14] should eventually provide
practitioners with evidence-based design guidelines that balance facility design and
operational integration for best outcomes at the lowest capital and operational cost.
delayed getting to the dayroom for their meals. The last arrivals were about 30 min past
the hour. One possible mitigation strategy is staggering meal start times. This could
take the form of even numbered cells (on all tiers) starting meals at the top of the hour,
while odd numbered cells start meals at half past the hour. In-cell personal hygiene
would occur in the non-meal portion of the three hours per day designated for these
purposes.
the opposite direction. One way to visualize it is to think about the circulatory system in
the human body. It never reverses direction because it is a loop (also because its
‘secured openings’ include valves that only allow one-way movement). If all detainee
movement could occur in a closed loop, conflicts would only be temporal, not spatial.
However, the redundancy and resilience of the human circulatory system is limited,
thus myocardial infarctions and strokes.
Wait time – Aggregate wait time at controlled openings (elapsed time from
presentation of credentials to lock cycling) – at the mean of the
distribution and at the 95% percentile (two deviations from the
mean).
% Compliance – How many detainees in the simulation succeed in receiving five
hours per day of education, vocational, and therapeutic program-
ming, a goal of the City of New York [15].
% Utilization – The amount of time a space is utilized during the hours it can be
utilized, which varies based on the space. This metric helps uncover
potential overutilization in the form of bottlenecks or instances of
underutilization, where multi-use spaces might be considered.
Additional KPIs will be considered and some subset of these will be included as
simulation-generated data based on how informative they might be, especially over
large numbers of simulation runs.
to group actions like riots, to acts of God like an airplane crashing into the building,
causing a multi-alarm fire. Basically, we intend to start small and study how the KPIs
respond when we “throw a wrench into the system” so we can understand the levels of
resilience and redundancy extant in the design and operations as currently conceived.
5 Conclusions
The City of New York is dramatically reforming its justice system [16]. The new policy
is to minimize the use of detention and close Rikers Island [17]. Towards that end, four
new high-rise jails are proposed. A central concern is the high probability of trans-
ference of NYC DOC’s dysfunctional organizational culture to these new facilities. To
address this issue, policy, staffing, and administrative measures have been proposed,
and the best minds [7] are actively engaged with the issue. This paper outlines an
approach which explores the design of mini-jails and housing units that actively mit-
igate or eliminate the possibility of negative interactions between all actors, limiting
human interaction to spaces under direct supervision. To support such design and
operational integration, discrete event simulation is employed to refine the design, to
fine tune the scheduling, to facilitate operations in the form of unescorted movement,
and to pseudo-validate the modeling and simulation so employed. Although the work
continues, the results to date suggest that architectural design, systems integration,
operational integration, and discrete event simulation and modeling all have a role to
play in pursuit of culture change and positive outcomes.
Appendix A
References
1. Carter, S.: Career extensions: the footprints of David Parrish. Correct. News 14(7), 19–21
(2008)
2. Farbstein, J., Wener, R.: A Comparison of “Direct” and “Indirect” Supervision Correctional
Facilities. National Institute of Corrections, Washington, DC (1989)
3. Tartaro, C., Levy, M.: Density, inmate assaults, and direct supervision jails. Crim. Justice
Policy Rev. 18(4), 395–417 (2007)
4. Lester, H., McKay, G., Lester, E.: Sociotechnical systems for high rise detention. In: 26th
ISTE International Conference on Transdisciplinary Engineering (TE 2019) Conference
Proceedings. IOS Press, Amsterdam (2019, Forthcoming)
5. Ricci, K., Greene, F., Genevro, R.: What Jails Can’t Do. Urban Omnibus, New York (2017)
6. New York City Board of Correction: Annual Lockdown Report. Board of Correction, New
York (2019)
7. Jacobson, M., DeWolf, E., Egan, M., Hafetz, D.: Beyond the Island: changing the culture of
New York City jails. Fordham Urban Law J. 45(2), 373–436 (2018)
8. Drapkin, M.: Developing Policies and Procedures for Jails. United Book Press, Baltimore
(1996)
9. American Jail Association: Resolution: Direct Supervision Jails [Principles of Direct
Supervision]. Adopted by the Board of Directors on November 14, 1992. Re-affirmed on
May 3, 2008, by the AJA Board of Directors in Sacramento, CA. Amended on October 17,
2013, by the AJA Board of Directors in St. Louis, MO. https://www.americanjail.org/files/
About%20PDF/_AJA%20Resolutions%20-%20January%202017.pdf. Accessed 12 Aug
2019
10. Kerle, K.: Exploring Jail Operations. American Correctional Association, Hagerstown
(2003)
11. Hall, E.: The Hidden Dimension. Doubleday, Garden City (1966)
12. Craft, L., Perna, F.: The benefits of exercise for the clinically depressed. Prim. Care
Companion J. Clin. Psychiatry 6(3), 104–111 (2004)
13. Hutchinson, V.: Managing inmate behavior in jails. Correct. Today 67(5), 28–31 (2005)
14. Parrish, D.: The evolution of direct supervision in the design and operation of jails. Correct.
Today 62(6), 84–88 (2000)
15. Mayor’s Office of Criminal Justice: Justice Implementation Task Force Culture Change
Working Group Meeting #1. Mayor’s Office of Criminal Justice, New York (2017)
16. Mayor’s Office of Criminal Justice: Breaking the Frame? Remaking the Criminal Justice
System in New York City. Mayor’s Office of Criminal Justice, New York (2019)
17. Glazer, E.: Progress on Closing Rikers Island. New York City Council’s Committee on Fire
and Criminal Justice Services, New York (2017)
ADIoT Workshop
A Basic Theory of Lightweight
Hierarchical Key Predistribution Scheme
1 Introduction
Internet of Things (IoT) is an emerging technology that has been recently incor-
porated into our life to make it more comfortable. Resource constraint (low
power, computation power, memory etc) devices such as sensors, systems-on-
chips, microcontrollers are basic devices low cost IOT. These devices need to
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 325–340, 2020.
https://doi.org/10.1007/978-3-030-42048-2_21
326 D. K. Dalai
1.1 Motivation
Constraints in resources (i.e., energy, storage and computation) of IoT devices,
like sensors of WSN, forbids implementations of public key cryptosystems for
secure communication in them. Therefore, relatively inexpensive symmetric key
cryptosystems (SKC) are adapted in these low cost networks. As the SKC require
the end parties to possess the same encryption-decryption key(s), the scheme
requires the implementation of an adequate key predistribution scheme (KPS).
Several existing approaches to KPS are available [10,12–16,19,24] in literature.
See Sect. 2.1 for a brief survey of some prominent KPS and their limitations in
adapting to low cost hierarchical setups.
The situation becomes more critical when the number of nodes increases and
the network of the organization is already in hierarchical in nature; forcing us
to implement a hierarchy via labeling of special nodes that decentralizes the
network. Practical instances of such networks are ad-hoc networks that have
volatile inter nodal communications or an internal network of a multinational
company which has its offices all across the world. More critical applications
include military networks or health care or automated processes that must always
be ready for an emergency.
Such networks may be well served by a protocol that ensures pair-wise shared
key(s) for a node at same as well as different levels. Of course, due to the large
(and growing) size of the network and limited storage in each node, subsets of
users must share some common key(s). For example, for a military network,
during a normal course of action, regiment heads are responsible for their team
member’s (child’s) performance. Being fewer, they may well be given extra phys-
ical protection and be allowed to listen in all their children’s communications.
All existing works [5,13,17,18] are designed to support such systems.
Time of emergency may demand children of various subdivision to securely
collaborate among themselves. Involvement of respective parents (or even worse,
A Basic Theory of Lightweight Hierarchical Key Predistribution Scheme 327
the root central authority) in such interactions may not be feasible due to prac-
tical constraints like distance and traffic overload. These practical instances are
not captured by existing works [3–6,8,12,13,15–18,21,22] and motivate our hier-
archical key predistribution proposal. In fact solutions with hierarchy provided
by state-of-the-art protocols like [13,17,21] relies on random graph theory [11],
identity-based key agreement protocol of Sakai et al. [20] and PBIBD [19] respec-
tively. These have some issues of practicality which are briefly discussed in
Sect. 2.1.
The primary contribution in this paper is a proposal for an efficient and secure
hierarchical KPS (HKPS). To extend the idea of designing deterministic KPS
from the combinatorial design (see Sect. 2.4), we introduced the concept of com-
binatorial hierarchical design for building an HKPS. This idea of mapping can
be a basic theory to design several HKPS. Hence, we propose a simple and ele-
gant product based HKPS next. This HKPS is built from a set of basic KPS,
where a KPS is chosen at each level of the hierarchy. As a result, the designer
has the freedom to choose an appropriate KPS as per the requirements (e.g.,
number of nodes, size of the key ring, efficiency, resiliency, connectivity) at that
level. Further, to ensure the security of parent nodes after a capture of several
child nodes, we proposed the hash chain technique. This hash chain technique
ensures resilience against compromise of any number of lower level nodes. Such
a hash chain based HKPS is denoted as HC-HKPS. Further, we have discussed
an instant of such HKPS by building over a very efficient KPS Sensornet. Our
contribution is theoretical and starting point of such a design. Therefore, the
paper does not contain any experimental and comparative results.
1.3 Organization
We have already outlined our research objectives. Rest of the paper is orga-
nized as follows. Important preliminary requirements for this work are drafted
in Sect. 2. A brief survey of KPS and HKPS are presented in Sects. 2.1 and 2.2
respectively. Two important metrics resilience (Rt ) and connectivity (C)for the
evaluation of a KPS is defined in Sect. 2.3. Further, Sect. 2.4 presents the defini-
tion of useful combinatorial designs and the mapping to build a KPS from it.
The main contribution is presented in Sect. 3. In this section, the hierarchical
set design is introduced which can be mapped for the construction of an HKPS.
Further, an HKPS is designed from the proposed product based hierarchical
design. To improve the resilience at the parent level, a hash chain technique is
implemented to build a secure HKPS. Section 4 investigates instantiation of the
proposed HKPS to a class of a very efficient KPS, Sensornet. Section 5 concludes
the paper and states relevant research directions.
328 D. K. Dalai
2 Preliminary
2.1 Key Predistribution Scheme: An Overview
An approach for authentication in open network [22] assume that trust on nodes
is rather risky as nodes are vulnerable to several attacks. Further, mutual sharing
of keys is impractical for the large networks (with hierarchy) due to storage over-
heads. These considerations led to the evolution of key predistribution schemes
(KPS).
Ideally, any KPS should have small key rings, yet support large networks
with appreciable resilience and connectivity. However, the authors in [15,16]
indicate the impossibility of constructing such a ‘perfect KPS’ that meets all
these criteria. This motivates proposals of several designs that are robust for the
specific purpose(s). Next, we recall the steps executed by any KPS.
– Offline generation and preallocation of keys: A large (ν) collection of keys
(key pool := K; |K| = ν) and their identifiers (ids) are generated offline.
Equal sized subsets of keys (rank := z) are preallocated into each sensor node
at the time of their joining the network. Each key gets shared by the same
number of nodes (degree := r). Fixed value of z and r may ensure equal load
(desirable).
– Key establishment: This process is one of two cases, as described below:
– (i) Shared key discovery phase establishes shared key(s) among the par-
ticipants. Individual users broadcast all their key ids or, node id. From
each other’s ids, each sensor matches them or, follows a computation to
trace their mutual shared key id(s), hence common key(s).
– (ii) Path key establishment: establishes an optimized path key between
a given pair of nodes that do not share any common key and involves
intermediate nodes.
Depending on whether the above processes are probabilistic or deterministic,
such schemes are broadly classified into two types: (a) random and (b) deter-
ministic. We present below a brief overview of individual type of schemes.
other nodes establishing multiple links. This way of communication becomes vul-
nerable to information leak and attack, slows down the performance, consumes
more energy etc. Hence a good KPS must have high connectivity. The metric
connectivity (C) in a KPS is defined as the probability that two nodes can estab-
number of possible links in the network
lish a link in the network i.e., C = .
number of pair of nodes in the network
We name the new design as a hierarchical set system (or, hierarchical design) of
depth l and denote as (X , B 1 , B 2 , · · · , B l ). The blocks in B i are too called the
blocks at depth i.
With the above condition, B i can have a disjoint partition such that every
block in a partition of B i is a subset of a block in B i−1 . If a block B ∈ B i is a
subset of more than one blocks in B i−1 , we put the block B in exactly one of the
partitions. Hence, the subset relation in (X , B 1 , B 2 , · · · , B l ) can be viewed as a
tree structure where the elements in a partition of B i are children of an element
in B i−1 . The superscripted index i in B i refers to the depth/level of hierarchy in
the design or the height of the element in the tree. The set of blocks at depth i
is B i and the number of blocks in depth i is |B i |. The following is an example of
a hierarchical design of depth 2.
Example 1. The set system (X , B 1 , B 2 ), where
X = {1, 2, 3, 4, 5, 6}, B1 = {{1, 2, 3, 4}, {3, 4, 5, 6}} and
B2 = {{1, 2, 3}, {1, 2, 4}, {1, 3, 4}, {2, 3, 4}, {3, 4, 5}, {3, 4, 6}, {3, 5, 6}, {4, 5, 6}},
is a hierarchical design of depth 2. Figure 1 presents the tree structure of the
hierarchical design.
1, 2, 3, 4, 5, 6
1, 2, 3, 4 3, 4, 5, 6
1, 2, 3 1, 2, 4 1, 3, 4 2, 3, 4 3, 4, 5 3, 4, 6 3, 5, 6 4, 5, 6
1. the v varieties of X to the set of keys in the scheme (i.e., key pool K),
2. the blocks in B i to the key ring of the nodes at depth i in HKPS.
The elements of X mapped as the key identifier of the keys in K and the nodes
are assigned the key identifiers in the mapped block and the corresponding keys
of the identifiers. A KPS from a hierarchical design S = (X , B 1 , B 2 , · · · , B l ) is
called as HKPS S. This basic theory of building an HKPS by the mapping from
a hierarchical design tempts to design several well suited hierarchical designs.
We propose such a hierarchical design and equivalent HKPS in next.
a1, a2, a3, a4, b1, b2, b3, b4, c1, c2, c3, c4
B1 B2 B3
a1, a2, a3, a4, b1, b2, b3, b4 b1, b2, b3, b4, c1, c2, c3, c4 a1, a2, a3, a4, c1, c2, c3, c4
B1,1 B1,2 B1,3 B1,4 B2,1 B2,2 B2,3 B2,4 B3,1 B3,2 B3,3 B3,4
a1, a2, a3, a1, a2, a4, a1, a3, a4, a2, a3, a4, b1, b2, b3, b1, b2, b4, b2, b3, b4, a1, a2, a3, a1, a2, a4, a1, a3, a4, a2, a3, a4,
b1, b3, b4,
b1, b2, b3 b1, b2, b4 b1, b3, b4 b2, b3, b4 c1, c2, c3 c1, c2, c4 c1, c3, c4 c2, c3, c4 c1, c2, c3 c1, c2, c4 c1, c3, c4 c2, c3, c4
bd }.
– bd , z d and rd are defined in Theorem 1. Note that, the superscript ‘d’ rep-
resents the corresponding values at depth d and not to be confused with the
exponential power.
d
– αij = |Adi ∩ Adj | denotes the number of common elements between the blocks
Adi and Adj in the set design Sd .
at depth d. The link L can fail if and only if all d links lw , 1 ≤ w ≤ d between
the nodes Awiw and Awjw in the KPS Xw fail. On capture of t nodes at depth
d, the probability of the fail of the link lw is Fw (t) = 1 − Rw (t) for each
w(1 ≤ w ≤ d). Hence the fail probability of the link L on capture of t nodes
d
is Rd (t) = 1 − w=1 (1 − Rw (t)).
6. Similar to the proof of Item 5., there is a link L between two nodes
at depth d if and only if there are links lw , 1 ≤ w ≤ d between the nodes Awiw
and Awjw in the KPS Xw . Hence, the connectivity of the network at depth d
d
is C d = w=1 Cp .
One can verify the above theorem with Example 2. Now we summarize the effects
of Theorem 1 on the HKPS designed from Construction 1.
A Basic Theory of Lightweight Hierarchical Key Predistribution Scheme 335
1. Theorem 1.1 provides the maximum number of nodes can be allocated in each
depth. It can be observed that the number of blocks in depth d increased by
bd times than the number of blocks in depth d − 1. As we need a larger
number blocks in higher depth, we can choose the designs of a higher number
of blocks.
2. Theorem 1.2 provides the maximum size of key ring of nodes at each depth.
It can be observed that the size of key ring in depth d − 1 is vzdd times than the
size of key ring in depth i. To minimize the storage space, we need to choose
smaller vd or larger zd at lower depth nodes than the higher depth nodes. This
can be achieved as in our network design goal, we have a smaller number of
nodes (but more powerful) at lower depth than the number of nodes at higher
depth.
3. Theorem 1.3 presents the degree of a key (i.e., the number of nodes are having
a key). This value is involved to compute the connectivity and resilience of
KPS.
4. Theorem 1.4 provides the number of common keys between two nodes. The
common keys between two nodes are used to establish a link between them.
This value is involved to compute the resilience of the KPS.
5. Theorem 1.5 presents the resilience of the network among the nodes at a
particular depth. That is, the failure probability of a link in the network at
depth d is computed after the capture of t nodes from the same network. The
failure probability may vary if a number of nodes are captured from different
depths. For example, if a node is captured at depth d, then all the nodes in
the subtree with root as the captured node are exposed. Hence, all the links
involving a node in the subtree fail.
Further, capture of a node at higher depth exposes some keys in nodes at
lower depth and it is possible that capturing some nodes in higher depth may
completely expose a node at lower depth. Hence, this HKPS does not provide
the resilience against the compromise of any number of nodes at higher depth
in the hierarchy or leave nodes.
6. Theorem 1.6 presents the connectivity of the network among the nodes at
a particular depth. That is, the probability that two nodes at a particular
depth d can set a link between them. Further, as a parent node stores all keys
of a child node under its subtree, the parent node and its child can always
establish a link between them.
chain idea is used to achieve the goal. We too use the same hash chain idea in
the direction of depth of the hierarchy as follows.
1. Given an HKPS (X , B 1 , B 2 , · · · , B l ) as described in Subsect. 3.1.
2. Given a key K and a full domain cryptographic hash function h : K
→ K, let
inductively define hi (K) = h(hi−1 (K)) for i ≥ 1 and h0 (K) = K.
3. Depth of the hierarchy is used to discriminate the initially preloaded KPS
keys as described below:
(a) instead of the original key, K, a node at depth d is preloaded with the
key hd (K), for each key K to be distributed to the node in the HKPS.
(b) thus, two nodes at depth d and e that shared the same key K in the
HKPS end up possessing hd (K) and he (K) respectively;
(c) if d < e then the node at depth d (lower depth) can compute the key
at higher depth e i.e., he (K) = h(e−d) (hd (K)) and due to the preimage
resistant property of the cryptographic hash function h, the node at depth
e can not compute the key at lower depth d i.e., hd (K).
4. Key establishment of the nodes at depth d and e where d ≤ e, establish the
shared secret key Kd,e = he (K), that can be computed at the both end where
K is a shared key due to the HKPS.
5. Capture of a node at depth d exposes all its keys hd (K) to the adversary.
Then, the adversary
(a) can expose the keys of the nodes at higher or same depth if that possess
a key he (K) where d ≤ e;
(b) can not expose a single key of the nodes at lower depth.
So, the hash chain based HKPS (i.e., HC − HKP S) is resilient against the
compromise of any number of lower level users. Further, the connectivity, storage
overhead and communication overhead remains the same as the HKPS. On the
computation point of view, the nodes at lower depth have to pay for hash function
computations, which is not bothersome as nodes at lower depth are supposed to
be more powerful than the nodes at a higher depth. Hence the product based
hierarchical scheme HC − HKP S satisfies the following criteria.
1. decentralized hierarchy of a fixed number of depths (l). Decentralization is
necessary for large networks to distribute computational burden and security
threat to lower level users by reducing the burden of the central authority.
2. resilient against compromise of (i) any number of lower level nodes; and (ii)
a threshold number of nodes of same level in the hierarchy. This is achieved
by exploiting the hash chains technique in HC − HKP S.
3. non-interactive which saves bandwidth and energy.
4. deterministic KPS which implies the nodes in the network have predictable
behaviour of key rings.
5. efficient as it uses hash chains rather than any public key based key exchange
or bilinear maps.
6. free to choose any basic KPS at any level/depth of hierarchy as per the
requirement of that level.
7. simplicity in design as it uses the Cartesian product and hash chains over the
underlying basic schemes.
A Basic Theory of Lightweight Hierarchical Key Predistribution Scheme 337
An interested reader can refer to [10, Section 4.1] for some examples of nets. Now
to construct the class of KPS, Sensornet, consider a net in Vn that comprises
of the design (X , A), where X = Vn is set of varieties and the set of blocks is
A = {α + Ei : α ∈ Eic and 0 ≤ i < s}. The design of the KPS class, Sensornet
is a combination of [10, Theorem 1] with Sect. 2.4, that is restated in Result 1.
The KPS Sensornet is designed by assigning a sensor node Nij to the j-th coset
of Ei (i.e., Aij = αij + Ei ) as its set of key identifiers for 0 ≤ i < s, 0 ≤ j <
pm . Sensornet(s, pm ); whereas Sensornet denotes the generic KPS class. Due
to the affine structure of blocks in Sensornet, a key
establishment process is
exceptionally efficient and is of O(n3 ) = O (log N )3 .
Lemma 1. Given a KPS Sensornet which is a (pn , spm , s, pm )–configuration,
we have,
1. the resilience after a capture of t nodes R(t) = (1 − p−m )t ;
2. the connectivity of the KPS is C = 1 − 1s .
338 D. K. Dalai
Proof. The proof for connectivity and another metric for resilience (fail(t)) is
available in [10, Corollary 1, Corollary 2] respectively. We will prove for the
resilience R(t).
Here, each node stores pm keys from pn keys. Let a node is captured by the
adversary. Anmarbitrary link L secured by a key from the captured node is with
probability ppn = p−m . The link L is not affected with probability 1 − p−m .
Hence, the link L is not affected by a capture of t nodes is (1 − p−m )t . Therefore,
R(t) = (1 − p−m )t .
i )–configuration for 1 ≤ i ≤
Consider the Sensornets which are (pi , si pi , si , pm
ni mi i
d
3. Each element x ∈ X belongs to rd = i=1 si blocks in B d at depth d. Further,
l d
the element x belongs to d=1 i=1 si blocks in the whole hierarchical design.
4. (a) The number of common elements between two blocks Bi1 ,i2 ,··· ,id and
l
Bj1 ,j2 ,··· ,jd at depth d is 0 or i=d+1 pni i .
(b) The number of common elements between two blocks Bi1 ,i2 ,··· ,id and
e l
Bj1 ,j2 ,··· ,je of depths d and e, with e ≥ d, is 0 or i=d+1 pm i
i ni
i=e+1 pi .
5. The resilience among the nodes in dthe network at depth d after a capture of t
nodes at depth d is Rd (t) = 1 − i=1 (1 − (1 − p−m i
i t
) ).
In particular, the resilience among the nodes in the network of leave nodes is
Rl (t) = 1 − i=1 (1 − (1 − p−m
l i t
) ).
i d
6. The connectivity in the network at depth d is C d = i=1 (1 − s1i ).
In particular, the connectivity in the network of leave nodes (i.e., at depth l)
l
is C l = i=1 (1 − s1i ).
As Sensornet is not fully connected (i.e., connectivity is less than 1), HKPS-
Sensornet is not fully connected. Since Sensornet is very efficient and a CID with
a high value of μ, it is not troublesome to establish a link through an intermediate
node. Further, the designer can choose appropriate pi , ni and si to have enough
nodes and storage space to store the keys in the hierarchy. For example, the
A Basic Theory of Lightweight Hierarchical Key Predistribution Scheme 339
designer may choose smaller pni i for smaller value of i to have smaller key ring.
Moreover, the designer can choose larger si to have more number of nodes and
higher connectivity at a depth.
5 Conclusion
Realizing the need of HKPS with desirable properties to address the problem of
key management in low cost networks, we propose a basic theory of HKPS by
introducing the hierarchical design. Then we propose an HKPS on the basis of
the product of different KPS. Further, to achieve leaf-resilient in the HKPS, we
propose HC-HKPS by involving hash chain. For the investigation purpose, we
instantiate the HKPS on the KPS Sensornet.
Each level/depth in the hierarchy allows to use different KPS as per the
requirement. This freedom in the design of product based HKPS and introduc-
tion of a basic theory certainly open the doors for future research for the design
of different HKPS as per the requirements of the metrics.
References
1. Bechkit, W., Bouabdallah, A., Challal, Y.: Enhancing resilience of probabilistic
key pre-distribution schemes for WSNs through hash chaining. In: Proceedings of
the 17th ACM Conference on Computer and Communications Security, CCS 2010,
4–8 October 2010, Chicago, Illinois, USA, pp. 642–644 (2010)
2. Bechkit, W., Challal, Y., Bouabdallah, A.: A new class of hash-chain based key
pre-distribution schemes for WSN. Comput. Commun. 36(3), 243–255 (2013)
3. Çamtepe, S.A., Yener, B.: Combinatorial design of key distribution mechanisms
for wireless sensor networks. In: Samarati, P., Ryan, P., Gollmann, D., Molva, R.
(eds.) ESORICS 2004. LNCS, vol. 3193, pp. 293–308. Springer, Heidelberg (2004).
https://doi.org/10.1007/978-3-540-30108-0 18
4. Chakrabarti, D., Maitra, S., Roy, B.: A key pre-distribution scheme for wireless sen-
sor networks: merging blocks in combinatorial design. In: Zhou, J., Lopez, J., Deng,
R.H., Bao, F. (eds.) ISC 2005. LNCS, vol. 3650, pp. 89–103. Springer, Heidelberg
(2005). https://doi.org/10.1007/11556992 7
5. Chakrabarti, D., Seberry, J.: Combinatorial structures for design of wireless sensor
networks. In: Zhou, J., Yung, M., Bao, F. (eds.) ACNS 2006. LNCS, vol. 3989, pp.
365–374. Springer, Heidelberg (2006). https://doi.org/10.1007/11767480 25
6. Chan, H., Perrig, A., Song, D.: Random key predistribution schemes for sensor
networks. In: IEEE Symposium On Security and Privacy, pp. 197–213. IEEE Com-
puter Society (2003)
7. Chen, C.Y., Chao, H.C.: A survey of key predistribution in wireless sensor net-
works. Secur. Commun. Netw. 7(12), 2495–2508 (2014)
8. Dalai, D.K., Sarkar, P.: Key predistribution schemes using bent functions in dis-
tributed sensor networks. In: Chen, K., Lin, D., Yung, M. (eds.) Inscrypt 2016.
LNCS, vol. 10143, pp. 367–385. Springer, Heidelberg (2016). https://doi.org/10.
1007/978-3-319-54705-3 23
340 D. K. Dalai
9. Dalai, D.K., Sarkar, P.: Enhancing resilience of KPS using bidirectional hash chains
and application on sensornet. In: Yan, Z., Molva, R., Mazurczyk, W., Kantola, R.
(eds.) NSS 2017. LNCS, vol. 10394, pp. 683–693. Springer, Cham (2017). https://
doi.org/10.1007/978-3-319-64701-2 54
10. Dalai, D.K., Sarkar, P.: Sensornet - a key predistribution scheme for distributed
sensors using nets. In: Proceedings of the 6th International Conference on Sensor
Networks - SENSORNETS, pp. 49–58. INSTICC, ScitePress (2017)
11. Erdős, P., Rényi, A.: On the evolution of random graphs. In: Publication of the
Mathematical Institute of the Hungarian Academy of Sciences, pp. 17–61 (1960)
12. Eschenauer, L., Gligor, V.D.: A key-management scheme for distributed sensor
networks. In: ACM Conference on Computer and Communications Security, pp.
41–47 (2002)
13. Gennaro, R., Halevi, S., Krawczyk, H., Rabin, T., Reidt, S., Wolthusen, S.D.:
Strongly-resilient and non-interactive hierarchical key-agreement in MANETs. In:
Jajodia, S., Lopez, J. (eds.) ESORICS 2008. LNCS, vol. 5283, pp. 49–65. Springer,
Heidelberg (2008). https://doi.org/10.1007/978-3-540-88313-5 4
14. Kendall, M., Martin, K.M.: Graph-theoretic design and analysis of key predistri-
bution schemes. Des. Codes Cryptogr. 81(1), 11–34 (2016)
15. Lee, J., Stinson, D.R.: A combinatorial approach to key predistribution for dis-
tributed sensor networks. In: IEEE Wireless Communications and Networking
Conference WCNC 2005, New Orleans, USA, pp. 1200–1205 (2005)
16. Paterson, M.B., Stinson, D.R.: A unified approach to combinatorial key predistri-
bution schemes for sensor networks. Des. Codes Cryptogr. 71(3), 433–457 (2014)
17. Ramkumar, M., Memon, N., Simha, R.: A hierarchical key pre-distribution scheme.
In: Electro/Information Technology Conference, EIT 2005. IEEE (2005)
18. Ruj, S., Roy, B.K.: Key predistribution using combinatorial designs for grid-group
deployment scheme in wireless sensor networks. TOSN 6(1), 4:1–4:28 (2009)
19. Ruj, S., Roy, B.K.: Key pre-distribution using partially balanced designs in wireless
sensor networks. IJHPCN 7(1), 19–28 (2011)
20. Sakai, R., Ohgishi, K., Kasahara, M.: Cryptosystems based on pairings. In: Pro-
ceedings of SCIS 2000 (2000)
21. Sarkar, P.: Lightweight deterministic non interactive (ni) hierarchical key agree-
ment scheme (KAS). In: Yan, Z., Molva, R., Mazurczyk, W., Kantola, R. (eds.)
NSS 2017. LNCS, vol. 10394, pp. 315–331. Springer, Cham (2017). https://doi.
org/10.1007/978-3-319-64701-2 23
22. Steiner, J.G., Neuman, B.C., Schiller, J.I.: Kerberos: an authentication service for
open network systems. In: USENIX Winter, pp. 191–202 (1988)
23. Stinson, D.R.: Combinatorial Designs - Constructions and Analysis. Springer, Hei-
delberg (2004). https://doi.org/10.1007/b97564
24. Wei, R., Wu, J.: Product construction of key distribution schemes for sensor net-
works. In: Handschuh, H., Hasan, M.A. (eds.) SAC 2004. LNCS, vol. 3357, pp.
280–293. Springer, Heidelberg (2004). https://doi.org/10.1007/978-3-540-30564-
4 20
Adversarial Examples
for Hardware-Trojan Detection
at Gate-Level Netlists
1 Introduction
Demand of integrated circuits (ICs) has recently been increased due to introduc-
tion of Internet of Things (IoT) in foundries and home. In order to effectively
design and produce hardware devices at low cost, they have been become compli-
cated. Hardware design and production level can be divided into two steps: the
design step and the manufacturing step. In the design step, hardware vendors
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 341–359, 2020.
https://doi.org/10.1007/978-3-030-42048-2_22
342 K. Nozawa et al.
2 Related Works
# Description
1 The number of logic-gate fanins 4-level away from the target net n
2 The number of logic-gate fanins 5-level away from the target net n
3 The number of flip-flops up to 4-level away from the input side of the target net n
4 The number of flip-flops up to 3-level away from the output side of the target net n
5 The number of flip-flops up to 4-level away from the output side of the target net n
6 The number of up to 4-level loops from the input side of the target net n
7 The number of up to 5-level loops from the output side of the target net n
8 The minimum level to the primary input from the target net n
9 The minimum level to the primary output from the target net n
10 The minimum level to any flip-flop from the output side of the target net n
11 The minimum level to any multiplexer from the output side of the target net n
1
The number of Trojan nets identified as Trojan nets is called as true positive (TP).
The number of Trojan nets identified as normal nets is called as false negative (FN).
The true positive rate is obtained from TP / (TP + FN).
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 345
Given the fact of the development in machine learning utilizing neural net-
works, novel attack methods have been proposed. In [10], for example, a method
generating test data that cause misclassification in image recognition has been
proposed. Image recognition is frequently leveraged in physical world such as
self-driving and face recognition, thus both of the attack methods and defense
methods are getting much attention [7,8,18]. Although we cannot distinguish
adversarial examples from original ones, a classifier misclassifies them. In addi-
tion to image recognition, audio recognition [3], sentence recognition [15,16] and
graph data [5,30] become new targets these days. In this paper, we propose
a framework to generate AEs against hardware-Trojan detection by modifying
hardware designs. In image recognition, AEs are generated with minimizing visi-
ble impacts. Likewise, in this case, we aim to generate AEs which hardly degrade
circuit performance such as power consumption and path delay. The AEs cause
misclassification as a Trojan net mistakenly classified as a normal net.
Now we focus on AE attacks on hardware-Trojan detection utilizing neural
networks. In AE attacks against hardware-Trojan detection, an adversary aims
346 K. Nozawa et al.
to decrease |Et |. When |Et | is dramatically decreased, most of the Trojan nets
are mistakenly classified as normal nets. If the AE attack is realized, it becomes
hard to detect hardware Trojans in a netlist. In order to learn hardware design
information utilizing machine learning methods, we represent hardware circuits
as graph structures as shown in Sect. 2.1. The conversion from a graph structure
space into a feature space is one-way. Therefore, even if we add perturbations in
feature value space, we can not specify corresponding changes in graph structure.
In case of hardware-Trojan detection, it is difficult to completely represent
circuit structures as graph structures because circuit elements have several char-
acteristics such as particular I/O ports, power consumption, and delay time. In
particular, power consumption and delay time are important in hardware design.
We must take account of these characteristics towards AE attacks on hardware-
Trojan detection. For these reasons, we need to take a different approach from
conventional AE-attack methods.
For the purposes described above, adversaries may try to attack hardware
designs. We assume a white-box attack, the worst-case scenario of an AE attack
to analyze an attack scheme in an initial examination of AE attacks on hardware
design. This will give us the meaningful clues to further develop hardware-Trojan
detection methods based on machine learning. In this paper, we assume that
adversaries might know the following points.
I1 Adversaries have already known that the hardware vendor utilizes neural
networks to detect hardware Trojans.
I2 Adversaries have already known the structure and parameters of the neural
networks.
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 347
other hand, if adversaries modify the original circuit, original designers can easily
detect the modification because the original circuit itself is elaborately designed
to satisfy the requirements such as critical path delay and power consumption.
Therefore, adversaries should modify Trojan nets to achieve AE attacks.
Based on the discussion above, generating AEs against hardware-Trojan
detection is summarized as follows.
Point 1 Modified circuits are logically equivalent to the original ones.
Point 2 Only Trojan nets are modified.
Point 3 AEs degrade classification performance and conceal Trojan nets with
small modification.
In this paper, we propose Trojan-net concealment degree (TCD) and modi-
fication evaluating value (MEV), which indicates how likely the Trojan-nets are
classified as normal. Then, we design AEs for hardware design information and
examine its validity by those two evaluating values.
Based on these values, we further propose an AE generation method that
enables us to degrade the performance of hardware-Trojan detection in a short
time. It is impractical that we modify all the nets in a netlist because the num-
ber of modification patterns is exponentially increased. Our proposed method
chooses the modification whose MEV is the best at that time. Based on the
method, we realize to generate an appropriate AE against hardware-Trojan
detection in a practical time.
where K is the number of units in the output layer (K = 2 in the case of [12]),
p1 (x(e)) and p2 (x(e)) are the functions to return answer labels of x(e). When
e is a Trojan net, p1 (x(e)) equals to 0 and p2 (x(e)) equals to 1. q1 (x(e)) and
q2 (x(e)) are the functions to return prediction by the classifier. q1 and q2 satisfy
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 349
When TCD is large, the difference between the prediction and answer is large.
Therefore, if the value is large enough, adversaries can easily achieve their pur-
poses to conceal Trojan nets.
method chooses the modification whose MEV is the best at that time. The
proposed method is shown in Algorithm 1 below. Note that P is a set of possible
AE patterns by which a Trojan gate is modified to a logically equivalent circuit.
Examples of such patterns are provided in Sect. 4.2.
4 Experiments
In this section, we apply AE patterns shown in Sect. 4.2 to the hardware Trojan
circuits and classify them with the hardware-Trojan classifier proposed in [12]
under the conditions shown in Sect. 4.1. We perform the experiments in two
cases: (1) we do not consider the amount of modifications and (2) we consider
the amount of modifications. Then, we compare the results so that we can verify
the validity of MEV. We show the results when we do not consider the amount
of modifications in Sect. 4.3, and the results when we consider the amount of
modifications in Sect. 4.4.
Most of hardware Trojans have trigger conditions to hide their activity in normal
operation. In general, trigger conditions are set based on the values of internal
signals and internal states. If we modify the Trojan circuits with arbitrary con-
ditions, trigger conditions are also altered. Therefore, modified circuits must be
logically equivalent to the original hardware Trojan to satisfy the condition of
Point 1 mentioned in Sect. 3.1. In addition, these modifications should be applied
only to Trojan nets for condition of Point 2.
The hardware-Trojan detection method in [12] utilizes the distance from the
net to a near flipflop, multiplexer and primary input/output and thus modifica-
tions with changing the distance are efficient. For instance, the number of logic
352 K. Nozawa et al.
1'b0
NAND1
OR1
NAND2
AND1
NAND3 WIRE1
OR3 AND2
NAND4
AND3
NAND5
OR2
NAND6
NAND7
Fig. 1. Hardware Trojan (Trojan nets and Trojan gates) embedded in RS232-
T1000 [28].
In this experiment, we set λ in Eq. (5) to 0. We consider only the loss function
of neural networks on generating AE in this case.
Table 4 shows the results when we apply the six AE patterns to RS232-T1000.
Table 4 contains TN, FP, FN, TP, TPR, TNR (True Negative Rate)2 , accuracy
and MEV (Eq. (5)). From the viewpoint of MEV, the most efficient modification
to induce misclassification is t4 (WIRE1). The MEV becomes −1.69. In fact, t4
(WIRE1) shows the lowest TPR. TP is decreased from 34 to 26 by 8 compared to
the original circuit. Hence, TPR decreases to 68.42 % by 26.02 points. Note that
the number of Trojan nets is increased by 2 due to the modification. Therefore,
the denominator of TPR, the total number of Trojan nets (FN and TP), is also
increased.
1'b0
Gate Gate Gate Gate
1'b0
1'b1
1'b1
2
The number of normal nets identified as normal nets is called as true negative (TN).
The number of normal nets identified as Trojan nets is called as false positive (FP).
The true negative rate is obtained from TN/(TN + FP).
354 K. Nozawa et al.
Pattern Gate TN FP FN TP TPR TNR Accuracy The increased The increased MEV
number of number of
gates logic levels
None – 275 8 2 34 94.44% 97.17% 96.87% 0 0 −0.24
t1 OR1 275 8 4 34 89.47% 97.17% 96.26% 2 1 −0.25
t1 OR2 275 8 6 32 84.21% 97.17% 95.64% 2 1 −0.73
t2 OR1 275 8 3 35 92.11% 97.17% 96.57% 2 2 −0.31
t2 OR2 275 8 6 32 84.21% 97.17% 95.64% 2 2 −0.98
t3 OR3 277 8 5 33 86.84% 97.19% 95.98% 2 1 −0.42
t4 WIRE1 275 8 12 26 68.42% 97.17% 93.77% 2 2 −1.69
t5 AND1 277 8 4 34 89.47% 97.19% 96.28% 2 1 −0.28
t5 AND2 277 8 2 36 94.74% 97.19% 96.90% 2 1 −0.23
t5 AND3 277 8 3 35 92.11% 97.19% 96.59% 2 1 −0.33
t6 NAND1 275 8 2 36 94.74% 97.17% 96.88% 2 1 −0.23
t6 NAND2 275 8 5 33 86.84% 97.17% 95.95% 2 1 −0.79
t6 NAND3 275 8 5 33 86.84% 97.17% 95.95% 2 1 −0.39
t6 NAND4 275 8 3 35 92.11% 97.17% 96.57% 2 1 −0.59
t6 NAND5 275 8 5 33 86.84% 97.17% 95.95% 2 1 −0.94
t6 NAND6 275 8 4 34 89.47% 97.17% 96.26% 2 1 −0.86
t6 NAND7 275 8 5 33 86.84% 97.17% 95.95% 2 1 −0.67
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 355
Pattern Gate TN FP FN TP TPR TNR Accuracy The increased The increased MEV
number of number of
gates logic levels
None – 275 8 2 34 94.44% 97.17% 96.87% 0 0 −0.24
t4 WIRE1 275 8 12 26 68.42% 97.17% 93.77% 2 2 −1.69
t1 OR2 275 8 20 20 50.00% 97.17% 91.33% 4 3 −3.56
t1 OR1 275 8 23 19 45.24% 97.17% 90.46% 6 3 −3.93
t2 OR1 275 8 25 17 40.48% 97.17% 89.85% 6 4 −4.40
t3 OR3 277 8 12 30 71.43% 97.19% 93.88% 6 4 −1.43
t5 AND1 277 8 22 20 47.62% 97.19% 90.83% 6 4 −4.40
t5 AND2 277 8 20 22 52.38% 97.19% 91.44% 6 4 −3.39
t5 AND3 277 8 22 20 47.62% 97.19% 90.83% 6 4 −3.52
t6 NAND1 275 8 23 19 45.24% 97.17% 90.46% 6 4 −3.98
t6 NAND2 275 8 22 20 47.62% 97.17% 90.77% 6 4 −4.09
t6 NAND3 275 8 25 17 40.48% 97.17% 89.85% 6 4 −4.40
t6 NAND4 275 8 23 19 45.24% 97.17% 90.46% 6 4 −4.34
t6 NAND5 275 8 23 19 45.24% 97.17% 90.46% 6 4 −4.46
t6 NAND6 275 8 22 20 47.62% 97.17% 90.77% 6 4 −3.78
t6 NAND7 275 8 23 19 45.24% 97.17% 90.46% 6 4 −4.00
In this experiment, we set both λ1 and λ2 in Eq. (5) to 1. We consider the amount
of modifications as well as the loss function of neural networks on generating AEs.
Table 6 shows the results when we apply the six AE patterns to RS232-
T1000. From the viewpoint of MEV, the most efficient modification to induce
misclassification is t6 (NAND5). The MEV becomes 2.06. TP is decreased from
34 to 33 by 1 compared to the original circuit. Hence, TPR is decreased to 86.84
% by 7.60 points.
We repeatedly apply modifications to the circuit with t6 (NAND5). Table 7
shows the result when we repeatedly apply modifications up to three times. In
the same way as Table 6, the rows 1, 2 and 3 in Table 7 show the results of the
original circuit, the first modification and the second modification, respectively.
The following rows show the third modification. We pick up the modification
whose MEV is the lowest in each iteration. Finally, we pick up t6 (NAND5)
first, t6 (NAND6) second and t1 (OR2). From the viewpoint of MEV, the most
efficient modification to induce misclassification is t1 (OR2). The MEV becomes
4.29. TP is decreased from 34 to 27 by 7 compared to the original circuit. Hence,
TPR is decreased to 64.29 % by 30.15 points.
According to Table 5, applying t6 (NAND5) at the third iteration increases
six gates and four logic levels compared to the original circuit. On the other
hand, as shown in Table 7, applying t1 (OR2) at the third iteration increases six
gates and just one logic level compared to the original circuit. t1 (OR2) has less
amount of modification than t6 (NAND5). Reflecting this difference to MEV
356 K. Nozawa et al.
Pattern Gate TN FP FN TP TPR TNR Accuracy The increased The increased MEV
number of number of
gates logic levels
None – 275 8 2 34 94.44% 97.17% 96.87% 0 0 −0.24
t1 OR1 275 8 4 34 89.47% 97.17% 96.26% 2 1 2.75
t1 OR2 275 8 6 32 84.21% 97.17% 95.64% 2 1 2.27
t2 OR1 275 8 3 35 92.11% 97.17% 96.57% 2 2 3.69
t2 OR2 275 8 6 32 84.21% 97.17% 95.64% 2 2 3.02
t3 OR3 277 8 5 33 86.84% 97.19% 95.98% 2 1 2.58
t4 WIRE1 275 8 12 26 68.42% 97.17% 93.77% 2 2 2.31
t5 AND1 277 8 4 34 89.47% 97.19% 96.28% 2 1 2.72
t5 AND2 277 8 2 36 94.74% 97.19% 96.90% 2 1 2.77
t5 AND3 277 8 3 35 92.11% 97.19% 96.59% 2 1 2.67
t6 NAND1 275 8 2 36 94.74% 97.17% 96.88% 2 1 2.77
t6 NAND2 275 8 5 33 86.84% 97.17% 95.95% 2 1 2.21
t6 NAND3 275 8 5 33 86.84% 97.17% 95.95% 2 1 2.61
t6 NAND4 275 8 3 35 92.11% 97.17% 96.57% 2 1 2.41
t6 NAND5 275 8 5 33 86.84% 97.17% 95.95% 2 1 2.06
t6 NAND6 275 8 4 34 89.47% 97.17% 96.26% 2 1 2.14
t6 NAND7 275 8 5 33 86.84% 97.17% 95.95% 2 1 2.33
Pattern Gate TN FP FN TP TPR TNR Accuracy The increased The increased MEV
number of number of
gates logic levels
none – 275 8 2 34 94.44% 97.17% 96.87% 0 0 −0.24
t6 NAND5 275 8 5 33 86.84% 97.17% 95.95% 2 1 2.06
t6 NAND6 275 8 8 32 80.00% 97.17% 95.05% 4 1 3.48
t1 OR1 276 8 8 34 80.95% 97.18% 95.09% 6 1 5.51
t1 OR2 275 8 15 27 64.29% 97.17% 92.92% 6 1 4.29
t2 OR1 275 8 9 33 78.57% 97.17% 94.77% 6 2 6.48
t2 OR2 275 8 13 29 69.05% 97.17% 93.54% 6 2 5.44
t3 OR3 277 8 14 28 66.67% 97.19% 93.27% 6 2 5.88
t4 WIRE1 275 8 19 23 54.76% 97.17% 91.69% 6 3 5.76
t5 AND1 277 8 15 27 64.29% 97.19% 92.97% 6 2 5.90
t5 AND2 277 8 9 33 78.57% 97.19% 94.80% 6 2 6.47
t5 AND3 277 8 9 33 78.57% 97.19% 94.80% 6 2 6.46
t6 NAND1 275 8 8 34 80.95% 97.17% 95.08% 6 1 5.55
t6 NAND2 275 8 11 31 73.81% 97.17% 94.15% 6 1 5.05
t6 NAND3 275 8 11 31 73.81% 97.17% 94.15% 6 1 5.40
t6 NAND4 275 8 9 33 78.57% 97.17% 94.77% 6 1 5.19
t6 NAND7 275 8 11 31 73.81% 97.17% 94.15% 6 1 5.12
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 357
5 Conclusion
References
1. Akhtar, N., Mian, A.: Threat of adversarial attacks on deep learning in computer
vision: a survey. IEEE Access 6, 14410–14430 (2018)
2. Bhunia, S., Hsiao, M.S., Banga, M., Narasimhan, S.: Hardware Trojan attacks:
threat analysis and countermeasures. Proc. IEEE 102(8), 1229–1247 (2014)
3. Carlini, N., Wagner, D.: Audio adversarial examples: targeted attacks on speech-
to-text. In: 2018 IEEE Security and Privacy Workshops (SPW) (2018)
4. Chakraborty, R.S., Narasimhan, S., Bhunia, S.: Hardware Trojan: threats and
emerging solutions. In: Proceedings of International High-Level Design Validation
and Test Workshop (HLDVT), pp. 166–171 (2009)
5. Dai, H., et al.: Adversarial attack on graph structured data. In: Proceedings of
International Conference on Machine Learning (ICML) (2018)
6. Dong, C., He, G., Liu, X., Yang, Y., Guo, W.: A multi-layer hardware trojan
protection framework for IoT chips. IEEE Access 7, 23628–23639 (2019)
7. Eykholt, K., et al.: Physical adversarial examples for object detectors. CoRR (2018)
8. Eykholt, K., et al.: Robust physical-world attacks on deep learning models. CoRR
(2017)
9. Francq, J., Frick, F.: Introduction to hardware Trojan detection methods. In: 2015
Design, Automation and Test in Europe Conference and Exhibition (DATE), pp.
770–775. EDAA (2015)
10. Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial
examples. In: Proceedings of 2015 International Conference on Learning Represen-
tations (ICLR) (2015)
358 K. Nozawa et al.
11. Grosse, K., Papernot, N., Manoharan, P., Backes, M., McDaniel, P.: Adversarial
examples for malware detection. In: Foley, S.N., Gollmann, D., Snekkenes, E. (eds.)
ESORICS 2017. LNCS, vol. 10493, pp. 62–79. Springer, Cham (2017). https://doi.
org/10.1007/978-3-319-66399-9 4
12. Hasegawa, K., Yanagisawa, M., Togawa, N.: Hardware Trojans classification for
gate-level netlists using multi-layer neural networks. In: Proceedings of 2017 IEEE
23rd International Symposium on On-Line Testing and Robust System Design
(IOLTS), pp. 227–232 (2017)
13. Inoue, T., Hasegawa, K., Yanagisawa, M., Togawa, N.: Designing hardware Trojans
and their detection based on a SVM-based approach. In: Proceedings of Interna-
tional Conference on ASIC, pp. 811–814 (2018)
14. Inoue, T., Hasegawa, K., Yanagisawa, M., Togawa, N.: Designing subspecies of
hardware Trojans and their detection using neural network approach. In: Proceed-
ings 2018 IEEE 8th International Conference on Consumer Electronics in Berlin
(ICCE-Berlin) (2018)
15. Iyyer, M., Wieting, J., Gimpel, K., Zettlemoyer, L.: Adversarial example generation
with syntactically controlled paraphrase networks. In: Proceedings of the 2018
Conference of the North American Chapter of the Association for Computational
Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 1875–
1885. Association for Computational Linguistics (2018)
16. Jia, R., Liang, P.: Adversarial examples for evaluating reading comprehension sys-
tems. In: Proceedings of the 2017 Conference on Empirical Methods in Natural
Language Processing, pp. 2021–2031. Association for Computational Linguistics
(2017)
17. Jin, Y., Makris, Y.: Hardware Trojan detection using path delay fingerprint. In:
Proceedings of IEEE International Workshop on Hardware-Oriented Security and
Trust (HOST), pp. 51–57 (2008)
18. Kurakin, A., Goodfellow, I.J., Bengio, S.: Adversarial examples in the physical
world. In: Proceedings of 2017 International Conference on Learning Representa-
tions (ICLR) (2017)
19. Lamech, C., Plusquellic, J.: Trojan detection based on delay variations measured
using a high-precision, low-overhead embedded test structure. In: 2012 IEEE Inter-
national Symposium on Hardware-Oriented Security and Trust, pp. 75–82, June
2012
20. Liu, B., Qu, G.: VLSI supply chain security risks and mitigation techniques: a
survey. Integr. VLSI J. 55, 438–448 (2016)
21. Liu, D.C., Nocedal, J.: On the limited memory bfgs method for large scale opti-
mization. Math. Program. 45(1), 503–528 (1989)
22. Moosavi-Dezfooli, S.M., Fawzi, A., Frossard, P.: DeepFool: a simple and accurate
method to fool deep neural networks. In: IEEE Conference on Computer Vision
and Pattern Recognition, pp. 2574–2582 (2016)
23. Oya, M., Shi, Y., Yanagisawa, M., Togawa, N.: A score-based classification method
for identifying hardware-Trojans at gate-level netlists. In: Proceedings of 2015
Design, Automation and Test in Europe Conference and Exhibition, pp. 465–470
(2015)
24. Rostami, M., Koushanfar, F., Rajendran, J., Karri, R.: Hardware security: threat
models and metrics. In: Proceedings of International Conference on Computer-
Aided Design (ICCAD), pp. 819–823 (2013)
25. Salmani, H., Tehranipoor, M., Karri, R.: On design vulnerability analysis and
trust benchmarks development. In: 2013 IEEE 31st International Conference on
Computer Design (ICCD), pp. 471–474 (2013)
Adversarial Examples for Hardware-Trojan Detection at Gate-Level Netlists 359
26. Shakya, B., He, T., Salmani, H., Forte, D., Bhunia, S., Tehranipoor, M.: Bench-
marking of hardware trojans and maliciously affected circuits. J. Hardware Syst.
Secur. 1(1), 85–102 (2017)
27. Szegedy, C., et al.: Intriguing properties of neural networks. CoRR (2013)
28. Trust-HUB. http://trust-hub.org/benchmarks/trojan
29. Xiao, K., Forte, D., Jin, Y., Karri, R., Bhunia, S., Tehranipoor, M.: Hardware
trojans: lessons learned after one decade of research. ACM Trans. Design Autom.
Electron. Syst. (TODAES) 22(1), 1–23 (2016)
30. Zügner, D., Akbarnejad, A., Günnemann, S.: Adversarial attacks on neural net-
works for graph data. In: Proceedings of the 24th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining - KDD 2018, pp. 2847–2856.
ACM Press (2018)
Selective Forwarding Attack on IoT
Home Security Kits
Abstract. Efforts have been made to improve the security of the Inter-
net of Things (IoT) devices, but there remain some vulnerabilities and
misimplementations. This paper describes a new threat to home security
devices in which an attacker can disable all functionality of a device,
but to the device’s owner, everything still appears to be operational. We
targeted home security devices because their security is critical as peo-
ple may rely on them to protect their homes. In particular, we exploited
a feature called “heartbeat”, which is exchanged between the devices
and the cloud in order to check that the devices are still connected.
Even though network traffic was encrypted, we successfully identified
the heartbeats due to their fixed size and periodic nature. Thereafter,
we established a man-in-the-middle attack between the device and the
cloud and selectively forwarded heartbeats while filtering out other traf-
fic. As a result, the device appears to be still connected (because the
heartbeat traffic is being allowed through), while in reality the device’s
functionality is disabled (because non-heartbeat traffic is being filtered
out). We applied this exploit on a set of six devices, and five were found
to be vulnerable. Consequently, an intruder can use this exploit to dis-
able a home security device and break into a house without the awareness
of the owner. We carried out a responsible disclosure exercise with the
manufacturers of the affected devices, but the response has been limited.
This shows that IoT security is still not taken completely seriously and
many threats are still undiscovered. Finally, we provide some recommen-
dations on how to detect and prevent the threats posed by insecure IoT
devices, which ironically include IoT home security kits.
1 Introduction
The Internet has considerably changed in the last decade. It has become more
than just a platform for email exchanges, web browsing, instant messaging or
media streaming. The connected devices are no longer just servers, computers
and smartphones, but instead the Internet has become an Internet of Things
(IoT), of connected wearables, home appliances, biomedical devices, cars, cities,
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 360–373, 2020.
https://doi.org/10.1007/978-3-030-42048-2_23
Selective Forwarding Attack on IoT Home Security Kits 361
and many more. IoT is gaining more and more popularity and experts predict
that it will become an Internet of Everything in the near future [1]. It is estimated
that the number of connected devices will reach up to 20 billion by 2020 [2].
Despite its convenience, IoT and its applications introduce major privacy
threats and critical security risks. For instance, IoT devices may be compromised
to access personal information, or to gain control over industries, cities and public
organisations or disrupt their services. This has been shown by several incidents
like the Mirai botnet [3,4]. Mirai is a computer worm that compromised hundreds
of thousands of IoT devices, which were then used to mount a Distributed Denial
of Service (DDoS) attack to disrupt well-known services like Netflix and Twitter
[5]. IoT also imposes personal privacy threats through smart home devices like
cameras, personal assistants and home automation kits. For example, Trendnet
home security and monitoring cameras were found to be vulnerable, allowing
an attacker to access live video feeds of the camera without any authentication
[6]. Likewise, customers of Swann home cameras reported that they were able to
access recordings from cameras of other customers [7]. A very recent report of
security vulnerabilities in three specialist car alarm systems further illustrates
the danger of connecting your device to the Internet without proper security
testing [8]. These vulnerabilities allowed attackers to steal or hijack affected
vehicles through the compromised alarm system. What is ironic here is that
whoever bought these vulnerable car alarm systems did so out of a desire to
improve the security of their vehicle. But inadvertently, they introduced security
vulnerabilities that would allow attackers to take control of their vehicle. This
irony resonates with the message we aim to convey in our paper.
Evidently, the security of IoT devices is a major issue that needs to be con-
tinuously evaluated and addressed due to its impact on the physical world. This
motivated us to explore new and common vulnerabilities in a selected set of
consumer IoT products. Particularly, we targeted home security devices because
their security is critical as people may rely on them to protect their homes.
Contribution. The main contribution of this paper is the exploitation of a
vulnerability in the heartbeat exchange of IoT devices. By exploiting this vulner-
ability, an attacker can disable IoT home security devices without the awareness
of their owners. Particularly, the device will appear to be online and working
normally, but in fact it will be completely disabled. This was due to wrong
implementations of heartbeat messages exchanged between devices and their
cloud infrastructure. Our second contribution relates to a potential misimple-
mentation of the WPA2 four-way handshake protocol in some IoT devices. This
misimplementation allows an attacker to carry out an evil twin access point
attack, which would force the device to connect to the attacker’s LAN. This
would allow the attacker to exploit further vulnerabilities, or eavesdrop on the
communications between the device and the cloud.
The rest of this paper is organised as follows. Section 2 presents related
work. Section 3 outlines our methodology, while Sect. 4 presents our findings
and results. Section 5 describes the threat model to illustrate the feasibility
of our attack. Section 6 dissects the risks and consequences of the discovered
362 A. Hariri et al.
2 Related Work
Visan et al. [9] assessed the security of Samsung Smart Things hub. They first
attempted to extract credentials from the hub using various traffic sniffing meth-
ods. They also demonstrated that the hub is robust and secure. However, they
discovered that a DoS attack against the hub is possible, if the attacker has
access to the LAN. The attack would give an intruder an 8-minute window to
break into a house before notifying the owner. Visan et al. argue that home
security kits are not completely reliable.
Another good example is the work by Fernandes et al. [10], which demon-
strated again that despite the effort put into security, there may remain some
security issues due to the complex nature of the products. More precisely, they
discovered vulnerabilities inside the architecture, in the capability model and the
event subsystem of Samsung Smart Things, due to the numerous and complex
functionalities exposed to the user. By exploiting them, they managed to insert
backdoor pin-codes into a connected door lock, eavesdrop door lock pin-codes,
cause a fake fire alarm and disable the vacation mode of the device. More impor-
tantly, those vulnerabilities were significant as they targeted the architecture of
the application layer at its core, thus making them difficult to patch.
Apthrope et al. [11] proved that users’ privacy can be breached without
compromising devices or network communications. They showed that any party
having network access – e.g. Internet Service Providers (ISPs) – can infer sen-
sitive information just by analysing network traffic. They particularly analysed
DNS requests, IP and TCP headers and packet rates to identify device types and
user interaction. For instance, they were able to determine if a user is sleeping
by analysing a sleep sensor’s traffic, or if the user is moving inside a house by
analysing a motion sensor’s traffic. This work highlights that privacy is a critical
challenge in IoT security, and it cannot be achieved with cryptography only.
Jerry Gamblin discovered that Google home assistant can be controlled by
any device that has network access to the LAN. The device can send commands
to the assistant without any authentication and it can cause to reboot and even
to disconnect from Wi-Fi. This is due to an undocumented API that can be
exploited by sending rogue commands [12].
Very recently, OConnor et al. [13] uncovered a design flaw in the messaging
protocols of 22 IoT devices. The design flaw they uncovered is very similar to the
weakness we discuss in this paper. We independently carried out our research
and developed a proof-of-concept automated tool to exploit this flaw.
Selective Forwarding Attack on IoT Home Security Kits 363
3 Methodology
We chose smart home IoT devices because they provide a wider attack sur-
face due to their numerous interconnected components such as cameras, alarms,
motion detectors and many other sensors. We targeted a set of home security
devices including Swann Smart Home Security Kit (SWO-HUB01K), D-Link
Home Security Kit (DCH-G020), D-Link camera (DCS-935L), Panasonic Home
Monitoring and Control Kit (KX-HN6012EW), Telldus Smart Home Start-up
Kit and Samsung SmartThings (F-H-ETH-001). We performed DoS attacks on
the selected devices, mainly Wi-Fi deauthentication and blackhole attacks. We
used arpspoof tool [14] for a blackhole attack and aireplay-ng tool [15] for Wi-
Fi deauthentication attack. arpspoof tool was used to achieve a MITM position
and IP forwarding was disabled to complete the blackhole attack. Moreover,
we analysed network traffic thoroughly to find patterns that were then used
in our selective forwarding attack. Traffic analysis was mainly conducted using
Wireshark [16], in which several Wireshark filters were used to analyse and
identify patterns in network traffic of each device. In addition, our selective
forwarding attack requires achieving a MITM position on the Transport layer.
Thus, it was necessary to run a TCP proxy and force the devices to connect
to the proxy instead of connecting to their legitimate servers. DNS poisoning
was the best option to force the devices to connect to the proxy. For that, we
used Wireshark to view DNS requests and responses, and identify the domain
names and the IP addresses of the servers. Then, we used Bind9 and configured
it to resolve the identified DNS requests to the IP address of the machine that
is running the proxy. As a result, all devices connected to the proxy instead of
connecting to their servers. The proxy was developed as a Python script, and it
can be found with the Bind9 configurations on GitHub1 .
4 Results
We conducted blackhole and Wi-Fi de-authentication attacks on the selected
devices to determine the required time to alert or display that the device went
offline. The results showed that the devices took between 5 s and 2 min to alert
the user depending on the vendor. Thereupon, we deduced that the devices
must be exchanging periodic messages with the cloud to prove that they are
still online. Once the cloud stops receiving those messages, it reports to the
user application that the device was disconnected. Those messages are known as
heartbeats and they are defined as “a periodic signal generated by hardware or
software to indicate that it is still running” [17].
We realised that heartbeats will always have a fixed size during one session
because the payload is always the same and it is always encrypted with the same
algorithm over this session. Subsequently, we postulated that heartbeat messages
can be identified in network traffic even if they were encrypted due to their
periodic nature and constant payload size. Therefore, we presumed that we can
1
https://github.com/HaririAli/IoTHeartbeatProxy.git.
364 A. Hariri et al.
We discovered that the heartbeats are sent as 48-byte long UDP datagrams
every 15 s. Since the heartbeats are sent as UDP datagrams, there was no need
to run a TCP proxy and selective forwarding could be achieved using iptables
only. Subsequently, we wrote iptables rules to allow only 48-byte long UDP
packets and drop everything else. However, this exploit was only tested when
both the hub and the smartphone are on the same LAN. Testing the exploit
on WAN was not possible due to unknown technical issues in which the mobile
application was not able to access the hub. The exploit succeeded on the LAN
and the application displayed that the hub is online. However, the application
was completely unresponsive because it could not receive anything apart from
heartbeats.
Visan et al. [9] proved that under a DoS attack, Samsung Smartthings provides
a time-window of 8 min before notifying the user. To build on their work, we
decided to apply the heartbeat exploit on the same device to check if it can
366 A. Hariri et al.
provide an infinite time window. We analysed network traffic collected from the
hub and noticed that both the server and the hub exchange heartbeats every 31
s; the hub sends 49 bytes and the server responds with 55 bytes of data.
Unlike the other devices, Samsung SmartThings proved that its heartbeat
exchange is robust and secure. The hub disconnected immediately upon selec-
tively forwarding its heartbeats. Subsequently, we analysed the behaviour of the
device to understand how it detected the exploit. We discovered that upon block-
ing non-heartbeat messages, either the device or the server disconnects right after
the next heartbeat. This means that the server was reporting in its heartbeat
whether it received the last message or not. This allowed the device to detect
that the last message was not received by the server, thus it disconnected. As
such, we think that the heartbeat messages in this system have some sort of
checking whether other (non-heartbeat) messages have passed through or not.
4.7 Summary
The sizes, types and periods of heartbeats of each device are summarised in
Table 1. During the experiment, we developed custom proxies for each device
and used them to conduct selective forwarding exploits. Once we confirmed our
results, we developed a generic selective forwarding proxy (outlined in Sect. 3)
that takes a set of arguments to customise its behaviour.
The results of this exploit proved that the devices can be completely dis-
abled while their corresponding applications still display them as online and
operational. This exploit can be considered as an unnoticed DoS attack, because
the system was not really available, but the application displayed that it was nor-
mally working. The results of the heartbeat exploit are summarised in Table 2,
where the third column states if the device can be physically disconnected dur-
ing the attack (which indicates a more severe security violation), and the last
column presents relevant further details for each device.
Selective Forwarding Attack on IoT Home Security Kits 367
5 Threat Model
In this section, we demonstrate how an attacker can theoretically and practi-
cally exploit the vulnerability in the heartbeat exchange. We then describe an
automated attack that we have successfully applied to one of the devices.
governmental and private organisations in Lebanon and the UAE. The attackers
redirected email and VPN traffic to an IP address managed by them [21]. Sim-
ilarly, attackers can redirect heartbeat traffic to a TCP proxy using this attack
and eventually exploit the vulnerability in the heartbeat exchange.
6 Discussion
Exploiting the heartbeat pattern is not critical in all IoT applications like light
bulbs for example. However, it can be remarkably critical in home security
kits and in some health care applications which are used to remotely moni-
tor patients. Furthermore, this attack can be extended to identify other patterns
in encrypted traffic like sensor readings or controller commands and selectively
forward and drop those patterns. For instance, attackers can block some control
commands in water distribution systems to cause floods, or in electricity distri-
bution systems to cause blackouts. Autonomous vehicles can also be affected by
this attack in which attackers filter communications between the vehicles lead-
ing to deadly accidents. Therefore, detection and defence mechanisms must be
studied and developed to prevent such attacks and to reach a more secure IoT.
To exploit heartbeats, an attacker must redirect network traffic to a malicious
proxy. Nevertheless, attackers can find many ways to gain access over the net-
work. This was proven by Chapman [27] who compromised WLAN credentials
by exploiting a vulnerability in LIFX light bulbs. Alternatively, attackers can
compromise Wi-Fi credentials using social engineering techniques. Therefore, it
is necessary to secure the heartbeat exchange in IoT, even though its exploit
requires network access.
implemented on the application layer instead of the transport layer like TCP
keepalive packets. Secondly, heartbeats must include information about the last
message that was sent before the heartbeat. This would allow both endpoints to
detect if any messages were filtered out. To understand this mechanism better,
consider this scenario:
7 Conclusion
Although the security of IoT devices has substantially improved in the past few
years, some vulnerabilities remain undiscovered in many IoT systems. Our paper
describes a new threat to home security devices in which an attacker can disable
a device while making it appear to be working normally to the user.
We demonstrated that heartbeats can be identified in network traffic even if
they were encrypted. Subsequently, we proved that heartbeats can be exploited
using selective forwarding attack. In particular, we redirected network traffic
from the devices to a TCP proxy that only forwards heartbeats between the hub
and the server and discards any other data. As a result, the user application
displays that the device is online because heartbeats are passed. However, the
device is in fact disabled because the rest of its traffic was blocked. This allows
an intruder to disable home security devices and break into houses without the
awareness of their owners.
We applied this exploit on a set of home security devices and most of them
were found to be affected by this vulnerability. We disclosed our findings to the
affected companies to fix the issue and release the necessary patches. To mitigate
this vulnerability, we propose some recommendations to implement a robust and
secure heartbeat exchange. Our results confirm the findings of OConnor et al.
[13], although our research was carried out independently and on a different set
of devices.
Our work proves that further research is still needed for a more secure IoT. In
addition, researchers should work closely with industry to ensure that security
standards are implemented correctly. To build on this work, selective forwarding
exploits should be extended to cover any type of traffic patterns rather than just
heartbeats. This would help researchers to develop robust frameworks that can
prevent or at least detect any type of selective forwarding.
References
1. Evans, D.: How the internet of everything will change the world...for the better
IOE [infographic], April 2013. https://blogs.cisco.com/digital/how-the-internet-
of-everything-will-change-the-worldfor-the-better-infographic
2. Meulen, R.V.D.: Gartner says 8.4 billion connected “things” will be in use in
2017, up 31 percent from 2016’, February 2017. https://www.gartner.com/en/
newsroom/press-releases/2017-02-07-gartner-says-8-billion-connected-things-
will-be-in-use-in-2017-up-31-percent-from-2016
3. Antonakakis, M., April, T., Bailey, M., et al.: Understanding the mirai botnet.
In: 26th USENIX Security Symposium (USENIX Security 17), pp. 1093–1110.
USENIX Association (2017)
4. Cetin, O., Ganán, C., Altena, L., et al.: Cleaning up the internet of evil things:
real-world evidence on ISP and consumer efforts to remove Mirai. In: Network and
Distributed Systems Security (NDSS) Symposium (2019)
5. Kolias, C., Kambourakis, G., Stavrou, A., Voas, J.: DDoS in the IoT: Mirai and
other botnets. Computer 50(7), 80–84 (2017). https://doi.org/10.1109/mc.2017.
201
372 A. Hariri et al.
6. Zetter, K.: Flaw in home security cameras exposes live feeds to hackers, June 2012.
https://www.wired.com/2012/02/home-cameras-exposed/
7. Kelion, L.: Swann’s home security camera recordings could be hijacked, July 2018.
https://www.bbc.co.uk/news/technology-44809152
8. Simmons, D.: Security holes found in big brand car alarms, March 2019. https://
www.bbc.co.uk/news/technology-47485731
9. Visan, B., Lee, J., Yang, B., Smith, A.H., Matson, E.T.: Vulnerabilities in hub
architecture IoT devices. In: 2017 14th IEEE Annual Consumer Communications
Networking Conference (CCNC), pp. 83–88, January 2017
10. Fernandes, E., Jung, J., Prakash, A.: Security analysis of emerging smart home
applications. In: IEEE Symposium on Security and Privacy (SP), pp. 636–654
(2016)
11. Apthorpe, N., Reisman, D., Feamster, N.: A smart home is no castle: privacy
vulnerabilities of encrypted IoT traffic. CoRR, vol. abs/1705.06805 (2017). http://
arxiv.org/abs/1705.06805
12. Shaun, N.: This one weird trick turns your google home hub into a doorstop,
November 2018. https://www.theregister.co.uk/2018/10/31/google home api
13. OConnor, T., Enck, W., Reaves, B.: Blinded and confused: uncovering systemic
flaws in device telemetry for smart-home internet of things. In: Proceedings of the
12th Conference on Security and Privacy in Wireless and Mobile Networks, pp.
140–150. ACM (2019)
14. Whalen, S., Engle, S., Romeo, D.: An introduction to ARP spoofing, April 2001.
http://index-of.es/Misc/pdf/arpspoofing slides.pdf
15. Aircrack-NG tool. https://www.aircrack-ng.org/doku.php?id=aireplay-ng
16. Wireshark: Wireshark Tool. https://www.wireshark.org/
17. PC Magazine Encyclopedia: Heartbeat Definition. https://www.pcmag.com/
encyclopedia/term/44190/heartbeat
18. All the networks. found by everyone. https://wigle.net/
19. Nakhila, O., Attiah, A., Jin, Y., Zou, C.: Parallel active dictionary attack on
WPA2-PSK wi-fi networks. In: MILCOM 2015–2015 IEEE Military Communi-
cations Conference, pp. 665–670. IEEE (2015)
20. Viehböck, S.: Brute forcing wi-fi protected setup, December 2011. https://sviehb.
files.wordpress.com/2011/12/viehboeck wps.pdf
21. Krebs, B.: A deep dive on the recent widespread DNS hijacking attacks,
February 2019. https://krebsonsecurity.com/2019/02/a-deep-dive-on-the-recent-
widespread-dns-hijacking-attacks/
22. Greene, T.: Attackers can take over cisco routers; other routers at risk, too,
September 2015. https://www.networkworld.com/article/2984124/attackers-can-
take-over-cisco-routers-other-routers-at-risk-too.html
23. Cisco Security Advisory: Cisco security threat and vulnerability intelligence, Jan-
uary 2019. https://tools.cisco.com/security/center/content/CiscoSecurityAdvisory
/cisco-sa-20190123-sdwan-file-write
24. IEEE Computer Society LAN/MAN Standards Committee: IEEE Standard for
Information technology-Telecommunications and information exchange between
systems-Local and metropolitan area networks-Specific requirements Part 11:
Wireless LAN Medium Access Control (MAC) and Physical Layer (PHY) Specifi-
cations. IEEE Std 802.11 (2007)
Selective Forwarding Attack on IoT Home Security Kits 373
1 Introduction
(SSH) service. Its goal was to subvert a large number of devices and coordinate
Distributed Denial of Service (DDoS) attacks to multiple targets. Chalubo is
an evolution of Mirai botnet [18,35] that compromised over 100,000 IoT devices
(mostly IP cameras, home routers, and digital video recorders) and used them
for DDoS attacks. These and other similar incidents indicate that the security of
IoT devices and networks must be re-examined and appropriate solutions should
be developed to protect businesses, consumers, and critical infrastructure.
Two widely used Operating Systems (OSes) for IoT are ContikiOS [1,10]
and TinyOS [2]. The behaviour of ContikiOS can be simulated using the Cooja
simulator [30]. Cooja is particularly suitable for real-world experiments, since
the developed applications can be directly uploaded to real hardware.
In this work, the main aim is to develop an Intrusion Detection System (IDS)
to detect DoS attacks which exploit RPL control packets. In order to achieve
that, we first simulate the behaviour of benign and malicious smart sensors in
Cooja using the Zolertia Z1 motes [3]. The malicious nodes attempt to consume
the energy of benign nodes and to disrupt their communications by launching
DoS attack. In particular, these attacks exploit the features of RPL - the IPv6
Routing Protocol for Low-Power and Lossy Networks [41].
We utilise the obtained results to design and develop a lightweight IDS for
RPL-based IoT networks.
Our IDS comprises both centralised and distributed components/modules. In
particular, a centralised router hosts the detection module and acts as a firewall,
whereas a number of spatially-distributed lightweight modules are deployed at
the network edge for the purposes of traffic monitoring and local detection [25].
Distributed modules send periodic reports to the router. They also send alerts
whenever certain predefined thresholds are exceeded, indicating abnormal or
unexpected activity in their proximity.
One of the advantages of our approach is that no software modification of
current devices/sensors is required. Furthermore, the IDS modules support both
wireless and wired connectivity. The latter might be particularly useful to avoid
wireless attacks.
The key contributions in the paper are as follows:
– IoT-specific DoS attack is implemented in ContikiOS and is simulated in
Cooja simulator. Mean packet interval and mean number of messages were
utilised to show the effects of the attack and to help us develop a method for
detecting these kind of attacks.
– A lightweight IDS for RPL-based IoT networks is implemented for detecting
RPL attacks.
– The IDS detection module is evaluated in Cooja simulator.
The rest of the paper is organized as follows. In Sect. 2, we provide the
required background information. Subsect. 2.1 presents our considered 7-layer
IoT reference model. Subsect. 2.2 briefly provides an overview of the RPL pro-
tocol. In Sect. 3, we discuss the existing DoS attacks and defenses in IoT. In
Sect. 4, we describe our DoS attack implementation in Cooja. Subsects. 4.1 and
4.2 describe the considered scenarios and simulation parameters used to simulate
376 P. P. Ioulianou and V. G. Vassilakis
DoS attack, respectively. Subsect. 4.3, presents the results obtained from simu-
lating DoS attack. In Sect. 5, we describe our proposed IDS design, including a
high-level architecture and its main components. In Sect. 6, we present the eval-
uation of our IDS. Subsect. 6.1 describes the implemented algorithms. Subsects.
6.2 and 6.3 discuss the simulation settings and the considered scenarios used
to evaluate the IDS, respectively. Subsect. 6.4 presents the experimental results
from IDS evaluation. Finally, in Sect. 7, we conclude and discuss future work.
2 Background
2.1 IoT Reference Model and Attack Vectors
In the literature, a number of different reference models have been proposed for
IoT [27]. Among them, CISCO’s 7-layer model [9] provides sufficient level of
detail and has been considered in this study. The layers of CISCO’s model are
the following:
1. Physical Devices & Controllers (the “things” in IoT)
2. Connectivity (communication & processing units)
3. Edge (Fog) Computing (data element analysis & transformation)
4. Data Accumulation (storage)
5. Data Abstraction (aggregation & access)
6. Application (reporting, analytics, control)
7. Collaboration & Processes (involving people & business processes)
Our current work mostly contributes in the first three layers. Below, we briefly
describe each layer and review possible attack vectors against IoT devices. At
Layer 1, we have the physical IoT devices such as sensors and actuators. Possible
attacks are malicious modification of firmware, and DoS attacks such as battery
draining and resource exhaustion [43].
Layer 2 refers to the connectivity among the devices, which send or receive
generated/censored data within the same network or across different networks.
Here, an adversary may sniff network packets and try to modify or contaminate
the routing information by launching routing, and Man-in-the-Middle (MitM)
attacks [5]. Furthermore, IoT devices can be turned into bots and used for DoS
attacks against selected targets, as was the case with the infamous Mirai botnet
[19]. Layer 3 operations include data analysis and transformation performed at
the network edge. Some of the attacks are malicious input that cause data leaking
from a device or server.
Count (ETX) so that routes are formed in the DODAG. RPL supports both uni-
directional traffic towards DODAG root and bidirectional traffic between smart
devices and the root. Each node has an IPv6 address which is used as node’s
ID. Nodes keep a list of their DODAG neighbours and they have one or more
parents, except for the root. Moreover, RPL nodes have a rank indicating their
position with respect to the root, who has the lowest rank. The direction from
root to other nodes is called downward route, while the direction from other
nodes to root is called upward route.
RPL introduces new ICMPv6 control messages. The root begins by sending
out DODAG Information Option (DIO) messages. These messages are used for
keeping the DODAG and include information about the rank of the broadcasting
node (which is the distance of the node from the backbone network), the OF,
and the DODAG ID. As soon as a node gets a DIO message, it determines its
rank (according to the rank in the received message) and the cost of getting to
the node from itself. DIO messages are sent regularly from each node, triggered
by trickle timer [21]. Using a timer reduces the redundant transmissions of DIO
messages.
When a new node joins the network, it either waits for a DIO message or it
may multicast a DODAG Information Solicitation (DIS) message to request a
DIO message. Other nodes listening to the DIS message, begin sending DIO mes-
sages, and the new node can join the DODAG. Then, the new node also sends the
Destination Advertisement Option (DAO) message to its parent; parent nodes
may also request DAO messages by sending a DIO message to sub-DODAG.
DAO messages are used for advertising information needed to create downward
routes. Receivers of DAO messages update their routing table.
The RPL-specific DIS attack may be launched when a malicious RPL node
creates massive amount of traffic by sending DIS messages to other RPL nodes,
causing the recipient nodes to respond by sending DIO messages. As a result,
congestion is created in the network and nodes are energy exhausted.
3 Related Work
IoT devices have limited computation power, different network structures, and
various communication protocols. These features introduce new challenges that
should be addressed by an IoT-based IDS [44]. Below we briefly describe recent
works in the field of DoS attacks and the most important IDS solutions for IoT.
Islam et al. [17] discuss the security issues for IoT devices in the healthcare
domain. Among the different types of DoS attacks, the authors analyse the
permanent DoS attack in which the device’s functionality is permanently affected
due to the execution of specific Linux commands. BrickerBot is the name of such
a bot which spreads the malicious code [12].
Regarding DDoS attacks, many works exist in the literature. Kolias et al.
[19] analysed the Mirai botnet and tried to extract some attack vectors for using
them in detecting vulnerabilities of IoT devices. In another work, Lyu et al. [24]
evaluate the capability of eight household IoT devices to participate in reflective
378 P. P. Ioulianou and V. G. Vassilakis
DDoS attacks. Results show that devices can amplify an attack by a factor of
20 to inflict 1.2 Mbps of malicious traffic on a victim.
Several DDoS protection frameworks exist in the literature. A multi-level
DDoS detection framework was proposed in [42]. Authors use traditional mech-
anisms in different levels including fog computing, edge computing and cloud
computing to defend against DDoS attacks. Authors in [8] suggest an entropy-
based DDoS detection. This approach is one of the most effective in the literature
because of the low computing overhead.
Regarding password guessing attacks, it is increasingly popular for attackers
to compromise IoT devices by guessing default/weak passwords. Several works
such as [38–40] study password policies of various services. All highlight the huge
variety of password specifications and password management that the services
are implementing.
Raoof et al. [32] present a comprehensive study of RPL attacks. They classify
RPL attacks into WSN-inherited and RPL-specific attacks. They also discuss and
classify the latest IDS solutions for RPL-based networks. Authors report that
although some IoT-based IDSes exist, RPL-specific attacks such as DIS attack
have no dedicated mitigation method until today.
A recent work by Muna et al. [28] focuses on detecting malicious activities
in industrial IoT. They propose an anomaly-based IDS, which learns the nor-
mal behaviour of the TCP/IP traffic and automatically detects any abnormal
behaviour.
Kalis [26] is one of the first developed IoT-based IDSes that aims at protecting
smart devices irrespective of the IoT protocol or application used. Kalis is a
network-based, hybrid signature/anomaly-based, hybrid centralized/distributed,
online IDS. Kalis attempts to prevent DoS attacks based on the current network
topology, traffic analysis, and mobility information. Experimental results show
that Kalis has better detection performance in IoT environments compared to
the traditional IDSes. Another remarkable work in the field is the SVELTE
IDS [33]. This is a signature- and anomaly-based IDS, developed to protect IoT
devices from routing attacks based on RPL. Some of the considered attacks
include altering information, sinkhole forwarding, and selective forwarding.
In our previous works [15,16] we implemented two variants of DoS attacks,
namely DIS attack and version number modification. Results depicted that
these attacks negatively impact devices’ power consumption. We also provided a
high-level design of signature-based IDS for protecting IoT networks from these
attacks.
Despite good progress in developing IoT-based IDSes, current solutions have
several limitations. Kalis, for example, requires installation of specialised detec-
tion modules for detecting each type of attack. This could create a complex
network and could lead in poor detection performance. Moreover, Kalis relies on
WiFi. This means that interference between the smart sensors and Kalis nodes
is possible if they are in close proximity. SVELTE has also some limitations as
it requires the modification of sensors’ software. This, however, would be very
DoS Attacks and Countermeasures in RPL-Based IoT 379
inconvenient for networks with large numbers of sensors, which is a typical case
in many IoT application domains.
All in all, a new technologically improved solution is needed to protect IoT
networks from a wide range of possible attacks. The aforementioned limitations
have been taken into account when designing our proposed IDS solution.
4.1 Scenarios
Two main scenarios were simulated, normal and malicious. Initially, our aim is
to examine the behaviour of sensors in an environment without a compromised
node. Next, this behaviour will be compared to a scenario where one or more
nodes are compromised. The obtained results have been used to design and
implement an IDS detector for identifying malicious nodes and stopping the
DIS attack. The application used in IoT nodes/sensors is based on the UDP
client-server model. The hardware used for each node is Zolertia Z1 [3] running
ContikiOS.
In the normal scenario, each node regularly sends 1 DIS message per minute.
This is the default sending rate implemented in ContikiOS and was also used
by Le et al. [20]. In contrast, the malicious scenario has one or more nodes
malicious/compromised and have been modified to send 10 DIS messages per
second. This was implemented to demonstrate DIS attack which causes IoT
nodes to perform unnecessary computations and consume energy.
Each scenario uses a variable number of nodes. Table 2 describes the differ-
ent node types and the configuration used for compiling node’s firmware. The
number of nodes used in each scenario is depicted in Table 1. As it is indicated,
malicious motes are used only in the malicious scenario, while other nodes are
used in both scenarios.
Overall, we performed 7 and 6 simulations for the normal and malicious
scenarios, respectively.
number of other messages which shows the average number of DIO, DAO and
UDP messages in each simulation. Based on the simulation results, the normal
scenario starts with a very low number of DIO and DAO messages, while UDP
messages constitute the majority. This is an expected behaviour because nodes
form the network and communicate with root by sending UDP packets. On the
other hand, the number of other messages, especially DIO, rises dramatically if
malicious node exist. The reason is that nodes respond with DIO when malicious
nodes send DIS messages. As a consequence, malicious nodes affect the network
performance and the operation of benign nodes.
Other compromised devices may exchange traffic locally. Our design considers
all types of communications so that malicious nodes can be blocked. The BR
captures traffic from both WiFi and IEEE 802.15.4 interfaces and can detect
attacks from Zigbee/6LoWPAN devices.
In that case, a new rule with the node’s IP is created and the node cannot send
or receive data from the Internet.
As far as the placement strategy of IDS modules is concerned, a hybrid
approach has been adopted. The centralized node (i.e, BR) analyzes traffic,
and detects attacks originating from the sensors or coming from the Internet.
The decentralized nodes (i.e., IDS detectors), perform lightweight tasks such as
monitoring and reporting network data to the BR. This placement strategy helps
in capturing traffic and detecting attacks from all network segments.
6 IDS Evaluation
In this subsection, our aim is to evaluate the implemented IDS in terms of
detection rate and scalability. Moreover, simulations will help us to find out
which IDS configuration is the best for optimized performance.
– True positive (TP) rate: Percentage of malicious nodes that are correctly
detected as malicious.
– False positive (FP) rate: Percentage of normal nodes that are incorrectly
detected as malicious.
– IDS warnings: Indicates how many times the IDS generated a warning for a
malicious node.
– Messages sent to IDS root: Indicates the number of messages sent to IDS root
by IDS detectors.
Each metric is calculated after taking into account results from all the repe-
titions of individual scenario.
Evaluating an IDS requires testing various topologies. The IDS should be able to
detect compromised nodes in different environments. Therefore, scenarios were
created assuming that attackers target several edge nodes. DIS attack was con-
figured to be launched by compromised nodes after 30 s so that the network
is properly formed. In other words, compromised nodes attack the network by
sending 10 DIS messages every 30 s. A large network with 30 benign nodes,
a variable number of IDS detectors, 6 malicious nodes and one IDS root was
created. The number of malicious nodes remains fixed, whereas the number of
386 P. P. Ioulianou and V. G. Vassilakis
IDS detectors varies. Using a large number of sensors helps in evaluating the
scalability of proposed IDS. Sensors were deployed throughout the network as
depicted in Fig. 1. IDS detectors increase by one in each scenario. We created
10 scenarios that are repeated 5 times each for cross-validation. Each scenario
is simulated for 25 min.
Fig. 1. Topology used in IDS simulations. Scenarios with 1, 5 and IDS detectors shown.
They increase up to 10. Colours for node types: green = server, yellow = benign node,
purple= malicious node, orange = detector (Color figure online)
6.4 Results
Results from 5 repetitions of each simulation are very encouraging. Figure 2(a)
illustrates the TP and FP rates using different IDS detectors. As it is expected,
the TP rate is 100% if 3 or more detectors are used in the network. This percent-
age falls to 97% and 83% when detectors are one and two respectively. However,
FP rate increases if detectors are 8 or more. Therefore, this metric suggests using
between 3 and 7 including IDS detectors.
Figure 2(b) shows the number of warnings generated by IDS root. These
warnings are the output of the detection module algorithms. As it is depicted,
the number is below 2,500 when IDS detectors are less than 3. However, this
number is rocketed to over 7,000 when 3 or more detectors are deployed. The
number remains at similar levels when detectors are between 5 and 8, with a
small increase when 9 or more detectors exist in the network.
DoS Attacks and Countermeasures in RPL-Based IoT 387
Fig. 2. (a) TP and FP rates for each scenario, (b) Number of warnings generated by
IDS root
In conclusion, the IDS achieves high detection rate in almost all cases. This
means that all 6 malicious nodes are detected even in large networks. Further-
more, results suggest that more than 3 and less than 8 IDS detectors should be
deployed for best performance and low overhead. However, this could be different
if more compromised nodes exist.
In this work, we study the effects of IoT-specific DoS attack, called DIS attack,
in RPL-based networks and evaluate a proposed lightweight IDS for protecting
the IoT network and devices. Firstly, we simulated sensor nodes in Cooja which
supports application development for ContikiOS. Simulations were performed for
studying the characteristics of normal and malicious IoT environments. In attack
scenarios, compromised nodes perform DoS attack that relies on RPL control
messages. The attack introduces a large number of packets in the network that
388 P. P. Ioulianou and V. G. Vassilakis
constitute some nodes unreachable and may negatively impact their power con-
sumption. Based on these results we develop detection algorithms and implement
an IDS. The proposed system involves both centralised and distributed modules
for detecting intrusions originating from external networks as well as from inter-
nal compromised nodes. Experimental results showed that high detection rate
can be achieved if IDS detectors are 3 or more. However, the more the detectors,
the higher the overhead introduced in the network.
In our future work, we plan to minimize the introduced overhead and test the
proposed architecture in a real-world IoT environment. This will be achieved by
importing the IDS modules to ContikiOS devices. Also, we will implement more
types of routing attacks such as selective forwarding and rank attack. Mitigation
techniques will be developed to detect and prevent new attacks in IoT networks.
References
1. Contiki: The Open Source OS for the Internet of Things. http://www.contiki-os.
org/. Accessed 13 Aug 2019
2. TinyOS: An OS for Embedded, Wireless Devices. https://github.com/tinyos/
tinyos-main. Accessed 13 Aug 2019
3. Zolertia technical documentation (2017). https://github.com/Zolertia/Resources/
wiki/Zolertia-Technical-documentation
4. McAfee Labs Threats Report, September 2018. https://www.mcafee.com/
enterprise/en-us/assets/reports/rp-quarterly-threats-sep-2018.pdf
5. Alaba, F.A., Othman, M., Hashem, I.A.T., Alotaibi, F.: Internet of Things security:
a survey. J. Netw. Comput. Appl. 88, 10–28 (2017). https://doi.org/10.1016/j.jnca.
2017.04.002
6. Alohali, B.A., Vassilakis, V.G., Moscholios, I.D., Logothetis, M.D.: A secure scheme
for group communication of wireless IoT devices. In: Proceedings 11th IEEE/IET
International Symposium on Communication Systems, Networks, and Digital Sig-
nal Processing (CSNDSP), Budapest, Hungary, pp. 1–6, July 2018. https://doi.
org/10.1109/csndsp.2018.8471871
7. Ammar, M., Russello, G., Crispo, B.: Internet of things: a survey on the security
of IoT frameworks. J. Inform. Secur. Appl. 38, 8–27 (2018). https://doi.org/10.
1016/j.jisa.2017.11.002
8. Bhuyan, M.H., Bhattacharyya, D., Kalita, J.K.: An empirical evaluation of infor-
mation metrics for low-rate and high-rate ddos attack detection. Pattern Recogn.
Lett. 51, 1–7 (2015)
9. CISCO: The Internet of Things Reference Model (2014). http://cdn.iotwf.com/
resources/71/IoT Reference Model White Paper June 4 2014.pdf
10. Dunkels, A., Gronvall, B., Voigt, T.: Contiki - a lightweight and flexible operating
system for tiny networked sensors. In: Proceedings 29th IEEE International Con-
ference on Local Computer Networks, Tampa, FL, USA, pp. 455–462, November
2004. https://doi.org/10.1109/lcn.2004.38
11. Easton, T.: Chalubo botnet wants to DDoS from your server or IoT device, Octo-
ber 2018. https://news.sophos.com/en-us/2018/10/22/chalubo-botnet-wants-to-
ddos-from-your-server-or-iot-device/
12. Geenens, P.: BrickerBot - The Dark Knight of IoT (2017). https://blog.radware.
com/security/2017/04/brickerbot-dark-knight-iot/
DoS Attacks and Countermeasures in RPL-Based IoT 389
30. Osterlind, F., et al.: Cross-level sensor network simulation with COOJA. In: Pro-
ceedings of 31st IEEE International Conferene on Local Computer Networks,
Tampa, FL, USA, pp. 641–648, November 2006. https://doi.org/10.1109/lcn.2006.
322172
31. Pongle, P., Chavan, G.: A survey: attacks on RPL and 6LoWPAN in IoT. In:
Proceedings of International Conference on Pervasive Computing (ICPC), pp. 1–
6. IEEE (2015). https://doi.org/10.1109/pervasive.2015.7087034
32. Raoof, A., Matrawy, A., Lung, C.H.: Routing attacks and mitigation methods for
RPL-based internet of things. IEEE Commun. Surv. Tutor. 21, 1582–1606 (2018)
33. Raza, S., Wallgren, L., Voigt, T.: SVELTE: real-time intrusion detection in the
Internet of Things. Ad Hoc Netw. 11(8), 2661–2674 (2013). https://doi.org/10.
1016/j.adhoc.2013.04.014
34. Rghioui, A., Khannous, A., Bouhorma, M.: Denial-of-service attacks on
6LoWPAN-RPL networks: threats and an intrusion detection system proposition.
J. Adv. Comput. Sci. Technol. 3(2), 143–153 (2014). https://doi.org/10.14419/
jacst.v3i2.3321
35. Symantec Security Response: Mirai: what you need to know about the botnet
behind recent major DDoS attacks, October 2016
36. Tsiftes, N., Eriksson, J., Dunkels, A.: Low-power wireless IPv6 routing with Con-
tikiRPL. In: Proceedings of 9th ACM/IEEE International Conference on Informa-
tion Processing in Sensor Networks, pp. 406–407 (2010)
37. Wallgren, L., Raza, S., Voigt, T.: Routing attacks and countermeasures in the RPL-
based Internet of things. Int. J. Distrib. Sens. Netw. 9(8), 1–11 (2013). https://
doi.org/10.1155/2013/794326
38. Wang, D., Cheng, H., Wang, P., Huang, X., Jian, G.: Zipf’s law in passwords. IEEE
Trans. Inf. Forensics Secur. 12(11), 2776–2791 (2017)
39. Wang, D., Wang, P.: The emperor’s new password creation policies. In: Pernul, G.,
Ryan, P.Y.A., Weippl, E. (eds.) ESORICS 2015. LNCS, vol. 9327, pp. 456–477.
Springer, Heidelberg (2015). https://doi.org/10.1007/978-3-319-24177-7 23
40. Wang, D., Zhang, Z., Wang, P., Yan, J., Huang, X.: Targeted online password guess-
ing: an underestimated threat. In: Proceedings of the 2016 ACM SIGSAC Confer-
ence on Computer and Communications Security, pp. 1242–1254. ACM (2016)
41. Winter, T., et al.: RPL: IPv6 routing protocol for low-power and lossy networks.
RFC 6550, March 2012
42. Yan, Q., Huang, W., Luo, X., Gong, Q., Yu, F.R.: A multi-level DDoS mitigation
framework for the industrial internet of things. IEEE Commun. Mag. 56(2), 30–36
(2018)
43. Yang, Y., et al.: A survey on security and privacy issues in internet-of-things.
IEEE Internet Things J. 4(5), 1250–1258 (2017). https://doi.org/10.1109/jiot.
2017.2694844
44. Zarpelão, B.B., Miani, R.S., Kawakani, C.T., de Alvarenga, S.C.: A survey of
intrusion detection in internet of things. J. Netw. Comput. Appl. 84, 25–37 (2017).
https://doi.org/10.1016/j.jnca.2017.02.009
Study of DNS Rebinding Attacks
on Smart Home Devices
1 Introduction
The spread of smart appliances leads to increase networking between the devices
themselves and thus to smart homes. For attackers, this development represents
an increased attack surface. In particular, devices accessible via the Internet
are attractive targets. Consumers may assume that if smart home devices are
reachable locally only, they pose no risk to the home network. However, with DNS
rebinding attacks, it is possible to communicate with only internally accessible
devices. DNS rebinding allows unauthorized access to private networks.
DNS rebinding attacks are known since 1996 [5,6]. As a result, various attack
methods, as well as countermeasures, were already published [4,9]. However, a
recent study from 2018 demonstrates that DNS rebinding attacks are still feasi-
ble today [1]. Even the assigned CVEs (a total of 25) indicate that DNS rebind-
ing, since 2017 (11 out of 25 CVEs), experiences a revival [3]. This observation
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 391–401, 2020.
https://doi.org/10.1007/978-3-030-42048-2_25
392 D. Tatang et al.
correlates with the increasing number of Internet of Things (IoT) devices. World-
wide, approximately half a billion devices are estimated to be vulnerable to DNS
rebinding attacks in 2018 [2].
In this paper, we investigate DNS rebinding attacks on smart household
appliances, a subset of the IoT. We show that sensitive data can be extracted
and remote control from the Internet is possible. We analyze the execution of the
attacks systematically and summarize them. In addition, we investigate which
prerequisites must be fulfilled in order to carry out the attacks. In this way, we
discuss how serious the risk is for smart home device owners to become victims.
Finally, we analyze the top 100 Alexa web pages that communicate over HTTP
to investigate whether DNS rebinding is performed unnoticed on one of these
pages. It was demonstrated that protection mechanisms of dnsmasq do not detect
attacks on localhost (127.0.0.1), 4 out of 5 investigated devices are vulnerable,
and none of the top 100 sites performs DNS rebinding attacks.
To summarize, we make the following contributions:
– We systematically analyze DNS rebinding attacks on four smart home devices
and summarize our results in an overview.
– We investigate requirements for successfully DNS rebinding attacks and dis-
cuss the risk of becoming a victim of such an attack.
– We present a brief measurement study on the execution of DNS rebinding on
popular websites.
In the remainder of the paper, we first introduce basic knowledge and identify
requirements for successfully DNS rebinding attacks in Sect. 2. Afterwards, we
describe our conducted experiments in Sect. 3, followed by presenting the results
in Sect. 4. In Sect. 5, we discuss the results and limitations. Section 6 presents
some related work and we conclude our work in Sect. 7.
Fig. 1. High-level concept of DNS Rebinding attacks. The browser becomes by manip-
ulating the mapping of hostnames and IP addresses to the proxy into the internal
network (see ①). Direct access is blocked by the router firewall (see ②).
an IP address, he also receives the corresponding DNS request, which can then
be manipulated. The browser trusts the DNS response, and thus the connection
to a local network device can be established by the manipulated domain name
and IP address assignment. An attacker does not have to compromise a DNS
server; it is sufficient to generate valid DNS replies for requests to resolve his/her
domain. Note, DNSSec is not able to prevent this attack scenario because the
attacker only generates valid DNS responses to queries for his/her domain.
server provides a resource record set with IP addresses. The first entry in this
record set must contain the internal IP address of the target device in the local
network of the victim. The second IP address is identical to the IP of the attacker
server. In this way, an external attacker can use a Java applet to implement inter-
active access to devices in the local network of the victim. It is exploited that a
connection request is allowed by the Java system as soon as the IP address from
which the Java applet was loaded appears in the resource record.
However, this attack is no longer feasible as DNS pinning has been intro-
duced and the security policies for Java applets have been changed. Nowadays,
an applet can only establish connections to the IP address from which it was
originally loaded (Same Origin Policy) [8].
Attacker DNS
server
1 2 5 6
3
7 Attacker
Target device Victim
website
4
Local network
Fig. 2. Steps of a time varying DNS rebinding attack. The DNS request of the victim
① is answered by the attacker DNS server with his IP address and a short TTL ②.
The victim’s browser downloads malicious code from the attacker’s website ③ and ④.
When the malicious code is executed, the TTL has expired so that a new DNS request
must be sent to the attacker’s server ⑤. The response contains the IP address of the
target device in the victim’s local network ⑥. Thus, the request is redirected to the
target device ⑦.
the domain and IP address of the attacker is already deleted due to the very
short TTL. Thus, to resolve the domain, a new DNS request must be triggered
that the DNS server of the attacker responds with the private IP address of
the target device in the network of the victim. In this way, the domain of the
attacker in the DNS cache of the browser is assigned to the private IP address
of the device by the victim. As a result, the asynchronous connection request is
not sent to the Web server of the attacker, but the local network component of
the victim. Thus, the attacker succeeds in establishing an interactive session to a
device in the private network of the victim. The attack detection of routers can
easily detect this attack method when a private IP address is resolved according
to RFC1918 [13]. DNS pinning also prevents the attack in modern browsers. As
a result, this simple attack is no longer exploitable.
The simple time-varying attack described above can be blocked by DNS pin-
ning in the browser. The most straightforward strategy to bypass DNS pinning
is to make the malicious script wait with the asynchronous connection request
until the DNS entry expires in the cache of the browser. This trivial approach
is called anti-DNS pinning [7]. In 2013, Dai and Resig showed that it is possible
to significantly speed up the waiting period by flooding the DNS cache [4].
2.3 Countermeasures
DNS rebinding attacks have been known for a long time and so there are func-
tioning countermeasures. On the server side, every web server in the local net-
work can have its own authentication methods. Furthermore, communication
with the web server should be secured by TLS; thus, no DNS rebinding attack
is possible. The firewall settings should be such that requests from external host
names must not be resolved with internal IP addresses. The DNS settings should
also be configured so that external hostnames cannot be resolved with internal
IP addresses. This adjustment is straightforward to do by using DNS rebinding
protection mechanisms on many routers, e.g., dnsmasq uses this protection in
the default settings. On the client side, browser extensions such as NoScript can
be used when visiting web pages.
2.4 Requirements
In order to successfully perform a DNS rebinding attack today, several require-
ments must be fulfilled. We identified a total of six requirements and describe
them in the following.
1. No transport layer security (TLS): If TLS is used, no DNS rebinding attack
can be performed. This is because a TLS certificate is issued to a full hostname
or a unique IP address. When a new connection request is made to the target
device to perform a DNS rebinding attack, the TLS certificate verifies that
the domain of the new connection matches the information of the certificate.
Since the domain of the local target device differs from the information in
the TLS certificate the TLS handshake fails and the connection request is
rejected.
396 D. Tatang et al.
3 Descriptions of Experiments
In the following, we describe the experiments we conducted as part of our study.
We start with the description of the Attacker Model, continue with the setup,
and finally our measurement.
examined in the local network. We are examining a Google Home Mini, a Sonoff
Basic, a Foscam IP Camera, and a Bose Soundtouch.
The Google Home Mini is a voice-controlled speaker, i.e., a smart assistant
who is representative for other voice assistants in our study, such as the Alexa
speakers. The Sonoff Basic is a smart switch that can be controlled via WLAN.
The switch is flashed with the Tasmota firmware [16] and therefore representative
for all devices with the Tasmota firmware. The Foscam IP camera is an IP camera
with the exact designation FI9900p, and the Bose Soundtouch 10 is a WLAN
controllable speaker.
We used the DNS rebinding attack framework Singularity of Origin of the
NCC Group for the execution of the attacks [15]. This framework performs a
time-varying DNS rebinding attack as already introduced in Sect. 2.2. As part of
the work, we have accelerated an attack method of the framework by applying
the DNS cache flooding technique presented by Dai and Resig [4] to the method
used in the framework. In this way, the duration of the attack could be reduced
from 60 s to 5 s.
Sonoff Switch
(192.168.1.106)
Internet
User Router
Computer Aack Server
(192.168.1.181)
Bose Soundtouch 10
(192.168.1.131)
Foscam IP-Camera
(192.168.1.173)
To complement our study, we examine websites and check whether they per-
form DNS rebinding attacks in the background. We performed this measurement
398 D. Tatang et al.
by setting up a Ubuntu system and installing the DNS server dnsmasq. With
DNS Rebinding Attack Protection enabled by default, dnsmasq reliably detects
whether private IP addresses are contained in DNS packets. Subsequently, we
implemented a script that automatically visits websites and evaluates the log file
from dnsmasq to detect DNS rebinding attempts.
4 Results
In this section, we present the results of our investigations. First, we describe
the results of our conducted attacks on smart home devices. Second, we present
the results of our brief measurement of DNS rebinding on popular websites.
Table 1. Overview of the results of the conducted DNS rebinding attacks against four
different smart home devices. Three out of four devices were attacked successfully.
Sonoff Basic with the Tasmota firmware and default configurations is also vul-
nerable to DNS rebinding attacks. The firmware uses a Web API interface.
Therefore all Sonoff devices with the Tasmota firmware are potentially vulnera-
ble. The active control of hardware functions is limited to toggling the relay of
the switch. However, sensitive data such as usage habits and power consumption
are readable.
Bose Soundtouch also has a Web API interface without authentication mech-
anisms and no TLS support. Thus, this device is also potentially vulnerable
to DNS rebinding attacks. It provides access to hardware functions such as vol-
ume, as well as stored data such as MAC addresses of paired devices. Other Bose
devices also use the same firmware. Therefore, we suspect that these devices are
vulnerable as well.
Foscam FI9900p is the only tested device not potentially easily vulnerable to
DNS rebinding attacks. When setting up the device, the user is forced by the
vendor to set a username and password. When connecting to the web service of
the camera, the user has to authenticate with his credentials. For this reason, a
DNS rebinding attack is only possible if the attacker knows the credentials or is
able to break the authentication.
This measurement led to the result that none of the top 100 Alexa sites that use
HTTP execute DNS rebinding and is thus not surprising. However, the approach
can be used to check a more significant number of websites.
5 Discussion
The results demonstrate that four out of five examined devices are vulnerable to
DNS rebinding attacks. However, it should be noted that for the selection of the
test devices, devices with open Web services were explicitly selected. Accordingly,
this selection of test devices cannot be used to make a statement about all smart
home devices. However, since the total number of smart home devices is very
high, the number of potentially vulnerable devices should still be a non-negligible
amount. In addition, our insights also confirm the results of Acar et al. [1] that
many IoT devices are vulnerable to DNS rebinding attacks.
Furthermore, we have seen that for a successful DNS rebinding attack, many
requirements have to fit, which limits it as a real-world threat. However, as soon
400 D. Tatang et al.
as the attack is feasible, it can have serious consequences. For this reason, it
is important to check which conditions have to be fulfilled and to evaluate the
applied countermeasures. The results of the study indicate that manufacturers
do not focus sufficiently on the security of their products when developing them.
The potential vulnerability of smart home devices highlights the lack of security
of IoT devices, which has repeatedly attracted media attention in recent years. In
many cases, inadequate authentication was the cause of attacks. The well-known
Mirai botnet [12], for example, targeted IoT devices that were operated with
standard credentials and accessible to the public. Manufacturers of smart home
devices could protect their customers by introducing mandatory authentication
at the application layer. A mandatory change to the authentication credentials
was implemented for the Foscam IP camera, for example, so that no unauthorized
access to web services can be made.
6 Related Work
Since the attack technique is long-established, there is much work in this area.
The first publication on DNS rebinding attacks was published in 1996 [5], after
which further papers were presenting new variants of the attack [4,9,11]. Corre-
sponding work with countermeasures also exists [8,10]. One of the most recent
papers deals with DNS rebinding attacks on IoT devices [1]. Especially the anal-
ysis of new as well as already known attacks on IoT devices is important, as the
Mirai Botnet demonstrates [12].
We follow the course of the history of work on DNS rebinding and rely in
particular on the current paper in the field of IoT devices. We perform the
attack on smart home devices and give a rough overview of vulnerable devices.
In addition, we discuss the attack surface and conclude that due to various
countermeasures, despite the vulnerability of the devices themselves to DNS
rebinding attacks, it is not likely to become a victim.
7 Conclusion
We showed that DNS rebinding attacks in the world of IoT are reviving and
four out of five devices tested are vulnerable (three out of four smart home
devices and a router). Nevertheless, possible attack targets are limited, as many
countermeasures exist and several conditions must be fulfilled and as routers
often already contain DNS rebinding attack detection. The manufacturers of
smart home devices, therefore, rely on the router firewalls to provide security
against this attack. In many cases, this will also be the case, but ideally, the
smart home devices themselves should also be protected. In summary, this work
suggests that when connecting all things, one needs to keep in mind known
weaknesses and issues in order not to become vulnerable to attacks that have
been known for years.
Future work in the DNS rebinding attacks IoT area can extend the results of
our study by testing further smart home devices and also check a more significant
Study of DNS Rebinding Attacks on Smart Home Devices 401
number of websites to see whether a DNS rebinding attack is being carried out
in the wild.
Acknowledgment. We would like to thank the anonymous reviewers for their valu-
able feedback.
References
1. Acar, G., Huang, D.Y., Li, F., Narayanan, A., Feamster, N.: Web-based attacks to
discover and control local IoT devices. In: Proceedings of the 2018 Workshop on
IoT Security and Privacy (2018)
2. DNS Rebinding Exposes Half a Billion Devices in the Enterprise. https://armis.
com/dns-rebinding-exposes-half-a-billion-iot-devices-in-the-enterprise/. Accessed
06 June 2019
3. CVE - Common Vulnerabilities and Exposures. https://cve.mitre.org/cgi-bin/
cvekey.cgi?keyword=DNS+Rebinding. Accessed 06 June 2019
4. Dai , Y., Resig, R.: FireDrill: interactive {DNS} rebinding. In: 7th {USENIX}
Workshop on Offensive Technologies (2013)
5. Dean, D., Felten, E.W., Wallach, D. S.: Java security: From HotJava to Netscape
and beyond. In: IEEE Symposium on Security and Privacy (1996)
6. DNS Attack Scenario, February 1996. http://sip.cs.princeton.edu/news/dns-
scenario.html. Accessed 06 June 2019
7. Grossman, J., Fogie, S., Hansen, R., Rager, A., Petkov, P.D.: XSS Attacks: Cross
Site Scripting Exploits and Defense. Syngress (2007)
8. Jackson, C., Barth, A., Bortz, A., Shao, W., Boneh, D.: Protecting browsers from
DNS rebinding attacks. In: ACM Conference on Computer and Communications
Security (CCS) (2007)
9. Johns, M., Lekies, S., Stock, B.: Eradicating DNS rebinding with the extended
same-origin policy. In: USENIX Security Symposium (2013)
10. Johns, M., Winter, J.: Protecting the intranet against “JavaScript malware” and
related attacks. In: M. Hämmerli, B., Sommer, R. (eds.) DIMVA 2007. LNCS, vol.
4579, pp. 40–59. Springer, Heidelberg (2007). https://doi.org/10.1007/978-3-540-
73614-1 3
11. Karlof, C., Shankar, U., Tygar, J.D., Wagner, D.: Dynamic pharming attacks and
locked same-origin policies for web browsers. In: ACM Conference on Computer
and Communications Security (CCS) (2007)
12. Kolias, C., Kambourakis, G., Stavrou, A., Voas, J.: DDoS in the IoT: Mirai and
other botnets. Computer 50(7), 80–84 (2017)
13. Rekhter, Y., Moskowitz, B., Karrenberg, D., de Groot, G., Lear, E.: Address allo-
cation for private internets. RFC 1918, RFC Editor, February 1996
14. Roskind, J.: Attacks against the netscape browser. In: Talk at the RSA Conference
(2001)
15. Singularity of Origin. https://github.com/nccgroup/singularity. Accessed 06 June
2019
16. Fonoff-Tasmota. https://github.com/arendst/Sonoff-Tasmota. Accessed 06 June
2019
17. Tatang, D., Schneider, C., Holz, T.: Large-scale Analysis of Infrastructure-leaking
DNS Servers. In: Perdisci, R., Maurice, C., Giacinto, G., Almgren, M. (eds.)
DIMVA 2019. LNCS, vol. 11543, pp. 353–373. Springer, Cham (2019). https://
doi.org/10.1007/978-3-030-22038-9 17
Anomaly Detection in the HVAC System
Operation by a RadViz Based
Visualization-Driven Approach
Abstract. The appearance of the smart houses, buildings, and cities has
defined new attack scenarios targeting industrial information systems.
The paper suggests a visualization-driven approach to the analysis of the
data from heating, ventilating and conditioning system (HVAC). The key
element of the approach is the RadViz visualization that is used to form
daily operation patterns and can detect suspicious deviations that could
be the signs of fraudulent activity in the system. It is supplemented by a
matrix-based representation of the HVAC parameters that is constructed
in the way that allows highlighting changes in values of parameters being
analyzed. The distinctive feature of the proposed visualization models is
the ability to display data from different data sources. To demonstrate
and evaluate the efficiency of the proposed approach we used the VAST
MiniChallenge-2 2016 data set that contains logs from the HVAC system
and the access control system.
1 Introduction
The IoT technology has provided a convenient platform for constructing smart
houses, buildings, and cities. The application of the interconnected distributed
devices with embedded data mining algorithms allows one to create the intel-
ligent resource efficient control systems that provide conditions comfortable for
the habitants’ living and suitable for the equipment functioning.
The smart heating, ventilation and air conditioning (HVAC) system is an
important part of the smart building, it monitors outside and inside temperature,
humidity, electricity or gas consumption to fit better the needs of the buildings
usage or its occupants’ behavior.
However, the usage of the meters and other HVAC devices connected to
the Internet for control and analytics purposes has defined new attack vectors
against the industrial information systems [1].
The compromise of the HVAC system may result in failure of the temper-
ature sensitive equipment and degradation of the life quality of the building
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 402–418, 2020.
https://doi.org/10.1007/978-3-030-42048-2_26
Anomaly Detection in the HVAC System Operation 403
2 Related Work
Visual analytics techniques allows one to transform enormous volumes of hetero-
geneous and noisy security data into a valuable evidence of the fraudulent activity
and cyber crime in an intuitively clear form and do not require explicit applica-
tion of complex mathematical and statistical methods [10,11]. Due to the het-
erogeneity of the security data, many sophisticated visualization and interaction
techniques supporting the analysis of security incidents have been proposed [10].
To reveal connections between security events both complicated compound
visualization models requiring additional learning [12–15] and rather simple easy
interpretable visualization models [16–18] are suggested. For example, in [12] the
authors propose a compound visualization model consisting of several graphical
elements to analyze events from IDS, firewall, operating system and other avail-
able security sensors. The linked graph displays network topology and is placed
in the circle; the circle is the second graphical element and represents a ring of
security events that are arranged along the ring according to their type. The
sources of the events are linked to the corresponding segment of the ring. The
histogram located along the inner circumference of the ring is used to display the
distribution of the events of the given type in time, and the outer circumference
is used to display the overall number of the events per time unit. In [16] authors
propose to correlate data from different sources using standard linear graphs.
The events from different sources are combined on the basis either time scale or
selected nominal attribute.
Mansmann et al. [19] adopt a RadViz visualization technique to analyze
the behavior of the network hosts. The distribution of the network protocols
in the host network traffic determines the position of nodes on the chart. This
representation allows discovering anomalies in the behavior of hosts or higher
level network entities. In [20] authors apply the RadViz to analyze transactions
in the mobile money transfer services and demonstrate that it is helpful to reveal
different fraudulent activities including mobile botnet infection.
In [11] authors discuss and give recommendations how to choose a visualiza-
tion technique appropriate to the data type being analyzed and forensics task
being solved.
There is not much research devoted to the visual analytics for efficient smart
building management and investigation of anomalous deviations in its function-
ing. The most of visual analytic techniques is focused on the problem of the
energy consumption. For example, in [5] the authors present a map-based visu-
alization technique to analyze the energy consumption, they map the energy use
to the building elements (rooms, elevators) and color different building zones
according to the level of the energy consumption. Palm and Ellegård [6] adopt
the stacked bar chart with the time axis to display how different entities con-
sume energy for different everyday activities such as cooking, watching TV, etc.
The periods with energy consumption are shown by the colored bar, where the
color encodes the type of the entity’s activity. Abdelalim et al. [7] propose to
visualize energy and mass flows of the HVAC system using the Sankey diagram.
The sensor data are converted into estimated energy flows for each HVAC com-
ponent, the color of the flows indicate whether it is normal or abnormal.
Anomaly Detection in the HVAC System Operation 405
3 Visualization-Driven Approach
3.1 RadViz Visualization of the HVAC Data
The goal of the HVAC system is to provide conditions comfortable for the habi-
tants and suitable for the equipment. Its functioning depends on many factors
both internal and external, and metaphorically its operation can be represented
as balancing between different external factors and tuning internal parameters in
order to achieve some equilibrium to provide conditions required. The increase
of the one factor, e.g. temperature outside the building, may require increasing
or decreasing the setting points of the HVAC sensors. If the balance breaks then
this may be a sign of physical degradation of the sensor as well as a sign of some
potentially malicious activity. Such understanding of the HVAC system led us to
the idea to present the HVAC data and the data from the access control system
using the RadViz visualization [21].
406 E. Novikova et al.
The set S of HVAC sensors form the dimension anchors, and the vector
v̄ t defines the position of the data point corresponding to the state of the system
at the time moment t on the RadViz.
To support visual exploration of the data apart from manipulating the
anchors layout, we propose to use following interaction techniques:
(1) filtering mechanism based on the analysis of the selected values of the
attributes and
(2) interactive color encoding of the temporal attributes of the data.
It is possible to set color scheme depending on the day of week or the time
intervals of the day. The latter is useful in establishing life patterns depending
on time of the day or day of the week (work day or day off).
We also provide an analyst an ability to adjust the transparency of the color,
than it is possible to produce image close to the heat map, the more saturated
the area of the RadViz the more typical behavior HVAC system exposes.
408 E. Novikova et al.
Figure 2 shows the RadViz visualization of the HVAC data characterizing the
general state of the building during 8 working days. The color scheme is set to
highlight the day functioning pattern. The blue color is used for early morning
and late evening hours (from 0 am till 6 am and from 9 pm till 12 pm); data
points in violet color correspond to the working hours (from 9 am till 6 pm)
when the majority of the employees are on their working places.
Fig. 2. The RadViz visualization of the HVAC data (Color figure online)
The points in green color characterize transitive state of the HVAC system
when the employees arrive to their working place or leave it. The color of the data
point has a non–zero transparency, thus the overlapping points located close to
each other look darker, while scattered points look much lighter. The transitive
period of the HVAC is presented by a region of points with relatively low density,
it overlaps with the region of scattered points in blue. The dense cluster of the
blue points near the green points highlighted by the circle in Fig. 2 looks rather
atypical. The location of the points allows us to suggest that the possible problem
is in the increase of the Water Heater Tank Temperature parameter.
The certain problem of the RadViz visualization is that it places objects
with comparable values of attributes to one position. Moreover, an analyst needs
to know how the parameters change, and what their values are. To solve this
problem we suggest supplementing it with matrix-based presentation and line
charts of the attributes values described in the next subsection.
Anomaly Detection in the HVAC System Operation 409
To visualize the changes in the HVAC system over the time, we used the concept
of data slices. Its key idea is to aggregate the data into data slices in the real time
mode if the changes in them do not exceed some given threshold. The average
values of the attributes are then displayed using selected visualization technique.
In our approach the state of the HVAC system is represented by a point in the
multidimensional space. Then its functioning can be presented as a trajectory
of the point. If the position of the point does not change significantly, i.e. its
deviation from initial position does not exceed the given threshold during some
period of time, it is possible to group all data points generated during this period
of time into one data slice.
The usage of this metaphoric representation of the HVAC functioning let
us to use dbscan algorithm to form data slices [26]. The dbscan algorithm is a
density-based spatial clustering technique that aggregates the data points that
are close to each other according to some distance metric. This algorithm is
designed to analyze static data, however, in our case it is necessary to consider
the time of data points generation.
The idea of the modified dbscan algorithm is described briefly as follows.
When a new data point is generated, we calculate the distances to the points
included already in the current data slice. Then, if the maximum distance is
not greater then threshold value ε, and the point timestamp does not start a
new day, the point is included in the current data slice, otherwise a new slice is
formed.
We included the condition that checks the beginning of a new day in order
lately to have a possibility to group data slices by the day, this is helpful in
revealing daily patters in the HVAC system functioning.
The modified dbscan algorithm extracts groups of data points that have two
important temporal attributes - the beginning of the data slice and its duration.
Formally, the data slice is defined as follows:
dc = t, dur, v̄ tj , tj ∈ [t, t + dur) ,
where t is a timestamp of the first data point that starts the data slice, dur is a
duration of the data slice, and the vector v̄ tj defines the values of the analyzed
parameters at the tj moment. The attribute t is used to sort data slices in order
of their appearance.
We propose to use a matrix-based representation to visualize HVAC data
slices. It is constructed in the following way. The y-axis corresponds to the ana-
lyzed parameters, while the x-axis corresponds to the time or sequence number
of the data slice. In our approach each cell of the matrix represents the average
value of the corresponding attribute in the data slice. The saturation of the color
is used to encode the values. The higher the value, the darker the matrix cell.
Figure 3 shows the scheme of matrix-based representation of the HVAC data and
data from the access control system for the one day for the meeting room.
410 E. Novikova et al.
According to it, this location was visited once approximately at 11–12 am.
The data from the access control system are fully conformed by the HVAC data.
The lights and equipment power supply were on at the same period as well as
the parameters characterizing the room ventilation increased at the same period
of time. The CO2 concentration measured at the room’s return air grille was
also slightly increased after the visit. Note, that rather long time period from 12
am till 10 pm is represented by only one column of cells. This means that there
are no significant changes in the values of the HVAC and access control system
parameters during the given period of time.
It should be noted that threshold parameter ε used to group data into the
data slices defines the level of changes that will be highlighted using the matrix
presentation. The higher its value the coarser the aggregation of the data into
data slices, the more significant changes in the HVAC data are highlighted. Thus
by manipulating the parameter ε an analyst may control the data detail level.
The obvious disadvantage of the matrix-representation is the non-linear
transformation of the displayed time scale, as the size of the matrix element
does not depend on the duration of the data slice. On one hand, this solution
highlights changes that may have very short duration and present a certain inter-
est to an analyst; but on the other hand, it may be confusing for the analyst and
requires studying the visualization model. In addition, it does not show absolute
values of the attributes the analyst may be interested in. To support analysis of
the absolute values and give understanding of the real time scale we supplement
the matrix-based representation with a set of interactive line charts with time
Anomaly Detection in the HVAC System Operation 411
Fig. 4. The RadViz visualization of the HVAC data for 10 working days (Color figure
online)
The comparison of the matrix representations of the days with and without
energy anomaly give us a clue what the anomaly took place (Fig. 5). While the
rest parameters look very similar, the total energy consumption and the HVAC
energy consumption (HVAC Electric Demand Power and Total Electric Demand
Power parameters) have a burst in the morning at 7 o’clock on the abnormal
day, and then the HVAC system energy consumption falls to almost zero.
When analyzing the rest HVAC zones we kept in mind these anomalous days
in order to determine typical and atypical behavior of the system and started the
analysis of the zones with comparing days with and without energy anomaly. It
allowed us to detect the possible cause of the anomaly:the cooling and heating
set points were set to the maximum values. For example, for the room with
server equipment these parameters are typically set to the 18.1 ◦ C and 15.5 ◦ C
correspondingly, however, during the anomalous days these parameters were set
to 26 ◦ C and 29.5 ◦ C correspondingly.
The RadViz visualization helped us to detect another interesting anomaly
for this room. The color scheme used to color the data points encoded the days
of week. After filtering out days with anomaly, we noticed that there are several
groups of points (Fig. 6).
Anomaly Detection in the HVAC System Operation 413
a)
b)
Fig. 6. The RadViz visualization of the HVAC and access control system data for the
server room
Point parameters which were set to 0. This anomaly became a reason for the
visit of the three employees simultaneously that took place next day.
Combining the RadViz and the matrix representation of the HVAC and the
assess control system we detected all anomalous situations present in the VAST
Challenge, such as increase of the temperature up to 30 ◦ C in the HVAC zone
1 on the third floor, increase of the CO2 concentration in almost all zones due
to failures in the HVAC system or wrong cooling and heating set points, the
failures of the conditioning in the zone 8 on the second floor, etc.
When analyzing data with RadViz we noticed the following. The RadViz pro-
duces rather big clusters with even density for the days with normal functioning
of the system. The anomalous days are characterized by a presence of several
very dense clusters and/or single outliers. The visual patterns for the normal
functioning of the HVAC system for the locations with specific attributes such
as offices, canteen, server room, on the contrast, are characterized by compact
dense clusters of the linear form, while anomalous days are characterized by
large dispersion of the points. This fact could be explained that the HVAC sys-
tem for the locations with specific attributes has rather strict functioning routine
depending on day time or day of week; while the general parameters HVAC such
as total energy consumption, water heater tank temperature, water heater rate,
Anomaly Detection in the HVAC System Operation 415
supply air rate, etc. depend on a set of such routines, and thus, produce “cloud”
of colored points with even density. We assume that these visual patterns pro-
duced by the RadViz visualization can be used in further research to produce
automatic assessment of the HVAC state.
To evaluate the efficiency of the proposed visualization models we presented
our approach to the 10 specialists having both practical and research experi-
ence in the information security and intrusion detection techniques. We also
invited specialist in the industrial control systems. Firstly, we presented them
our approach, described the peculiarities of the RadViz visualization technique
and explained how to interpret its results. We also discussed what the matrix-
presentation displays and how data slices are formed and, then the participants
were given a test task formed on the basis of the VAST Challenge dataset. The
test tasks included the following questions: determine the typical patterns of
the HVAC system functioning, i.e. the dependencies between its parameters, the
character of their changes, how they depend on the presence of the employees
in the controlled zone, and weather outside the building; describe the possible
anomalies in the HVAC system if present, i.e. determine the parameters with
atypical values, describe the character of the anomalous changes, determine the
possible causes of anomalous changes. The participants were also asked to rank
the visual models (from very bad to very good) and give their feedback and their
suggestions to the enhancement of the visual models.
The overall assessment of the visualization models was good. Interestingly,
that the specialists in information security and data analysis captured the idea of
visual models quickly as they are more used to different non trivial visualization
techniques, while the specialist in complex system automation firstly was skeptic,
as he is more used to traditional line charts to monitor the object state. How-
ever, after studying them and doing the test he changed his mind and proposed
to adopt the models to the real time data monitoring. He suggested presenting
typical behavior of the system as a heat map on the RadViz visualization and
mapping the current position on it to monitor the state of the HVAC system.
All participants of the evaluation process marked that the matrix-representation
was helpful in explaining what is wrong for the particular state of the HVAC sys-
tem. They highlighted the possibility to analyze the data from the access control
system in context of the HVAC, as it gives understanding on why some HVAC
parameters change and, correspondingly, the changes in HVAC parameters may
explain further behavior of the employees. They also liked the possibility to
change the color scheme according to the values of the temporal attributes of
the HVAC parameters but recommended to add a possibility to highlight data
belonging to different HVAC zones, as it enables comparing HVAC zones ser-
vicing locations with similar characteristics, i.e. offices, lifts, corridors, etc. The
specialists also advised to enhance the interaction techniques linking the Rad-
Viz and the matrix-based visualization to simplify the analysis process. Another
interesting suggestion is connected with automation of the analysis process. The
experts suggested adding possibility to mark visual patterns for typical work
days and days off and use them as a baseline to assess automatically the rest
416 E. Novikova et al.
5 Conclusions
Application of the IoT technology to construct smart homes, buildings and cities
caused the appearance of the new attacks on the industrial information systems.
The compromised HVAC system may lead to the life threatening scenarios and
degradation of the temperature and humidity sensitive equipment.
In the paper we proposed a visualization driven approach to the exploratory
analysis of the HVAC data in order to investigate different suspicious events that
could be a sign of the fraudulent activity. The key element of the approach is
the RadViz visualization that is used to reveal typical life cycle of the HVAC
system and detect suspicious deviations in its state. The RadViz allows one to
analyze parameters from the HVAC system and other systems controlling smart
building. This RadViz ability may be helpful when detecting attack scenarios
that include fabricating HVAC sensors readings.
We demonstrated that it is possible to form graphical patterns for normal
and abnormal functioning of the HVAC system.
To overcome some problems of the RadViz visualization we complemented it
with the matrix representation of the HVAC data. It is used to display changes
in the state of the system, the data points characterizing the HVAC state are
aggregated in one data slice if the state of the HVAC system remains almost
unchanged, thus, it is possible to both reduce the volumes of the data being
analyzed and attract the analyst attention to the changes in data. However, in
case of necessity the analyst can get the values of the attributes using standard
line charts.
Both visual models allow one to investigate the data from different sources,
and we illustrated this by applying our approach to the data set provided by the
VAST Challenge 2016. It contains logs from the HVAC system and the access
control system. Analyzing these data simultaneously it is possible to understand
the behavior of the HVAC system better and explain the motifs of the unusual
employees’ moves.
In the paper we discussed the result obtained, and defined future directions
of work devoted to the enhancement of the prototype and usability evaluation
of the proposed visual analytical system.
Anomaly Detection in the HVAC System Operation 417
References
1. Ciholas, P., Lennie, A., Sadigova, P., Such, J.M.: The security of smart buildings:
a systematic literature review. https://arxiv.org/abs/1901.05837v3. Accessed 28
Jan 2019
2. Mirsky, Y., Guri, M., Elovici, Y.: HVACKer: bridging the air-gap by attacking the
air conditioning system https://arxiv.org/abs/1703.10454. Accessed 9 Aug 2019
3. Mundt, P., Wickboldt, P.: Security in building automation systems-a first analysis.
In: International Conference On Cyber Security And Protection of Digital Services
(Cyber Security), pp. 1–8. IEEE (2016)
4. Caviglione, L., Lalande, J.-F., Mazurczyk, W., Wendzel, S.: Analysis of human
awareness of security and privacy threats in smart environments. In: Tryfonas, T.,
Askoxylakis, I. (eds.) HAS 2015. LNCS, vol. 9190, pp. 165–177. Springer, Cham
(2015). https://doi.org/10.1007/978-3-319-20376-8 15
5. Truong, H., Francisco, A., Khosrowpour, A., Taylor, J.E., Mohammadi, N.: Method
for visualizing energy use in building information models. Energy Proc. 142, 2541–
2546 (2017). https://doi.org/10.1016/j.egypro.2017.12.089
6. Palm, J., Ellegård, K.: Visualizing energy consumption activities as a tool for
developing effective policy. Int. J. Consum. Stud. (2011). https://doi.org/10.1111/
j.1470-6431.2010.00974.x
7. Abdelalim, A., O’Brien, W., Shi, Z.: Development of sankey diagrams to visualize
real HVAC performance. Energy Build. 149, 282–297 (2017). https://doi.org/10.
1016/j.enbuild.2017.05.040
8. Janetzko, H., Stoffel, F., Mittelstädt, A., Keim, D.A.: Anomaly detection for visual
analytics of power consumption data. Comput. Graph. (2014). https://doi.org/10.
1016/j.cag.2013.10.006
9. Novikova, E.S., Murenin, I.N., Shorov, A.V.: Visualizing anomalous activity in
the movement of critical infrastructure employees. In: Proceedings of IEEE Con-
ference of Russian Young Researchers in Electrical and Electronic Engineering
(EIConRus), St. Petersburg, pp. 504–509 (2017)
10. Novikova, E., Kotenko, I.: Analytical visualization techniques for security infor-
mation and event management. In: Proceedings of 21st Euromicro International
Conference on Parallel, Distributed, and Network-Based Processing Belfast, Ire-
land, pp. 519–525 (2013)
11. Kolomeets, M., Chechulin, A., Kotenko, I., Chevalier, Y.: A visual analytics app-
roach for the cyber forensics based on different views of the network traffic. J.
Wirel. Mob. Netw. Ubiquit. Comput. Dependable Appl. 9(2), 57–73 (2018)
12. Song, H., Muelder, C.W., Ma, K.-L.: Crucial nodes centric visual monitoring and
analysis of computer networks. In: Proceedings of the 2012 International Con-
ference on Cyber Security (CYBERSECURITY 2012). IEEE Computer Society,
Washington (2012) https://doi.org/10.1109/CyberSecurity.2012.9
13. Zhao, Y., Zhou, F., Fan, X.: A real-time visualization framework for IDS alerts. In:
Proceedings of the 5th International Symposium on Visual Information Commu-
nication and Interaction (VINCI 2012). ACM, New York (2012) https://doi.org/
10.1145/2397696.2397698
14. Arendt, D.L., et al.: Ocelot: user-centered design of a decision support visualization
for network quarantine. In: Proceedings of 2015 IEEE Symposium on Visualization
for Cyber Security (VizSec), Chicago, IL (2015). https://doi.org/10.1109/VIZSEC.
2015.7312763
418 E. Novikova et al.
15. Chen, S., Guo, C., Yuan, X., Merkle, F., Schaefer, H., Ertl, T.: OCEANS: online col-
laborative explorative analysis on network security. In: Proceedings of the Eleventh
Workshop on Visualization for Cyber Security (VizSec 2014). ACM, New York
(2014). https://doi.org/10.1145/2671491.267149
16. Humphries, C., Prigent, N., Bidan, C.E., Majorczyk, F.: CORGI: combination,
organization and reconstruction through graphical interactions. In: Proceedings of
the Eleventh Workshop on Visualization for Cyber Security (VizSec 2014). ACM,
New York (2014). https://doi.org/10.1145/2671491.2671494
17. Cappers, B.C.M., van Wijk, J.: SNAPS: semantic network traffic analysis through
projection and selection. In: Proceedings of the IEEE Symposium on Visualiza-
tion for Cyber Security (VizSec 2015), pp. 1–8 (2015). https://doi.org/10.1109/
VIZSEC.2015.7312768
18. Landstorfer, J., Herrmann, I., Stange, J., Dork, M., Wettach, R.: Weaving a car-
pet from log entries: a network security visualization built with cocreation. In:
Proceedings of Visual Analytics Science and Technology (VAST). IEEE (2014)
19. Mansmann, F., Meier, L., Keim, D.A.: Visualization of host behavior for network
security. In: Goodall, J.R., Conti, G., Ma, K.L. (eds.) VizSEC 2007. Mathematics
and Visualization, pp. 187–202. Springer, Heidelberg (2008). https://doi.org/10.
1007/978-3-540-78243-8 13
20. Novikova, E., Kotenko, I.: Visual analytics for detecting anomalous activity in
mobile money transfer services. In: Teufel, S., Min, T.A., You, I., Weippl, E. (eds.)
CD-ARES 2014. LNCS, vol. 8708, pp. 63–78. Springer, Cham (2014). https://doi.
org/10.1007/978-3-319-10975-6 5
21. Ankerst, M., Berchtold, S., Keim, D.A.: Similarity clustering of dimensions for an
enhanced visualization of multidimensional data. In: Proceedings 1998 IEEE Sym-
posium on Information Visualization (INFOVIS 1998), pp. 52–60. IEEE Computer
Society (1998)
22. Hout, M.C., Papesh, M.H., Goldinger, S.D.: Multidimensional scaling. WIREs
Cogn. Sci. 4, 93–103. https://doi.org/10.1002/wcs.1203
23. van der Maaten, L.J.P., Hinton, G.E.: Visualizing high-dimensional data using
t-SNE. J. Mach. Learn. Res. 9(Nov), 2579–2605 (2008)
24. Kohonen, T., Honkela, T.: Kohonen network. Scholarpedia 2(1), 1568 (2007)
25. Di Caro, L., Frias-Martinez, V., Frias-Martinez, E.: Analyzing the role of dimension
arrangement for data visualization in radviz. In: Zaki, M.J., Yu, J.X., Ravindran,
B., Pudi, V. (eds.) PAKDD 2010. LNCS, vol. 6119, pp. 125–132. Springer, Heidel-
berg (2010). https://doi.org/10.1007/978-3-642-13672-6 13
26. Ester, M., Kriegel, H.-P., Sander, J., Xu, X.: A density-based algorithm for dis-
covering clusters a density-based algorithm for discovering clusters in large spatial
databases with noise. In: Simoudis, E., Han, J., Fayyad, U. (eds.) Proceedings of
the Second International Conference on Knowledge Discovery and Data Mining
(KDD 1996), pp. 226–231. AAAI Press (1996)
27. Vast Challenge Homepage. http://vacommunity.org/. Accessed 10 Aug 2019
Secure Location Verification: Why You
Want Your Verifiers to Be Mobile
1 Introduction
Many real-world distributed systems require sharing of location information
among network nodes. For example, in location-based access control or envi-
ronmental sensing applications, the location of individual nodes is often crucial
for distributed coordination, service delivery or decision making.
A common approach to sharing location information with neighboring nodes
is broadcasting them periodically over a wireless link (e.g., ADS-B, AIS, RTK,
WiFi, or Bluetooth). While this method has advantages in terms of simplicity
and scalablity, a known weakness of this scheme is that nodes may (intentionally
or not) advertise wrong location claims. In order to detect such false location
information, secure location verification schemes have been proposed in the liter-
ature with the aim to securely verify whether the advertised (“claimed”) location
c Springer Nature Switzerland AG 2020
S. Katsikas et al. (Eds.): ESORICS 2019 Workshops, LNCS 11980, pp. 419–437, 2020.
https://doi.org/10.1007/978-3-030-42048-2_27
420 M. Schäfer et al.
corresponds to the real position of the sender. Since Brands and Chaum first
addressed this problem in 1993 [2] and Sastry et al. later defined location (or
in-region) verification in 2003 [12], many solutions and methods have been pro-
posed in the literature to solve this problem. The existing solutions can broadly
be classified into methods based on distance bounding [2,10,12,15,19,20], time-
difference of arrival measurements (TDoA) [1,16,19,22], angle of arrival mea-
surements [6,8], or hybrid methods [3,4].
All of these techniques have in common that they verify location claims by
checking physical properties of the transmitted radio signals. For example, dis-
tance bounding protocols or TDoA systems exploit the fact that a radio signals
cannot propagate faster than the speed of light. A location claim violating this
condition must be false. While the majority of these schemes have been shown to
be secure within their assumptions, the requirements to the underlying systems
limit their applicability significantly. More specifically, TDoA-based methods
generally require many verifiers and tight time synchronization between verifiers.
This is usually achieved by additional infrastructure (e.g. wired networks) and
the exchange of synchronization information between nodes. This dependency,
however, significantly reduces the flexibility and increases the communication
overhead, rendering the approaches unsuitable for ad hoc or mobile scenarios,
especially when energy supply is limited. Distance bounding and angle of arrival
measurements, on the contrary, do not require time synchronization. However,
since they rely on active verification protocols and specialized hardware (see [11]
for more details), we argue that their applicability is also limited. For exam-
ple, they cannot be applied to systems that are already in place such as mobile
phones. Upgrading the billions of smartphones in use today to meet the require-
ments of distance bounding seems rather impracticable.
In our prior works [13] and [14], we have shown that by adding mobility of
nodes to the model, requirements of similar verification systems can be lowered
significantly. More specifically, we have shown that tracks and motion of moving
provers can be verified without any of the aforementioned limitations while at
the same time strong security can be provided. However, the downside of these
approaches is that they are only applicable to scenarios with moving provers.
They are not applicable to the classic location verification problem which con-
siders stationary provers at single locations or within certain areas.
In this work, we bridge this gap by bringing the benefits of mobility to the
problem of verifying single locations. We present a novel method based on mobile
verifiers which achieves strong security without limiting the attacker’s knowl-
edge (i.e., no “security by obscurity”) nor does it rely on time synchronization
or active verification protocols. We introduce the concept of “controlled mobil-
ity” and show that by being able to adjust the verifiers’ movements to the
claimed locations, provable security can be achieved with just two verifiers and
two location claim transmissions. Compared to existing approaches, this is both
lightweight and fast. In addition to that, we also analyze more general movement
strategies for scenarios that require batch verification, i.e., the simultaneous ver-
ification of multiple location claims. We conducted extensive simulations to find
Secure Location Verification: Why You Want Your Verifiers to Be Mobile 421
In line with the definition by Sastry et al. [12], we define the problem of secure
location verification as follows. A set of verifiers V = {V1 , V2 , . . . , Vn } wishes to
check whether a prover P is at a location l of interest.
We assume that verifiers are moving while P advertises its location m > 1 times.
This implies that each transmission is received by the verifiers from different
locations. To minimize the verification overhead, we also assume that verifiers
are passive receivers and there is no communication between prover and veri-
fiers other than the prover’s location advertisements. We further assume that
the inter-transmission time Δi,i+1 of two subsequent location advertisements is
known to each verifier. This can be achieved by either using a predefined constant
interval Δ, i.e., Δ(i,i+1) = Δ for all i, or by having the prover include transmis-
sion timestamps t(i)
P in the location claim broadcasts, i.e., Δ
i,i+1 (i+1)
= tP
(i)
− tP . It
is worth noting that the first option reduces the communication overhead since
less information needs to be transmitted by the prover while the second option
provides much more flexibility, e.g., to support random medium access protocols
such as ALOHA. Finally, we assume that each verifier knows its location at all
times and has a stable but unsynchronized local clock.
that the adversary is located at a location different from the claimed location,
i.e., A = l.1
As we aim at strong security rather than security by obscurity, we do not
limit the adversary’s knowledge. In effect, our adversary has perfect knowledge of
the verifiers’ locations at any point in time and it can even predict the verifiers’
future locations. This assumption is an important difference from the ranging-
based scheme proposed by Čapkun et al. in [21], where security is based on the
adversary’s lack of knowledge of the “hidden” verifiers’ location.
Finally, we assume that the verifiers are not compromised and they have
secure means to determine their locations. Consequently, we can consider loca-
tions Va(i) and timestamps t(i)
a of the verifiers trustworthy for all a ∈ {1, . . . , n}
and i ∈ {1, . . . , m}.
A specific real-world example for a system that could (and should) be extended
with our verification scheme are navigational aid systems used in aviation such
as non-directional radio beacons (NDB) or VHF omnidirectional radio range
(VOR) [7]. In both systems, ground stations at fixed locations transmit signals
that are used by aircraft for navigation. Each transmitter is assigned a unique
identifier that can be used by pilots and onboard systems to look up the station’s
location. Once their location is known, aircraft use them to stay on track by flying
towards or in a certain angle to the ground station.
As is the case with most systems used in aviation, security has not been part
of the design of NDB and VOR. As a consequence, they are highly vulnerable to
spoofing attacks which can be used to mislead pilots or automatic flight control
systems [17]. Although many aircraft have more accurate means of navigation
(GPS), many pilots around the world still rely on these systems. Our verifica-
tion scheme could be used to mitigate this threat in a scalable and light-weight
manner. The ground transmitters can be considered as stationary provers while
aircraft equipped with additional means of positioning (GPS) can act as moving
verifiers. Onboard verification systems can then detect fake signals and inform
ground personnel and other pilots.
Another use case is access control for services which should only be avail-
able to users within a certain physically restricted area. For instance, addtional
premium information about a sports match could be offered in a mobile app
to fans within a stadium. To prevent that people on the outside have access to
that service, drones or moving cameras can be used to verify that people are
in the stadium. Moreover, existing wireless technologies such as Bluetooth or
WiFi could be used by the app for verification without the need for additional
hardware.
1
As for verifiers and prover, we use A interchangeably for the adversary’s identity
and location whenever the meaning is clear from the context.
Secure Location Verification: Why You Want Your Verifiers to Be Mobile 423
holds. They estimate the propagation delay δa(i) using the known positions, i.e.,
(i) (i)
δa = dist(Va , P )/c with dist(·, ·) denoting the Euclidean distance between two
locations and c the signal propagation speed (usually speed of light). If the
equation is satisfied, the verifier remains silent. Otherwise it raises an alarm.
The verification procedure terminates successfully, i.e., P ’s location is verified,
after m transmissions without any alarm.
We assume that the verifiers are changing their locations between the location
claim transmissions, i.e., Va(i) = Va(i−1) . Verifiers can choose their next location
Va(i) within the physical limitations using different strategies. We call this con-
scious choice of the next position “controlled mobility” (as opposed to “oppor-
tunistic” or “random” mobility) and distinguish between coordinated and unco-
ordinated controlled mobility. In coordinated controlled mobility, verifiers choose
their direction of movement collaboratively to maximize security. To avoid com-
munication overhead, we assume that verifiers coordinate their movements solely
based on the claimed location and some fixed identifier. In the following security
analysis, we derive such a coordinated movement pattern and prove its security.
The disadvantage of this approach, however, is that the verification of multiple
location claims simultaneously (“batch verification”) is not possible. While this
is not a requirement per se, there are scenarios with many participants (e.g.,
verifying people’s locations in a stadium) that require a more scalable approach.
We therefore extend our analysis with simulations evaluating more general unco-
ordinated (yet controlled) movement patterns that allow for batch verification.
Based on our results, we can provide heuristics for verifier movements that max-
imize security while preserving a high efficiency in terms of verification time and
minimum required number of verifiers.
4 Security Analysis
The two main design goals of our protocol are security and efficiency. While
security as a goal is inherent to the problem, efficiency in terms of resources and
424 M. Schäfer et al.
verification time is crucial for the protocol’s applicability in mobile scenarios. For
instance, using antenna arrays for beamforming or high performance computers
for complex algorithms on mobile nodes such as drones is impracticable since
both weight and energy consumption must be low to maintain adequate operat-
ing times. We therefore start our security analysis by setting up the theoretical
foundations and then successively increase the transmission time (in terms of
number of transmissions) and number of verifiers until security is established. In
this way, we obtain the fastest and most resource-efficient configuration that can
provide strong security. For the sake of presentation, we conduct our analysis in
two-dimensional space. Extending the results to three dimensions is straightfor-
ward.
holds, where δA(i),a = dist(A, Va )/c is the propagation delay of the i-th transmission
from the adversary to verifier Va . Considering only a single verifier, this can easily
be achieved by simply choosing
(i−1,i) (i) (i−1)
ΔA = Δ + (δa(i) − δa(i−1) ) − (δA,a − δA,a )
(i) (i−1)
= Δ(i−1,i)
a − (δA,a − δA,a ) . (2)
In other words, the adversary can simply compensate for its unexpected prop-
agation delays to Va by choosing an inter-transmission interval equal to the
difference of the expected from the actual inter-arrival time. We conclude that a
single verifier cannot provide any security since an adversary can spoof arbitrary
locations.
We point out that this result is equal to the case of a single verifier in [13].
The only difference is that in [13], a moving sender is considered whereas here
we assume that the receiver moves. However, by adding another verifier in the
next step of our analysis, we diverge from the analysis in [13] since we then
face multiple moving nodes in our system whereas [13] always considers just a
single moving node. Facing multiple mobile nodes increases the complexity of
the analysis significantly.
Fig. 1. Example scenario with two verifiers and the resulting restrictions (implicit
curve) for the adversary’s location A. (Color figure online)
Theorem 1. If one verifier moves exactly towards P while another one moves
exactly away from P and not in line with the first one, then MoVers is secure
for m = 2.
(2) (1)
(Va/b − Va/b ) be the vectors describing the position changes of the two verifiers
between the two transmissions. Then there is an sa ∈ R with sa > 0 such that
P = Va(1) + v a · sa .
P = Vb(1) + v b · sb .
We further assume that the two verifiers are not in line with each other, i.e.,
there is no s ∈ R such that
Vb(1) = Va(1) + v a · s.
are satisfied. Let us now assume A would not be located in line with Va and P .
Then
(2) (1) v a
δA,a − δA,a > −
c
holds, since Va does not move exactly towards A. This means for the adversary
that it has to choose Δ(1,2)
A < Δ to compensate for the difference. Yet, then it
cannot satisfy Eq. 6 since it would be required that
(2) (1) v b
δA,b − δA,b > .
c
However, given the distance moved by Vb between the two transmissions and the
associated maximum possible distance change of v b , this is impossible. Hence,
A must be located in line with Va .
We can show analogously that A must also be in line with P and Vb in order
to satisfy Eq. 5. As a consequence, A must be located on two lines which both
cross P . Since Va and Vb are not in line with each other, these two lines are
different. Since furthermore two different lines can only have one intersection,
we can conclude that P is the only location from which a sender can satisfy both
equations at the same time. Thus, Theorem 1 holds and MoVers is secure.
We conclude that by adapting the verifiers’ movements to the claimed loca-
tion (“coordinated controlled mobility”), MoVers can provide provable security
with an efficient configuration of two verifiers and two transmissions.
Summary: The key results from this theoretical analysis are that (i) a single
verifier cannot provide any security, (ii) two verifiers can provide provable secu-
rity with coordinated controlled mobility (Theorem 1), and (iii) the security
increases with each additional transmission or verifier as more restrictions are
added for the adversary’s location.
6 Uncoordinated Mobility
the adversary by adding more implicit curves to the constraints for A. More
specifically, since the verifiers move between each transmission, the focal points
for the implicit curve defined by Eq. (4) change for every i ∈ {2, . . . , m} and each
pair Va , Vb ∈ V . As a result, A needs to be located at an intersection of (m−1)· n2
different implicit curves in order to remain undetected when claiming P = A.
Moreover, this set of intersections can be assumed to be finite since the curves
are not periodic. The number of such intersections can be considered a direct
measure of the attacker’s degree of freedom and thus the security of our scheme.
Our scheme is in particular secure if there is only one intersection of all curves
(which is P by construction) since false claims will then violate Eq. (1) for at
least one verifier.
Most related problems are of a simple hyperbolic nature (e.g. [13,14,18]) and
can often be analyzed algebraically. Unfortunately, having more than one mobile
node makes the exact analysis hard because each moving element contributes to
the equations. For example, in contrast to the analysis of intersections of a set of
hyperbolas, which is common, e.g., for TDoA or ranging-based approaches, we
face curves defined by intersections of intersections of hyperbolas with multiple
parameters. These curves are of a higher order than hyperbolas which makes
an exact analysis of the intersections extremely difficult. Although there exist
methods to decrease the computational complexity when computing the inter-
section of hyperbolas (e.g., homogeneous coordinates [9]), we could not find any
analytical method to analyze it in a general way, since the parameters that may
determine the hyperbolas are unknown. We therefore continue our analysis by
extending our theoretical findings with simulations analyzing the behavior of the
intersections with respect to the verifiers’ movements independent from P .
In the following simulations, we differentiate between opportunistic and
(uncoordinated) controlled mobility. In opportunistic (or random) mobility,
nodes are not moving according to any predefined pattern. This reflects sce-
narios where uncontrolled nodes act as verifiers (e.g., cellphones, agricultural
machines, or airplanes). In controlled mobility, verifiers follow certain patterns
aiming at improving the security of the verification scheme.
βa
Va(2)
αa
Va(1) Va(3)
40
50
20
25
0 0
25 50 75 100 0 100 200 300
Step width (m) β (°)
(a) Effect of step width/speed on the (b) Effect of relative movement (αa − αb )
number of intersections. The gray solid and different βa = βb = β on the number
line is the percentage of the cases in of intersections. This graph only shows
which the location was securely verified. the percentage of the 10.000 random sce-
The other lines represent the percentages narios that were secure, i.e., the set of in-
where an adversary could have been lo- tersections I only contained P .
cated at an increasing number of locations
other than P .
Effect of Speed v . The speed of the verifiers defines the distance covered by
a verifier between the periodic re-transmissions of a location claim. To evaluate
whether the resulting step width has an impact on the number of intersections,
we randomly generated 10.000 scenarios for the “cheapest” configuration n = 2
(1)
and m = 3. Each verifier starts at a random location Va/b and moves into a
random direction at different speeds 10 ≤ v ≤ 100 m/s. For each scenario, we
recorded the number of intersections |I |. We did not consider larger speeds since
they would be unrealistic given an area of interest of 209 × 255 m. The results are
shown in Fig. 4a. While the percentage of scenarios in which the claimed location
could be securely verified (gray solid line) slightly increased with increasing
step width, the percentage of |I | = 2 was constantly over 50%. For smaller
v , there were even about 20% of scenarios in which an adversary could have
chosen between two (dashed blue with squares) or three (green dashed with
pluses) locations different to P which also satisfied Eq. (1) for both verifiers. We
conclude that the step width (or speed) has only a minor effect on the number of
intersections. On the one hand, this means that the step width does not provide
much room for improving the security. However, on the other hand, this also
means that slow verifiers do not suffer big disadvantages.
432 M. Schäfer et al.
For the following simulations, we set the speed of the verifiers to that of
commercial off-the-shelf drones such as DJI’s Phantom 4, i.e., v = 20 m/s. The
turns of the verifiers between the two steps are controlled by βa and βb . To keep
our scheme light-weight, we assume that the verifiers do not communicate for
coordination and assume constant pre-defined βa = βb = β . However, since the
curves determining |I | do not only depend on β but also on αa and αb , we further
analyze how the difference between the two angles, i.e., the relative direction of
the verifiers to each other affects the intersections. As before, we conducted
10.000 random simulations for different combinations of β and αa − αb .
The results are shown in Fig. 4b. The graph shows that both the effect of β
and that of αa − αb on |I | are almost independent from each other. Regardless of
the difference in direction, any β close to 0◦ (respectively 360◦ ) should be avoided.
For large direction differences αa − αb , the best choice for β is around 110◦ or
250◦ . Note that both angles represent the same absolute change in direction since
360◦ − 250◦ = 110◦ .
An interesting special case is β = 180◦ , i.e., the third location of each verifier
is the same as the first one (V (1) = V (3) ). As a result, the implicit curve generated
by the first two transmissions coincides with that one of the second and third
transmission. In other words, the third transmission does not impose a new
constraint on the adversary and it is only limited to locations on the implicit
curve (compare Sect. 4.2).
More specifically, let us assume two verifiers Va and Vb receiving three trans-
missions of a location claim for P . According to Sect. 4.2 a potential adversary’s
location A must satisfy the following system of instances of Eq. (4):
(2) (1) (2) (1) (2)
(δA,b − δA,b ) = (δA,a − δA,a ) + kP
(3) (2) (3) (2) (3)
(δA,b − δA,b ) = (δA,a − δA,a ) + kP
and thus
(3) (2) (3) (2) (3)
(δA,b − δA,b ) = (δA,a − δA,a ) + kP
(1) (2) (1) (2) (2)
⇔ (δA,b − δA,b ) = (δA,a − δA,a ) − kP
(2) (1) (2) (1) (2)
⇔ (δA,b − δA,b ) = (δA,a − δA,a ) + kP
Consequently, the third transmission does not impose a new constraint on the
adversary’s location A if β = 180◦ .
Regarding the direction difference αa −αb , we can summarize that the closer it
is to 180◦ , the higher the percentage of locations which could be securely verified
434 M. Schäfer et al.
92.83
100
Controlled Mobility
Uncontrolled Mobility
80
Share [%]
54.35
60
31.65
40
9.47
20
5.01
4.36
2.16
0.13
0.04
0
0
0
0 1 2 3 4 5
Number of Intersections
after the third transmission. In fact, we also did the simulations for αa −αb > 180◦
but the results were identical to those for 360◦ − (αa − αb ).
We conclude from our simulations that with β = 110◦ or β = 250◦ and a
direction difference of |αa − αb | = 180◦ , more than 92.5% of all scenarios could be
securely verified with two verifiers and three transmissions of the location claim.
This is a huge improvement compared to random movement as shown in Fig. 5.
7 Related Work
As already mentioned in the introduction, many solutions and methods have
been proposed in the literature to solve the problem of secure location verifi-
cation. Existing solutions can broadly be classified into methods based on dis-
tance bounding [2,12,15,19,20], time-difference of arrival measurements (TDoA)
[16,19,22], angle of arrival measurements [6,8], or hybrid methods [3,4]. As men-
tioned in the introduction, each of these schemes comes with limiting require-
ments such as tight time synchronization, specialized hardware, directional
antennas, or limited attacker knowledge. We therefore argue that they are not
applicable to scenarios where passive and lightweight solutions are required.
Only a few works have tackled the case of mobile verifiers for secure location
verification. However, these protocols differ significantly from ours. Čapkun et
al. [21] proposed a location verification scheme, in which a mobile verifier initi-
ates a challenge-response protocol from a known position and then moves to an
unknown position to receive the response. The response is sent simultaneously
via ultrasound and RF so that the verifier can estimate its distance to the prover
based on the time-difference of arrival of the two signals due to their differing
propagation speed (ranging). The security of the approach derives from the fact
that although dishonest provers could modify the transmission times of the two
response signals, they would need to correctly guess the verifier’s new location
in order to mimic the expected time-difference of arrival. This scheme, however,
is cooperative and requires nodes to be equipped with two transceivers (ultra-
sound and RF). Moreover, we challenge the assumption ‘untraceability’ of the
Secure Location Verification: Why You Want Your Verifiers to Be Mobile 435
moving verifier. Even though the verifier does not actively transmit revealing
signals from its new location, a more sophisticated adversary could track the
verifier via reflections of signals of opportunity and passive radar techniques [5].
In [10], Perazzo et al. propose a location verification system in which a veri-
fier drone performs distance bounding with a prover consecutively from several
different locations. The locations are carefully chosen such that they form a tri-
angle containing the prover’s location. In this way an adversary claiming a false
location inside the triangle needs to mimic a shorter distance to at least one of
the locations chosen by the drone. As shown by Čapkun et al. in [20], this is
infeasible and hence, the scheme is secure. However, their approach inherits all
the aforementioned system requirements from distance bounding and is there-
fore not well-suited for location verification in existing systems or systems with
limited resources.
Baker and Martinovic proposed a TDoA-based scheme in [1]. Their scheme
relies on two verifiers, one fixed and the other one moving, to measure the TDoA
of multiple location broadcasts by the prover. Since one verifier is changing loca-
tion between each of the prover’s transmissions, different TDoAs are expected
each time. Analogously to traditional multilateration, each TDoA measurement
reduces the set of possible locations of the transmitter to one arm of a hyperbola.
By repeating the measurements at least three times (in 2D) and comparing the
expected to the measured TDoAs, the adversary can be localized by intersecting
the resulting hyperbolas. As mentioned, however, TDoA measurements require
tight time synchronization and extra communication to collect all measurements
at a central processing unit which our protocol does not require.
Finally, we want to highlight the difference of this work to our related works
on track verification [13] and motion verification [14]. First, the underlying prob-
lem considered in this paper (verification of locations) is different to that consid-
ered in [13] (verification of sequences of locations) or [14] (verification of motion).
The seemingly strong similarity is largely a result of the common theoretical
foundations on which these works are based on. This work, however, diverges
significantly in terms of problem statement, use cases, and security properties
from our previous works. More specifically, the theoretical analysis conducted
in this paper considers multiple moving nodes at the same time, whereas the
analyses of [13] and [14] are only applicable to systems with one moving node.
As a result, the analytical nature of the security guarantees of our scheme is not
hyperbolic anymore, making them much harder to analyze.
8 Conclusion
References
1. Baker, R., Martinovic, I.: Secure location verification with a mobile receiver. In:
Proceedings of the 2nd ACM Workshop on Cyber-Physical Systems Security and
Privacy (CPS-SPC), October 2016
2. Brands, S., Chaum, D.: Distance-bounding protocols. In: Helleseth, T. (ed.) EURO-
CRYPT 1993. LNCS, vol. 765, pp. 344–359. Springer, Heidelberg (1994). https://
doi.org/10.1007/3-540-48285-7 30
3. Chiang, J.T., Haas, J.J., Choi, J., Hu, Y.C.: Secure location verification using
simultaneous multilateration. IEEE Trans. Wirel. Commun. 11(2), 584–591 (2012)
4. Chiang, J.T., Haas, J.J., Hu, Y.C.: Secure and precise location verification using
distance bounding and simultaneous multilateration. In: Proceedings of the 2nd
ACM Conference on Wireless Network Security (WiSec), March 2009
5. Howland, P.: Editorial: passive radar systems. IEE Proc. - Radar Sonar Navigat.
152(3), 105–106 (2005). https://doi.org/10.1049/ip-rsn:20059064
6. Hu, L., Evans, D.: Using directional antennas to prevent wormhole attacks. In:
Network and Distributed System Security Symposium (NDSS), February 2004
7. International Civil Aviation Organization (ICAO): International Standards and
Recommended Practices, Annex 10: Aeronautical Telecommunications, 6 edn. Vol-
ume I: Radio Navigation Aids (2006)
8. Lazos, L., Poovendran, R., Čapkun, S.: ROPE: robust position estimation in wire-
less sensor networks. In: Proceedings of the 4th International Symposium on Infor-
mation Processing in Sensor Networks (IPSN). IEEE Press, April 2005
9. Li, H., Hestenes, D., Rockwood, A.: Generalized homogeneous coordinates for com-
putational geometry. In: Sommer, G. (ed.) Geometric Computing with Clifford
Algebras, pp. 27–59. Springer, Heidelberg (2001). https://doi.org/10.1007/978-3-
662-04621-0 2
10. Perazzo, P., Ariyapala, K., Conti, M., Dini, G.: The verifier bee: a path planner for
drone-based secure location verification. In: Proceedings of the 16th IEEE Inter-
national Symposium on A World of Wireless, Mobile and Multimedia Networks
(WoWMoM), June 2015
11. Rasmussen, K.B., Čapkun, S.: Realization of RF distance bounding. In: Proceed-
ings of the 19th USENIX Conference on Security (2010)
12. Sastry, N., Shankar, U., Wagner, D.: Secure verification of location claims. In:
Proceedings of the 2nd ACM Workshop on Wireless Security (WiSe), September
2003
13. Schäfer, M., Lenders, V., Schmitt, J.B.: Secure track verification. In: IEEE Sym-
posium on Security and Privacy, May 2015
14. Schäfer, M., Leu, P., Lenders, V., Schmitt, J.: Secure motion verification using the
doppler effect. In: Proceedings of the 9th ACM Conference on Security and Privacy
in Wireless and Mobile Networks (WiSec), July 2016
15. Singelee, D., Preneel, B.: Location verification using secure distance bounding pro-
tocols. In: IEEE International Conference on Mobile Adhoc and Sensor Systems
Conference (MASS), November 2005
16. Strohmeier, M., Lenders, V., Martinovic, I.: Lightweight location verification in
air traffic surveillance networks. In: Proceedings of the 1st ACM Workshop on
Cyber-Physical System Security (CPSS) (2015)
17. Strohmeier, M., Schäfer, M., Pinheiro, R., Lenders, V., Martinovic, I.: On percep-
tion and reality in wireless air traffic communication security. IEEE Trans. Intell.
Transp. Syst. 18(6), 1338–1357 (2017)
Secure Location Verification: Why You Want Your Verifiers to Be Mobile 437
18. Tippenhauer, N.O., Pöpper, C., Rasmussen, K.B., Capkun, S.: On the requirements
for successful GPS spoofing attacks. In: Proceedings of the 18th ACM Conference
on Computer and Communications Security (CCS), October 2011
19. Čapkun, S., Hubaux, J.P.: Securing position and distance verification in wire-
less networks. Technical report, École polytechnique fédérale de Lausanne (EPFL)
(2004)
20. Čapkun, S., Hubaux, J.P.: Secure positioning of wireless devices with application
to sensor networks. In: Proceedings IEEE 24th Annual Joint Conference of the
IEEE Computer and Communications Societies (INFOCOM), vol. 3, March 2005
21. Čapkun, S., Rasmussen, K.B., Čagalj, M., Srivastava, M.: Secure location verifi-
cation with hidden and mobile base stations. IEEE Trans. Mob. Comput. 7(4),
470–483 (2008)
22. Čapkun, S., Čagalj, M., Srivastava, M.: Secure localization with hidden and mobile
base stations. In: Proceedings of the 25th IEEE International Conference on Com-
puter Communications (INFOCOM), April 2006
Author Index