A Survey On Security and Privacy of Federated Learning
A Survey On Security and Privacy of Federated Learning
PII: S0167-739X(20)32984-8
DOI: https://doi.org/10.1016/j.future.2020.10.007
Reference: FUTURE 5875
Please cite this article as: V. Mothukuri, R.M. Parizi, S. Pouriyeh et al., A survey on security and
privacy of federated learning, Future Generation Computer Systems (2020), doi:
https://doi.org/10.1016/j.future.2020.10.007.
This is a PDF file of an article that has undergone enhancements after acceptance, such as the
addition of a cover page and metadata, and formatting for readability, but it is not yet the definitive
version of record. This version will undergo additional copyediting, typesetting and review before it
is published in its final form, but we are providing this version to give early visibility of the article.
Please note that, during the production process, errors may be discovered which could affect the
content, and all legal disclaimers that apply to the journal pertain.
of
Viraaji Mothukuria , Reza M. Parizia , Seyedamin Pouriyehb , Yan Huanga , Ali
Dehghantanhac , Gautam Srivastavad,e,∗
a The Department of Software Engineering and Game Development, Kennesaw State
pro
University, GA 30060, USA
b The Department of Information Technology, Kennesaw State University, GA 30060, USA
c Cyber Science Lab, School of Computer Science, University of Guelph, Ontario, Canada
d Department of Math and Computer Science, Brandon University, MB, Canada
e Research Center for Interneural Computing, China Medical University, Taichung 40402,
Abstract re-
Federated learning (FL) is a new breed of Artificial Intelligence (AI) that
builds upon decentralized data and training that brings learning to the edge or
directly on-device. FL is a new research area often referred to as a new dawn
in AI, is in its infancy, and has not yet gained much trust in the community,
mainly because of its (unknown) security and privacy implications. To advance
the state of the research in this area and to realize extensive utilization of the
lP
FL approach and its mass adoption, its security and privacy concerns must be
first identified, evaluated, and documented. FL is preferred in use-cases where
security and privacy are the key concerns and having a clear view and under-
standing of risk factors enable an implementer/adopter of FL to successfully
build a secure environment and gives researchers a clear vision on possible re-
search areas. This paper aims to provide a comprehensive study concerning
FL’s security and privacy aspects that can help bridge the gap between the
rna
∗ Corresponding author
Email addresses: [email protected] (Viraaji Mothukuri ),
[email protected] (Reza M. Parizi), [email protected] (Seyedamin Pouriyeh),
[email protected] (Yan Huang), [email protected] (Ali Dehghantanha),
[email protected] (Gautam Srivastava)
of
Federated Learning, Federated Machine Learning, Security, Privacy.
1. Introduction
pro
In traditional machine learning (ML), the efficiency and accuracy of models
rely on computational power and training data of a centralized server. Shortly
speaking, in traditional ML, user data is stored on the central server and utilized
for training and testing processes in order to develop comprehensive ML models
ultimately. Centralized-based ML approaches, in general, are associated with
different challenges including computational power and time, and most impor-
tantly security and privacy with respect to users’ data that has been neglected
for a long time. Federated Learning proposed by [1] has recently emerged as a
technological solution to address such issues.
re-
Federated Learning (FL) [2] offers a way to preserve user privacy by decen-
tralizing data from the central server to end-devices and enables AI benefits to
domains with sensitive data and heterogeneity. This paradigm came to light
mainly for two reasons: (1) The unavailability of sufficient data to reside cen-
trally on the server-side (as opposed to traditional machine learning) due to
direct access restrictions on such data; and (2) Data privacy protections using
local data from edge devices, i.e., clients, instead of sending sensitive data to
lP
the server where network asynchronous communication comes into play. Pre-
serving data privacy provides feasibility to leverage AI benefits enabled through
machine learning models efficiently across multiple domains. Additionally, com-
putational power is shared among the interested parties instead of relying on
a centralized server by iterative local model training processes on end-devices.
With its decentralized data concept, FL is one of the growing fields in the area
of ML in recent years, as it comes with security and privacy features promising
rna
to abide by emerging user data protection laws [3, 4]. In addition to privacy,
FL enables ML benefits to smaller domains where sufficient training data is not
available to build a standalone ML model.
As phrased by authors in [1] “FL brings the code to the data, instead of the
data to the code and addresses the fundamental problems of privacy, ownership,
and locality of data”. Even before we can fully appreciate the fact that in FL,
data stays intact on the users’ device and the traditional upload of data over
the network can be gracefully skipped, we are sharing the model parameters
and global ML model with each and every client, which opens numerous ways
Jou
2
Journal Pre-proof
to security and privacy aspects. Therefore, we can question what type of security
of
and privacy issues are we facing now and can we imagine occurring in the future
as a consequence of the adoption of this technology? This paper aims to provide
the answer to these types of questions and shed light on possible unwanted future
security and privacy states that we should be mindful of and prepare for.
Privacy-preserving promises of FL attracts different domains that may con-
pro
tain sensitive data. To an extent, FL does solve privacy concerns of sensitive
data in ML environments however at the same time model parameter sharing
and an increased number of training iterations and communications exposes the
federated environment to a new set of risks and opens new avenues for hacking
[5] as well as curious attackers to trace vulnerabilities to manipulate ML model
output or get access to sensitive user data. To ensure that we out-turn the
benefits of FL over risks and utilize the features of FL properly, we have an
immediate need to be on top of this area of research to investigate all possi-
ble security and privacy attacks on FL. Without precise information and clear
1.1. Contributions
While research already exists on this topic, sufficient progress has not been
made concerning understanding FL for its security and privacy risks. This
work hopes to contribute a comprehensive overview of FL security in terms of
rna
3
Journal Pre-proof
of
The rest of this paper is structured as follows. Section 2 on page 4 gives
background information about FL and its underlying working process. Section
3 on page 6 classifies approaches and techniques related to the FL. Section 4 on
page 17 provides the research questions on the security of FL and presents the
research results. Similarly, Section 5 on page 28 gives the research questions
pro
on the privacy of FL and presents the research results. Section 6 on page 34
gives the related work. Section 7 on page 35 summarizes the future directions
for security and privacy in the FL domain. Finally, Section 8 on page 37 gives
the concluding remarks.
In recent years, three major contributing factors have helped in the success of
re-
ML. The first and major contributing factor is the availability of a huge amount
of data collected over the years (thanks to big data!). The second is computa-
tional power, technology has evolved so much that we have moved away from
traditional computing devices to highly scalable and integrated microcircuit de-
vices which help to train models faster and deploy directly on devices with less
computational costs. For instance, we now have AI-ready smartphones with a
pre-installed AI chip to make smartphones much smarter and intelligent and as-
sist humans efficiently in day-to-day tasks. The third contributing factor is Deep
lP
Learning Models, which add much-needed intelligence to ML models. Models
trained on self-taught deep learning algorithms are showing commendable suc-
cess rates. For instance, Alpha-Go’s [6] board game victory over the world
champion has amazed people with its ability to master the game and win over
human Intelligence. In spite of the immense success of ML [7], many domains
can only wish to leverage its benefits but can not do so because of two major
rna
obstacles:
• Concerns on user data privacy and confidentiality and the laws that oversee
them.
• Inability to build an ML model due to inadequate data or training cost
on ML implementation of the computational cost involved for training an
ML model.
which are able to bring ML knowledge and computational power at a cost how-
ever still we see that privacy and confidentiality concern remain unaddressed
[9, 10, 11, 12, 13, 14]. In an effort to address such obstacles, the community has
seen a promising ML-based framework, known as Federated Learning. FL ad-
dresses these concerns by providing a highly trained ML model without the risk
of exposing training data. FL also tackles the issue of having inadequate data by
providing a trust factor among heterogeneous domains. Such privacy-preserving
4
Journal Pre-proof
of
pro
re-
lP
Figure 1: FL Process Flow
3. Aggregation of local models: Local models are trained at the client level
and updates are sent to the central server in order to aggregate and train
the global ML model. The global model is updated and the improved
model is shared among the individual clients for the next iteration.
FL is in a continuous iterative learning process that repeats the training
steps of 2 & 3 above to keep the global ML model updated across all the clients.
Figure 1 visualizes the FL architecture and training approach as discussed in
the steps above.
5
Journal Pre-proof
In reality, FL has already shown its footprint in mobile applications for next
of
word prediction on keyboards [15, 16, 17, 18] like Gboard by Google on Android
mobile phones, and wake word detection [19], which enables voice assisting
apps to detect wake word without risking exposure of sensitive user data. In
medical domains, FL can be utilized to keep patient data private and enhance
ML capabilities in assisting medical practitioners similar to the work in [20]
pro
which demonstrates the benefits of FL in the medical domain.
Apart from live production applications, there have been several research
proposals with useful application use-cases experimenting with the use of FL
for constructing privacy-persevering (or even secure) machine learning solutions
in various domains. For instance, the research works in [21, 22] summarize
possible applications for wireless communications with FL by avoiding commu-
nication overheads. For application use-cases in the security domain, FL has
been advocated for malware classification [23], human activity recognition [24],
anomaly detection [25], and intrusion detection [26], to name a few. For ap-
re-
plication use-cases in the intelligent transportation industry, there have several
proposals that use an FL-based approach. Data sharing between autonomous
cars and driving [27, 28], For preventing data leakage in vehicular cyber-physical
systems [29], traffic flow prediction [30], and the detection of attacks in aerial
vehicles [31] are examples of such works. For application use-cases in the com-
puter vision domain, Fedvision for object detection using secure FL has been
proposed by Webank in [32]. As for application use-cases in the medical domain,
lP
the attack detection in Medical cyber-physical systems that maintain sensitive
information on patient’s health records [33], and managing digital health record
with FL [34] are more examples of FL applications. It is worth mentioning
here that the primary focus of this paper is to investigate the potential security
and privacy-related issues within FL. Thus, examining the applications of FL
in different domains is slightly out of the scope with the purpose of this paper.
rna
3. FL categorization of techniques/approaches
FL is under active development and uses a variety of techniques and ap-
proaches to realize its underlying technology in practice. For an emerging tech-
nology, the categorization of its techniques and approaches is a crucial first step
that helps to understand and explore beyond the outlined big picture. This
section gathers and gives an overview of the inner workings of such techniques
from various points of view which will be used for a deeper understating of se-
curity and privacy aspects in later sections. Figure 2 shows our classification.
Jou
6
Journal Pre-proof
of
pro
re-
lP
rna
The focus of this category is the underlying architecture of FL, the way
components are built to achieve a FL environment. Based on network topology,
FL can be categorized as centralized or fully decentralized [35].
7
Journal Pre-proof
a global model, and share it back with all clients. This is mainly preferred to
of
establish a third party system for building a trust factor among clients. Single
server and multiple client topology/hub-and-spoke topology [36] are followed
which ensures a centralized authority to monitor and manage the continuous
learning process. Unlike the traditional centralized server which hosts data and
trains a given model on shared data, the centralized server in the FL envi-
pro
ronment works only on a shared model through synchronous or asynchronous
updates from clients. Gboard for Android keyboard developed using Tensorflow
federated from Google can be exemplified as a centralized approach of the FL.
Figure 1 represents the typical steps of such an approach. The current common
implementations of FL in practice use a centralized approach.
The clustering technique is proposed to improvise FL which aims to address
heterogeneity in FL clients’ data in the centralized network topology. Clustering
is one of the techniques which can help in detecting malicious updates. Research
work from google in [37] proposes three algorithms to achieve a personalized
re-
version of the local model with minimal communication rounds. Authors suggest
leveraging the algorithms as a combination or individually based on the use
cases of user clustering, data interpolation, and model interpolation. In the
user clustering algorithm, Clusters are created with a group of clients with
similar data distribution, and with each client, an intermediate model is created
to which helps the global model to converge faster. Hypothesis based clustering
technique is followed for identifying clusters. A Federated Multi-task learning-
lP
based approach is proposed in [38] where FL clusters of clients are identified
based on cosine similarity of local models. The initial set of clients are taken
and split into clusters recursively based on the stopping criteria set based on
the calculated cosine similarity.
Research work in [39] proposes Federated Stochastic Expectation Maximiza-
tion (FedSEM) to train multiple global ML models and reach a solution. Loss
function called Distance-based Federated loss (DF-Loss) is the objective of
rna
multi-center FL which is to find the optimal global model among the different
global models from a multi-cluster environment. Authors in [40] propose the
Iterative Federated Cluster Algorithm (IFCA) framework which tries to mini-
mize the loss function of each FL client and tags the client to a cluster in each
training round. Based on the source for averaging IFCA, it is proposed in two
variants called model and gradient averaging. IFCA applies random initializing
and multiple restarts to identify clients with a specific cluster and reach optimal
values. Experimental results of IFCA show that the proposed approach works
well in the linear model, convex, and non-convex neural networks.
Jou
3.1.2. Fully-decentralized FL
The fully-decentralized approach excludes dependency on the central server
for model aggregation. Centralized authority is replaced with algorithms to
establish trust and reliability. As demonstrated in [41], there is no concept of
a global model and each participant improves their model by sharing informa-
tion with neighbors. Peer-2-peer topology is followed and central authority is
used once to establish a protocol to be followed in the network during training
8
Journal Pre-proof
of
on technologies or algorithms are proposed. Authors in [42] propose an Adaptive
Averaging Algorithm which is based on the Byzantine concept, assuming more
than 23 rd of systems involved in FL are honest. With this approach, a group
of clients from different domains with a common goal collaborate to share data
and build an ML model and leverage benefits of high accuracy [43, 44] without
pro
the need to rely on a third-party centralized server. Authors in [45] propose
a framework called MATCHA to address network delays by providing critical
links for the clients to communicate with one another. Research work in [46]
demonstrates FL in peer-peer network. To help make this FL approach clear,
we have visualized its process in Figure 3.
9
Journal Pre-proof
of
areas of interest. In real-time the information, or to be more specific the statisti-
cal information derived from a set of user data, is utilized for many applications
of various domains. In this era of digitalization, every click a user makes in
cyberspace is captured for making derived statistical information, use-cases on
such derived data may belong to applications in identical domains or different
pro
domains. Similarly, user data to develop such derived statistical information
can be collected from domains belonging to dissimilar areas. Categorization of
this section can be considered as “pre-work" for setting up an FL environment.
The pre-work done on a set of interested parties with valuable user data can
be vital for the FL process overall. Distribution characteristics of data, i.e., the
differentiating & the colliding factors across heterogeneous data & clients par-
ticipating in FL can be broadly categorized as Horizontal, Vertical, and Transfer
learning (HFL1 , VFL2 , TFL3 ). These subcategories differ based on the data flow
between the parties involved in the FL environment. Figure 4, 5 ,6 illustrates
data partitioning in FL.
re-
3.2.1. Horizontal Federated Learning
Horizontal federated learning 4 is defined as the circumstances where datasets
on the devices share the same features with different instances. In this category
of FL, clients have similar features in terms of domain, usage style of derived
statistical information, or any other outcome from FL. The classic release of FL
lP
from Google [47] falls under this category. ML models which can predict the next
possible words while a user key in the text input, works more accurately with
real data feed as input from users. In this scenario, the Google keyboard app
(known as Gboard) enhances itself with continuous learning from users of An-
droid mobile devices. FL approach is implemented smartly by taking averaged
updates from the user usage stats without tagging the user identity. Another
example is from the medical domain, where a group of researchers working on
rna
an ML model which can analyze the medical images and predict the probability
of possible occurrence of cancer cells. Medical images are considered as user
sensitive data and cannot be shared as is due to constraints and laws of private
medical data. However with FL, information on such private data can be shared
securely either through a secure aggregated updates from each client.
pants using this approach prefer to have an intermediate third party organiza-
tion/resource to provide encryption logic to ensure that the only common data
stats are shared. However it is not mandatory to have a 3rd party intermediate
10
Journal Pre-proof
of
pro
re-
Figure 4: Data Partition: Horizontal Federated Learning
can serve customers better with relevant offers and online shopping domains can
revise their points allocation for customers using credit cards.
Jou
11
Journal Pre-proof
of
Federated Transfer learning 6 is an implementation of the existing classic
ML transfer learning technique which is a way to train a new requirement on
a pre-trained model that is already trained on a similar dataset for solving a
completely different problem. In the ML world, training on a pre-trained model
gives much better results in comparison to training done on a fresh model built
pro
from scratch. In [49, 50, 51, 52, 53], the authors explain about the implementa-
tion of FL in transfer learning modes. A real-time example would be similar to
vertical federated learning with a few changes. Instead of restricting conditions
to share only matching data information, participants can benefit from larger
datasets and well-trained ML model stats to serve their individual requirements.
As explained in [49], a real-time example is to apply a global model on a cloud
to a personalized user which can be further used to provide a personalized model
on wearable Internet of Things (IoT) devices of a specific user. This example
is illustrated in Figure 6. There are few research works [54, 55, 56] called Fe-
re-
dRL, which combine the reinforcement ML algorithms with FL in applications
aiming for personalized AI. FedRL utilizes FL to apply transfer learning from a
well-trained secure model with which direct transfer learning is not permitted.
lP
rna
Jou
12
Journal Pre-proof
3.3.1. Cross-silo FL
of
In this case, clients are typically of small-scale numbers ranging from 2-100
devices, usually indexed and are almost always available for training rounds.
Training data can be classified into either horizontal or vertical learning. Com-
putation and communication bottlenecks are major issues. Cross-silo FL is more
flexible compared to cross-device FL and used in scenarios within organizations
pro
or within groups of organizations to train the ML model with their confiden-
tial data. As in vertical and transfer learning implementations, an encryption
technique is preferred to restrict the inference of information from each client.
Built on FATE [57] framework, the research work in [58] demonstrates cross-silo
with homomorphic encryption and proposes gradient quantization [59] based
batch encrypt algorithm to reduce computation and communication costs in FL
environment [60].
3.3.2. Cross-device FL
re-
The FL approach with a large number of clients in a similar domain with sim-
ilar interests from the global model is known as Cross-device FL. Due to the huge
number of clients, it is difficult to track and maintain transaction history logs.
Mostly clients connect using unreliable networks where selection/participation
in training rounds happens randomly. Similar to data partitioning in hori-
zontal federated learning, resource allocation strategies [61] like client selec-
tion/importance [62] , device scheduling [63] are used to choose updates from
lP
better contributing clients. Incentive mechanisms like game theory [64] are de-
signed to motivate clients to contribute towards FL. For use-cases with a huge
number of clients such as IoT or mobile applications [16], this type is a great
fit.
13
Journal Pre-proof
of
mentation of FL based on the pre-config parameters. The current algorithms
are discussed below:
pro
As part of FedAvg algorithm logic, the central server acting as a coordi-
nator or orchestrator starts the process of FL training by sharing global
parameters and global model to a group of selected clients referred to as
mini-batch, who trains the ML model with local training data and global
model parameters and shares the trained model weights with the central-
ized server. The global model is generated by utilizing averaging logic to
compute the weighted sum of all the received local model updates at the
centralized server. Configurable criteria, number of training rounds is the
stopping condition for the coordinator to stop the training rounds and
re-
average the local model updates.
• SMC-Avg: As explained in [66], secure aggregation is based on the concept
of the Secure Multiparty Computation (SMC) algorithm, which aggregates
private values of mutually distrustful parties without revealing informa-
tion about their private values. Designed for addressing challenges of the
mobile device-based FL environment. This algorithm has a fault tolerance
limit, which means it works well even if 31 rd users fail to participate in the
lP
designed protocol.
• FedProx: A modified version of the FedAvg [1] algorithm is proposed
in [67] to tackle heterogeneity in FL. The experimental results of this
paper indicated positive results for FedProx in heterogeneous networks.
However, the experimental study [65] performed on FedAvg and FedProx
demonstrated that FedAvg achieves the highest accuracy among federated
rna
14
Journal Pre-proof
FedMA can also use communication as a variant that sends global model
of
matching results to clients at the start of each training rounds and adds
new neurons in the local model as an update to the global model instead
of matching, to achieve better overall performance.
• Scaffold: Stochastic Controlled Averaging for FL (Scaffold) proposed in
[70] addresses gradient dissimilarity/client-drift issue faced in federated av-
pro
eraging algorithm 3.4.1 for FL approaches where clients are stateful with
which algorithm can maintain/control variants at the client and server-side
to ensure that client updates are in moving towards with global conver-
gence.
on local data, the second one a local copy of the global model, and third
a personalized ML model built using the mixing parameter. Based on the
average distribution of data across the FL clients a mixing parameter is
calculated at each client node which keeps changing based on the distance
between the three (local, global, and personalized) models. Experimental
results of APFL perform better than federated averaging algorithm (i.e.,
FedAvg) in achieving a personalized model for each FL client.
15
Journal Pre-proof
share single value for sample instead of model parameters and perform
of
many local updates before sharing parameters with other clients. Evalua-
tion of this approach states promising results in achieving desired accuracy
rates.
• FedAttOpt: Attentive Federated Aggregation (FedAttOpt) proposed in
[76] adds an attention-augmented mechanism to model aggregation at the
pro
central server of FL which calculates the attention score based on the
contribution of each client. The attention score or the contribution factor
of the client is calculated based on the gap in common knowledge between
client and global model. FedAttOpt utilises attention score to help client
node to train and accumulates useful common knowledge to all the client
nodes.
• Asynchronous FL Training Rounds: The higher the number of clients the
higher the risk of communication bottlenecks and computational costs.
re-
There are few research works that address communication efficiency by
targeting minimal communication costs during the training rounds of FL.
Research work in [77] explores strategies to minimize communication costs
in FL by using the layer-wise asynchronous update on neural networks.
Similarly, TOR [78] is based on the asynchronous aggregation of models
and few strategies to perform even with low-communication bandwidth.
Asynchronous aggregation implies to the process where the central server
lP
waits for the client device to be online to share updates asynchronously.
The authors in [79] explore the Co-op algorithm, which is designed for
FL in asynchronous implementation, where aggregation of local models is
done on an offline basis based on the availability of clients.
• Communication costs: The approach proposed in [80] sets predefined
rules for selecting client updates during FL training rounds which helps
rna
16
Journal Pre-proof
of
ductionized version in Gboard, which enables Android mobile users to
predict the next word while using the keyboard on their mobile phones
[82, 83, 84, 85, 86], is one of the first attempts in the community to bring
FL into reality. TFF provides integration with Google Kubernetes En-
gine (GKE) [87] or a Kubernetes cluster for orchestrating interaction with
pro
clients and the central server of FL. It also provides docker images to de-
ploy a client and connect through gRPCs [88] calls. TFF uses FL specific
training datasets generated using LEAF [89] provided in Tensorflow APIs.
• PySyft: written in Python on top of the PyTorch framework, Pysyft
(Pysyft 6 ) provides a virtual hook for connecting to clients through a
WebSocket port [90, 91]. An aggregator or orchestrating server maintains
pointers to the ML model and sends it to each participating client to train
with their local data and gets it back for federated averaging. Federated
re-
averaging algorithm averages the model weights and scales them to main-
tain consistency in irregular coverage of data across clients. Apart from
the basic approach of FL, PySyft provides support for asynchronous and
synchronous approaches of FL and integration with existing encryption
strategies like differential privacy.
• FATE: from Webank developers called FATE (FATE 7 ), which is being
improvised with every release. FATE provides a framework to imple-
lP
ment FL in Horizontal, vertical, and transfer learning modes. It can be
implemented with either docker images or manual steps. There is an
open-source GitHub code, which provides training datasets for simulating
known attacks on FL. The authors in [92, 57] utilize such malicious user
datasets to explore the impact of attacks in FL. This framework provides
production-ready APIs with Kubernetes integration.
rna
There are more frameworks [93, 94, 95, 96, 97] for implementing FL that are
being experimented with. Based on the experience of researchers in the field,
it seems as though PySyft provides a more stable environment to develop FL
solutions as opposed to TFF which is in an early experimental stage.
5 TFF https://www.tensorflow.org/federated/federated_learning
6 Pysyfthttps://blog.openmined.org/tag/pysyft/
7 FATE https://fate.fedai.org
17
Journal Pre-proof
of
proposing frameworks to mitigate the risks is very limited.
We set out to investigate the following research questions on the security
aspect of the FL8 :
• RSQ1: What are the source of vulnerabilities in FL ecosystem?
pro
• RSQ2: What are the security threats/attacks in FL domain?
• RSQ3: What are the unique security threats to FL in comparison to
distributed ML solutions?
• RSQ4: What are the defensive techniques for security vulnerabilities of
FL?
In the following sections, we discuss the results based on each research ques-
tion and provide an analysis of the strengths and weaknesses of the current
re-
works.
Our results show that there are five various sources, listed below, that can be
considered as weak points for exploitation.
• Communication Protocol : FL implements an iterative learning process
with randomly selected clients which involves a significant amount of com-
munication over a given network. The FL approach suggests a mixed net-
work [78], which is based on public-key cryptography that keeps source
and message content anonymous throughout the communication. As FL
has more rounds of training, a non-secure communication channel is an
Jou
open vulnerability.
• Client Data Manipulations: FL in a larger landscape has numerous clients
that are open for attackers to exploit model parameters and training data.
Access to the global model may be further vulnerable to data reconstruc-
tion attacks.
18
Journal Pre-proof
of
secure, the central server is responsible for sharing initial model parame-
ters, aggregating local models and sharing global model updates to all the
clients. The cloud-based or physical server picked for this job should be
checked to ensure that open vulnerabilities of the server are not exploited
by curious attackers.
pro
• Weaker Aggregation Algorithm: The aggregation algorithm is the central
authority. In other words, as the local model’s update, it should be in-
telligent to identify abnormality with client updates, and it should have a
configuration to drop updates from suspicious clients. Failing to configure
a standardized aggregation algorithm will make the global model vulner-
able. Section 3.4 gives a list of algorithms proposed for FL aggregation
logic.
• Implementer’s of FL Environment: Intentionally or unintentionally a team
re-
of architects, developers, and deployers involved in the implementation of
FL may turn-out to be a source of a security risk. Either due to the con-
fusion or lack of understanding in what is consider as sensitive user data
and what is not can be the reason for the breach of security and privacy.
The risk from the implementers may be due to the basic fact that they
miss taking proper measures to scan for sensitive data and plan the usage
of it, hiding of the facts while taking consent of users on data usage.
lP
4.2. RSQ2: What are the security threats/attacks in FL domain?
Threat/attack is the possibility of a vulnerability being exploited by a ma-
licious/curious attacker impacting the system security and violating its privacy
policies. In FL, generally, the malicious agent utilizes vulnerabilities [99] to
take control of one or more participants (i.e., clients) in order to manipulate
rna
the global model ultimately. In such a scenario, the attacker targets different
clients with hopes of accessing local data at rest, training procedure, hyper-
parameters, or updated weights in transit [100] to modify and launch attacks in
the global model. The security threats/attacks are classified and their respective
descriptions are discussed in the following sub-sections.
4.2.1. Poisoning
An attack with a major possibility of occurrence in FL is known as poisoning
[101, 102], as each client in FL has access to the training data the possibility
Jou
of getting tampered data weights added to the global ML model is very high.
Poisoning can occur during the training phase and can impact either the training
dataset or the local model in-turn/indirectly tampering the global ML model
performance/accuracy. In FL, model updates are taken from a large group of
clients. That is, the probability of poisoning attacks from one or more clients’
training data is high so is the severity of the threat. Poisoning attack targets
at various artifacts in the FL process. Next we present a brief description of
poisoning attack classifications:
19
Journal Pre-proof
of
gorithms was for the first time presented by authors in [103] where the
attacker targets the vulnerability of the support vector machines algo-
rithm and tries to incorporate malicious data points in the training phase
in hopes of maximizing the classification error. Since then, a wide variety
of approaches have been proposed to mitigate data poisoning attacks in
pro
ML algorithms in different settings including centralized and distributed
environments respectively. While the FL environment enables clients to
actively contribute to training data and sending model parameters to the
server, it provides this opportunity for malicious clients to poison the
global model by manipulating the training process. Data poisoning in FL
is defined as generating dirty samples to train the global model in hopes
of producing falsified model parameters and send them to the server.
Data injection can be also considered as a subcategory of data poisoning
where the malicious client may inject malicious data into client’s local
re-
model processing. As a result, the malicious agent can take control over
multiple client’s local models and ultimately manipulate the global model
with malicious data.
• Model Poisoning: While in data poisoning the malicious agent aims to
manipulate the global model using fake data, in model poisoning, the
malicious agent targets the global model directly. Model poisoning attacks
lP
have been shown to be more effective compared to data poisoning attacks
in recent researches [104, 100, 105]. In fact, the effectiveness of model
poisoning attacks tends to rise when there is a large-scale FL product in
a place with many clients. In general, in the model poisoning attack, the
malicious party can modify the updated model before sending it to the
central server for aggregation and as a result, the global model can be
easily poisoned.
rna
4.2.2. Inference
Inference attacks are more of a threat to privacy (as pointed in Section
5.1) yet we are including it here for overall comparison of threats in FL. The
severity of inference attacks is highly similar to poisoning attacks as there is
a very high possibility of inference attacks from either the participants or a
malicious centralized server in the FL process.
20
Journal Pre-proof
of
Poisoning and Inference attacks are more transparent compared to backdoor
attacks. A backdoor attack is a way to inject a malicious task into the existing
model while retaining the accuracy of the actual task. It is difficult and time-
consuming to identify backdoor attacks as the accuracy of the actual ML task
may not get impacted right away. The authors in [92, 100] experiment on
pro
how backdoor attacks are implemented. Furthermore, the authors in [108, 109]
suggest model pruning and fine-tuning as a solution to mitigate the risks of
backdoor attacks. The severity of backdoor attacks is high as it takes significant
time to identify the occurrence of the attack. Moreover, the impact of the attack
is high as backdoor attacks have the capability to confuse ML models and predict
the false positives confidently. Trojans threats [110, 100, 111, 112, 113] are a
similar category of backdoor attacks which try to retain the existing task of ML
models while performing a malicious task in stealth mode.
4.2.4. GANs
re-
Generative Adversarial Network-based attacks in FL have been experimented
and analyzed by many researchers [114]. With their ability to launch poisoning
and inference attacks, GAN based attacks pose a threat to both the security
and privacy of a given system, discussed more in Section 5.1.3. Research work
in [115] demonstrates how GANs can be used to get training data through in-
ference and use GANs to poison the training data. As all the possibilities of a
lP
GAN based threat cannot be foreseen, it is categorized as a high impact and
prioritized threat. More information on GAN based attacks can be found in
[116, 117]
In cross-device FL, most of the work is done at the central server. From
selecting the model parameters to deploying the global model. Compromised
or malicious servers have a huge impact, and honest but curious or malicious
servers can easily extract private client data or manipulate the global model to
utilize shared computational power in building malicious tasks in the global ML
model.
21
Journal Pre-proof
of
One of the challenges in training an ML model from multiple heterogeneous
device’s data is communication bandwidth. In the FL approach, the commu-
nication cost is reduced by transferring trained models instead of sending huge
amounts of data but still, we do have the need to preserve communication band-
width. There are few algorithms [78, 1, 118] based on asynchronous aggrega-
pro
tion of models and few strategies to perform well even with low-communication
bandwidth. There are various research studies [119, 120] on preserving com-
munication bandwidth in FL environment, discussed in section 3.4.2 . The
severity of this threat is high, as communication bottlenecks can disrupt the FL
environment significantly.
4.2.10. Eavesdropping
In FL, we have an iteration of the learning process which involves commu-
nication rounds from clients to the central server. Attackers may eavesdrop
and extract data through a weak communication channel if one exists. Eaves-
dropping can be considered a medium severity threat on attacking FL models,
since black-box models in general are tough to attack. Attackers would rather
takeover a client with weaker security which will readily offer model parameters
Jou
22
Journal Pre-proof
4.3. RSQ3 What are the unique security threats to FL in comparison to dis-
of
tributed ML solutions?
Distributed Machine Learning (DML) solutions proposed so far aim to solve
challenges of Big Data and computational power while training the ML model.
Data and computation power are shared to train a common ML model. A
parameter server or multiple server nodes are configured which distribute data
pro
or assign a task to client nodes of DML. From an architecture perspective,
DML shares a few common properties with FL and there are research works
addressing security and privacy concerns in DML. However, FL is unique from
existing DML solutions and by default comes with higher security and privacy
levels. This section aims to discuss unique threats of FL and common threats
shared between FL and DML. This helps to understand existing work on DML
[123, 124], and explore reusable/adaptable research ideas from DML to FL.
Security and privacy risks specific to DML are out of scope in this paper and
are not discussed, we focus only on common risk factors of FL and DML.
re-
Table 1 presents an in-depth summary of the classification of security threats/
attacks. Poisoning threat, Backdoor attacks are for both DML [125] and FL as
the data at client nodes can be modified at client nodes which are common in FL
and DML architecture. Parameter server [126, 127] of DML and Central server
is prone to attacks leading to breach in security. Communication bottleneck
threat in DML and FL requires much attention as both the frameworks need to
communicate with their respective client nodes.
lP
Table 1: Threats in FL
Threats Severity ML Framework Source of Vulnerability
Poisoning High DML/FL Client Data Manipulations, Compromised Central Server
Inference High FL Client Data Manipulations, Compromised Central Server
Backdoor Attacks High DML/FL Client Data Manipulations
GANs High FL Client Data Manipulations, Compromised Central Server
Malicious Server High DML/FL Compromised Central Server
Communication bottlenecks High DML/FL Weaker Communication bandwidth
rna
the global model can be poisoned from many sources like local model updates,
malicious servers, and many others. The higher the possibility of the threat,
the higher the impact of the attack on FL. Similar to the poisoning threat, the
Inference threat has a high severity since the source of vulnerabilities that it
can launch is fairly diverse. The threat of backdoor attacks is of higher severity
as well since it is difficult to identify such an attack and the impact has the
possibility of destroying the authenticity of the global model. GAN-based at-
tacks also have high severity as well due to their unpredictability and capability
23
Journal Pre-proof
of
pro
Figure 7: Severity of Threats in FL
re-
to impact the security and privacy of user data. Malicious server threat is of
higher severity, due to the open vulnerability of cloud-based/physical servers.
System disruption as well as IT downtime is tagged with low severity level as
the impact would be less with every client holding the global model individu-
ally. Communication bottleneck in FL is a well-researched topic as it can be a
showstopper in the learning process. The impact and priority of this threat are
lP
always considered high in the FL environment with a huge number of clients
posting updates with every iteration of the learning process. The free-riding
threat impact is set as medium as this can be possible in fewer scenarios and
the impact of it is not severe as the learning process continues with other clients
anyway. Unavailability can halt the learning process where the aggregation al-
gorithm is not robust to handle dropouts, the impact of the threat is tagged as
the medium as there are algorithms [66], which are designed to handle dropouts.
rna
Eavesdropping has medium severity and priority as well. The reason for this is
that from the adversary point of view eavesdropping through a communication
channel would be the least preferred way to steal information as there is an
option to take over a client terminal and gain readily available sets of training
data, global model parameters and the global model itself. Interplay with data
protection laws can be an initiation loophole which can be created knowing or
unknowingly by implementers of FL environment, and the impact stands low as
the possibility is less and rectifiable.
Jou
4.4. RSQ4: What are the defensive techniques for security vulnerabilities of FL?
Defense techniques help to protect against the known attacks and reduce the
probability of risks. There are two classifications of defenses, namely proactive
and reactive. Proactive defense is a way to guess the threats and risks associ-
ated with and employ a cost-effective defense technique. The reactive defense
is later work done when an attack is identified and as part of the mitigation
process, where a defense technique is deployed as patch-up in the production
24
Journal Pre-proof
environment. Apart from the defense technique mentioned, there exists addi-
of
tional research work to enhance FL security capability with new add-on tech-
nologies/algorithms. Examples of such enhancements are the integration of FL
with one more successful technology such as blockchain [128]. Many research
works [129, 130, 131] target the combination of FL with one more decentral-
ized applications, mainly blockchain, which provides an immutable ledger to
pro
save and persist information. In FL, the use of blockchain technology serves
mainly two purposes. One is to provide incentives for major contributors to the
global ML model as a motivation to get more contributors from clients. The
second purpose is to save global model parameters, weights on a blockchain
ledger to ensure the security of the global ML model. In [129, 130], the authors
demonstrate how coordination and trust among federated clients are achieved
through the use of blockchain technology. As for the incentive mechanism con-
cept where ML training administrative hub encourages clients to proactively
contribute to the learning process by giving incentives, there exists ongoing
research [132, 133, 134, 135].
re-
Table 2 summarizes the current defense techniques for FL and the types of
threats they mitigate.
• Sniper :
Different poisoning attacks including data and model poisoning attacks
have been investigated in a centralized setting in FL for a long time.
Jou
However, there are only a few works that explore the poisoning attacks
in a distributed environment where multiple malicious participants with
the same attack goal aim to incorporate poisoned training samples to
the training procedures. Although it seems that distributed poisoning
attacks are a bigger threat in FL, the effectiveness of distributed poisoning
with multiple attackers is still unclear when compared to the traditional
poisoning with a single attacker. Recently, the authors in [136] proposed
the Sniper approach where it can recognize legitimate users and decrease
25
Journal Pre-proof
of
attackers are involved.
• Knowledge distillation: It is a variant of the model compression technique,
where a fully trained neural network transfers knowledge to a small model
step by step on what needs to be done. Knowledge distillation saves
computational cost involved in training a model. The concept of sharing
pro
the knowledge only instead of model parameters can be leveraged in FL
to enhance the security of the client data. The authors in [137] proposed
a federated model distillation framework, which provides flexibility to use
personalized ML models and uses translators to collect knowledge to be
shared with each client.
rejection) & ERR (Error rate based rejection (ERR)) which are inspired
from existing ML defenses such as RONI (reject on negative impact) [152]
& TRIM [153] to identify negatively impacting updates from clients.
• Moving target defense: The moving target defense concept was introduced
in the U.S. National Cyber Leap Year Summit in 2009 [154]. The Federal
Cybersecurity Research and Development Program [155] defined moving
target defense as a way to deploy diverse mechanisms and strategies that
continually change over time in order to increase the cost and complex-
Jou
ity for attackers. Moving target defense also increases system resiliency
while limits the disclosure of system vulnerabilities and opportunities for
attack. Moving target defense is the best of its kind to protect against
intrusion at server-level, network-level, and application level. It is a proac-
tive defense architecture built to obscure the vulnerability source from the
attackers. defense techniques like IP hopping, pooling of virtual IPs for
DNS pings are efficient approaches to moving target defense techniques.
26
Journal Pre-proof
of
vent eavesdropping-based attacks.
• Federated MultiTask Learning: Federated Learning provides training ML
models collaboratively over a large number of mobile devices consider-
ing their local data privacy. This setting can be also extended to a fed-
erated multi-task learning environment where multi-task learning drives
pro
personalized but shared models among devices. For example, the MOCHA
Framework proposed in [146] is designed to speed up the learning process
with fault tolerance, making it able to work even in the scenarios of de-
vice drop. This framework is proposed to address statistical and system
challenges like high communication cost, stragglers and fault tolerance is-
sues in the FL environment. Their experiments show that the framework
can handle variability in system heterogeneity and being robust to sys-
tem drop. Few other works on federated Multi-task learning are given
re-
in [156, 157, 158, 159].
• Trusted Execution Environment (TEE): A Trusted Execution Environ-
ment (TEE), in general, is defined as a high-level trusted environment
for executing code. This technique has been also utilized for privacy-
preserving in different ML models where private areas of computing re-
sources are isolated for a particular task [160]. This approach is also
applicable in federated learning where we have very limited computing
lP
resources. TEE is a tamper-resistant processing environment which pro-
vides integrity and confidentiality of the code executed in the secure area
of the main processor. Authors in [143] discuss this proactive defense tech-
nique. The authors in [144, 161] propose a framework that uses TEE in
the federated environment.
• Data Sanitization: Sanitizing training data proposed by [142] for the first
rna
fied updates to the central server. This type of attack from compromised
clients can break the security and authenticity of the FL environment.
Authors in [145] propose a Foolsgold approach that is efficient against
Sybil-based, label flipping, and backdoor poisoning attacks.
• Pruning: It is a technique in FL that minimizes the size of the ML model
to reduce the complexity and improve the accuracy. In FL, clients have
relatively low computational power and communication bandwidth. The
problem arises when we have large-sized deep neural networks to train
27
Journal Pre-proof
of
power and communication bandwidth when compared to machines in huge
data centers. To address such issues authors in [108, 109] propose a prun-
ing technique. As it is not required to share the full-fledged model in
this approach, it helps to address backdoor attacks and communication
bottlenecks more efficiently.
pro
One other defense technique proposed for the security of FL that is GAN
based is called PDGAN which is proposed in [165] to help defend data poisoning
attacks from malicious clients of FL. There has also been a defense technique
using a vertical federated learning approach as proposed in [166] known as the
Secureboost Framework which is based on gradient boosted tree algorithm.
In the following sections, we discuss the results based on each research ques-
tion obtained from a thorough analysis of all current works in the FL domain.
Jou
28
Journal Pre-proof
partially reveal each participants’ training data in the original training dataset
of
based on their uploaded parameter. Such critical threats in FL can be general-
ized into different categories of inference based attacks
pro
details. Membership Inference attack aims [167] to get information by checking
if the data exists on a training set. The attacker misuses the global model
to get information on the training data of the other users. In such cases, the
information on the training data set is inferred through guesswork and training
the predictive model to predict original training data. The authors in [107]
explore the vulnerability of the neural network to memorize their training data
which is prone to passive and active inference attacks.
framework for exploring GAN-based attacks on FL. mGAN-AI attacks are ex-
perimented on a malicious central server of the FL environment. It explores
user-level privacy leakage against the federated learning by the attack from a
malicious server. mGAN-AI framework’s passive version analyzes all the client
inputs and the active version works on isolating a client by sending the global
update to only to the isolated instance. The inference attack gains the high-
est accuracy with mGAN-AI framework because it does not interfere with the
training process.
Jou
29
Journal Pre-proof
5.2. RPQ2: What are the techniques to mitigate identified threats in RPQ1 and
of
enhance the general privacy-preserving feature of FL?
This section covers the mitigation strategies for the identified privacy threats.
The approach of retaining data at the client device level is the major built-
in privacy feature of FL. The state-of-the-art algorithms to enhance privacy-
preserving and mitigate threats in FL are mainly based on two categories: Secure
pro
Multi-party Computation (SMC) and Differential Privacy (DP).
ential privacy and encrypting the model update. Experiments performed with
this approach claim to have high accuracy while ensuring protection against the
honest but curious server and other users of FL. The authors in [173] propose a
privacy-enhanced FL approach, which provides privacy even after completion of
the training process. This approach is similar to the integration of encryption
and client level DP proposed in [172].
Therefore, as a promising solution, there are several remaining challenges
for SMC based solutions. The main challenge is the trade-off between efficiency
and privacy. SMC based solutions need more time expense than typical FL
Jou
frameworks which may negatively affect the model training. Such a problem
will be enhanced in data-freshness aware training tasks since longer training
time means data value loss. Besides, how to design a lightweight SMC solution
for FL clients is still an open problem.
30
Journal Pre-proof
of
protected. Meanwhile, the statistic data quality loss caused by the added noise
of each user is relatively low compared with the increased privacy protection. In
FL, to avoid inverse data retrieval, DP is introduced to add noise to participants’
uploaded parameters. DPGAN framework is proposed in [175] that utilizes DP
to make GAN based attacks inefficient in inferencing training data of other users
pro
in a deep learning network. Similarly, there is DPFedAvgGAN framework [176]
for FL specific environments. The authors in [177] explain about DP benefits
and definitions of DP properties such as randomization, composition, and ex-
emplifies DP implementation in a multi-agent system, reinforcement learning,
transfer learning, and distributed ML. Both works in [178, 172] combine secure
multiparty computation and differential privacy to achieve a secured FL model
with high accuracy. The authors in [179] improve privacy guarantee of the FL
model by combining the shuffling technique with DP and mask user data with an
invisibility cloak algorithm. However, such solutions bring uncertainty into the
re-
upload parameters and may harm the training performance. Furthermore, these
techniques make the FL server more difficult to evaluate the client’s behavior
to calculate payoff.
5.2.3. VerifyNet
VerifyNet [180] is a privacy preserving and verifiable FL framework. It gets
listed as a preferred mitigation strategy to preserve privacy as it provides double-
lP
masking protocol which makes it difficult for attackers to infer training data.
It provides a way for clients to verify central server results which ensures the
reliability of central server. Apart from providing security and privacy enhance-
ments, the VerifyNet framework is robust to handle multiple dropouts. Only
issue with this framework is the communication overhead as the central server
has to send verifiable proofs with each client.
rna
31
Journal Pre-proof
sharing techniques as well as proposes secure boost protocol for training rounds
of
and a secure predict protocol for secure prediction on the trained global model.
There are also a few defense techniques based on GANs such as in [183] called
Anti-GAN which helps to prevent inference attacks with the use of WGANs
[184, 185] to generate fake training data. This occurs at each client node which
makes it difficult to inference actual training data from the global ML model.
pro
Similar defense techniques of fake data generation at client node is proposed
with FedGP [186] with a slight modification to allow data share instead of
model sharing.
5.3. RPQ3: What are the unique privacy threats to FL compared to distributed
ML solutions?
FL offers user-data privacy by default resulting in very few privacy threats
that are specific to FL. As discussed and experimented in [187], FL performs
well in comparison with DML in terms of protecting the privacy of user data.
re-
In DML solutions with a parameter server, launching an inference attack (as
discussed in Section5.1) to steal information from other clients will be the least
preferred approach as the data is readily accessible either on the parameter
server or through the updates from clients. However, for DML applications,
e.g., [8], where a well-trained ML model is outsourced as a paid service there is
a high possibility of inference based attack [188, 189].
GANs based inference attacks (discussed in Section 5.1.3) are feasible in the
lP
FL environment but a less suitable approach for DML solutions. This makes
GANs based inference attacks specific to only FL. GANs involve two neural net-
works to launch an attack and DML provides space for attackers to launch more
simple and straightforward attacks. GANs are complex and demand more com-
putational resources which makes it a less preferred approach in DML solutions.
However, in FL’s high-level privacy environment GANs based inference attacks
[114] are worth the efforts yielding expected results which wouldn’t have been
rna
achievable with less complex strategies proposed for DML. Detailed strategies of
inference attacks in FL and ML with centralized data solutions are discussed in
[107]. The findings from this particular research suggest the possible strategies
of passive and active inference attack in stand-alone ML and FL environment.
In the FL version, an attacker can be at an aggregator server or a client observ-
ing updates to the global model trying to infer data from other clients. And
in the standalone version, an attacker utilizes BlackBox inference to retrieve
sensitive information.
Jou
Few FL proposals [190] include buyers/end-users who only leverage the final
product of FL without involving in FL training rounds. For such 3rd party end-
users of FL existing research work on the privacy of standalone and distributed
ML solutions can be leveraged and adapted in FL for protecting the privacy
of the hosted ML model. We recommend readers to sections 5.2 and 5.4 to
explore the defenses and costs proposed for add-on privacy of FL, and combine
additional privacy protection measures if required. Moreover, the training phase
of FL provides much-needed privacy to the user data by skipping most of the
32
Journal Pre-proof
privacy threats observed in DML solutions making the existing work on the
of
privacy of DML less suitable for FL adaption.
5.4. RPQ4: What is the associated cost with the privacy-preserving techniques
identified in RPQ2?
Every add-on enhancement comes with its own set of additional costs and
pro
implications. The cost here defines an overhead or a consequence incurred due
to the enhancement approach implemented. Secure Multi-party Computation
and Differential Privacy boost the privacy protection capability of FL, but at the
expense of higher cost with respect to accuracy and efficiency. In cryptography-
based methods in Secure Multi-party Computation, each client is required to
encrypt all the uploaded parameters. Therefore, each client needs to spend
additional computational resources to perform the encryption. This could be
concerning if the client device is computationally constrained such as is com-
mon with IoT devices. Therefore, to enhance the privacy of user-data using
re-
encryption, the efficiency of the ML model could be compromised as a trade-off.
In an empirical work [191] related to cost analysis of FL, the authors simulate
on Reddit datasets to understand the accuracy of DP-enabled FL global model.
Their experimental results show varied results, DP-FL, and nonDPFL environ-
ments show almost the same accuracy for the dataset with similar vocabulary
size. For experiments with participants who have varying lengths of vocabulary
size, DP-FL accuracy is less when compared to nonDPFL. Experiments from of
lP
this paper conclude that DP in the heterogeneous environment has a negative
impact on accuracy.
DP based methods add noise to the uploaded parameters to enhance privacy
protection during the communication and on the server. However, the added
noise will perturb the accuracy inevitably. And it may further influence the
convergence the global aggregation. That means there is a trade-off between
privacy protection strength and accuracy loss as well as efficiency (convergence
rna
time). If the FL model preserves more privacy, it loses more accuracy and costs
more time to converge. On the contrary, if the FL model needs to preserve a
certain degree of accuracy or converge time, it needs to estimate if the privacy
protection level is acceptable or not. Table 3 summarizes the privacy-preserving
techniques and their associated characteristics discussed in this section.
Efficiency loss
Secure Multi-party Computation Encrypt uploaded parameters [170, 173, 192]
due to encryption
Accuracy loss
Differential Privacy Add random noise to uploaded parameters [191, 174, 177, 193]
due to added noise in client’s model
Subdued cost
Hybrid Encrypt the manipulated parameter [194, 178, 172]
on both efficiency and accuracy
Double-masking protocol
VerifyNet Communication overhead [180]
Verifiable aggregation results
Computation power,
Include adversarial samples
Adversarial Training training time for [181, 195, 196]
in training data
adversarial samples
33
Journal Pre-proof
of
There have been a handful of survey-like studies proposed in the literature,
which aim to explore aspects limited to challenges/problems, architecture style
of FL for different domains/use-cases of Federated Learning by providing col-
lective insights and views.
pro
Authors in [197] focus on four challenges in FL which are expensive commu-
nication, systems heterogeneity, statistical heterogeneity, and privacy concerns.
The authors suggest ways to deal with each challenge but lack in-depth cate-
gorization of possible breaks in the integrity of FL environment. The authors
in [198] categorize existing FL models and provides a summary of each cate-
gory. It emphasizes building a robust FL environment by evaluating the issues
around FL from the system perspective. A taxonomy of FL is defined based
on data privacy levels, machine model, data partition across domains and basic
building blocks of FL. The focus of the paper is limited to categorizing FL from
re-
different perspectives with a brief overview of privacy. Another survey work
in [199] gives a detailed architecture and implementation of FL in horizontal,
vertical, and transfer learning approaches with relevant uses cases. This work
gives a great summary of definitions and information on existing research works
in the FL space with details on implementing encryption techniques utilized in
the vertical federated learning approach.However, the focus stays on implemen-
tation details of FL in different scenarios and lack insights into possible risks in
lP
FL.
There are few papers focusing on domain-specific areas related to FL, which
aim only to introduce FL in different real-world domains with possible useful
use-case scenarios. One such work is a comprehensive survey on FL in mobile
devices presented in [200] which gives insights of FL with mobile edge networks
and issues such as communication cost, privacy & security of data, and resource
allocation in mobile IoT devices specific to FL implementation. The authors
rna
34
Journal Pre-proof
of
communication efficiency and cost in each training round as well as federated
fairness to avoid bias [207] in FL are other suggested areas of future research.
Although open problems of FL are explored in detail in this paper, the authors
do not discuss the privacy and security risks of FL in depth as we did in this
work.
pro
In summary, the current state-of-the-art research surveys and reviews in the
field provide remarkable work from various perspectives of FL with different
goals and concentration. Despite their importance, the examination of security
and privacy of FL has not been sufficiently addressed in the literature, resulting
in a missing piece of work with an organized review of studies to dive deep into
the cybersecurity aspect of the FL. Our research work endeavored to provide
this focused research to fill the gap in an attempt to help the community and
newcomers with in-depth information and knowledge of FL security and privacy.
re-
7. Future Directions in FL Security and Privacy
Federated learning has a set of challenges that need further research. Based
on our observations and results, we have identified the following issues that
could form future avenues of research.
logical reality and blindly rely on human-made AI. There are a few preliminary
works leveraging blockchain technology [210, 211, 212] with FL to provide and
trace transaction updates to the global ML model [129, 130, 131], hoping to
achieve a more transparent tracing of the training process in deep learning ML
models.
35
Journal Pre-proof
of
FL is a fairly new approach that requires a detailed analysis of all the pros
and cons tagged with different approaches. Standardized techniques need to
be defined to support the emerging requirements of FL in different domains.
As privacy is a key factor in FL further research needs to be done focusing
on enhancing privacy and standardizing approaches for each requirement and
pro
define a process (with generic APIs) to implement such enhanced approaches.
in FL. Research work in [213] suggests optimal approaches, but still there is a
need to have a standardized approach for each ML algorithm use-case in FL.
36
Journal Pre-proof
of
Research work in [214] proposes an optimal strategy that helps the central
server to set an optimal trigger point to stop/restart the training rounds. Similar
research work needs to be done with respect to different models & domains of ML
applications which can help in understanding FL specific hyperparameters and
possible trigger conditions to configure in FL training rounds. As FL training
pro
rounds are time, cost and computational consuming having vision on setting
optimal values will help in establishing robust and cost-efficient FL solutions.
8. Conclusion
References
[3] White house report. consumer data privacy in a net- worked world: A
framework for protecting privacy and promoting innovation in the global
digital economy. journal of privacy and confidentiality, 2013 (2013).
37
Journal Pre-proof
of
URL https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=
CELEX:32016R0679
[5] A. Narayanan, V. Shmatikov, Robust de-anonymization of large sparse
datasets, in: 2008 IEEE Symposium on Security and Privacy (sp 2008),
2008, pp. 111–125. doi:10.1109/SP.2008.33.
pro
[6] J. X. Chen, The evolution of computing: Alphago, Computing in Science
& Engineering 18 (4) (2016) 4.
[7] State of art papers in ml, web.
URL https://paperswithcode.com/sota
[8] M. Ribeiro, K. Grolinger, M. A. M. Capretz, Mlaas: Machine learning
as a service, in: 2015 IEEE 14th International Conference on Machine
ICMLA.2015.152.
re-
Learning and Applications (ICMLA), 2015, pp. 896–902. doi:10.1109/
38
Journal Pre-proof
of
out-of-vocabulary words (2019).
URL https://arxiv.org/abs/1903.10635
[18] S. Ramaswamy, R. Mathews, K. Rao, F. Beaufays, Federated learning for
emoji prediction in a mobile keyboard (2019). arXiv:1906.04329.
pro
[19] D. Leroy, A. Coucke, T. Lavril, T. Gisselbrecht, J. Dureau, Federated
learning for keyword spotting, ICASSP 2019 - 2019 IEEE International
Conference on Acoustics, Speech and Signal Processing (ICASSP)doi:
10.1109/icassp.2019.8683546.
URL http://dx.doi.org/10.1109/ICASSP.2019.8683546
[20] Nvidia clara federated learning to deliver ai to hospitals while protecting
patient data, web.
URL https://blogs.nvidia.com/blog/2019/12/01/
re-
clara-federated-learning/
[21] S. Niknam, H. S. Dhillon, J. H. Reed, Federated learning for wireless
communications: Motivation, opportunities and challenges (2019). arXiv:
1908.06847.
[22] M. Chen, H. V. Poor, W. Saad, S. Cui, Wireless communications
for collaborative federated learning in the internet of things, ArXiv
lP
abs/2006.02499.
[23] K. Lin, W. Huang, Using federated learning on malware classification, in:
2020 22nd International Conference on Advanced Communication Tech-
nology (ICACT), 2020, pp. 585–589.
[24] K. Sozinov, V. Vlassov, S. Girdzijauskas, Human activity recog-
nition using federated learning, in: 2018 IEEE Intl Conf on
rna
39
Journal Pre-proof
of
for ultra-reliable low-latency v2v communications, in: 2018 IEEE Global
Communications Conference (GLOBECOM), 2018, pp. 1–7.
[29] Y. Lu, X. Huang, Y. Dai, S. Maharjan, Y. Zhang, Federated learning
for data privacy preservation in vehicular cyber-physical systems, IEEE
Network 34 (3) (2020) 50–56.
pro
[30] Y. Liu, J. J. Q. Yu, J. Kang, D. Niyato, S. Zhang, Privacy-preserving
traffic flow prediction: A federated learning approach, IEEE Internet of
Things Journal (2020) 1–1.
2002.10619.
[38] F. Sattler, K.-R. Muller, W. Samek, Clustered federated learning: Model-
agnostic distributed multitask optimization under privacy constraints,
IEEE Transactions on Neural Networks and Learning Systems (2020)
1–13doi:10.1109/tnnls.2020.3015958.
URL http://dx.doi.org/10.1109/TNNLS.2020.3015958
40
Journal Pre-proof
of
federated learning (2020). arXiv:2005.01026.
[40] A. Ghosh, J. Chung, D. Yin, K. Ramchandran, An efficient framework for
clustered federated learning (2020). arXiv:2006.04088.
[41] P. Vanhaesebrouck, A. Bellet, M. Tommasi, Decentralized collaborative
pro
learning of personalized models over networks (2016). arXiv:1610.05202.
[42] L. Muñoz-González, K. T. Co, E. C. Lupu, Byzantine-robust federated
machine learning through adaptive model averaging (2019). arXiv:1909.
05125.
[43] Z. Jiang, A. Balu, C. Hegde, S. Sarkar, Collaborative deep learning in
fixed topology networks (2017). arXiv:1706.07880.
[44] J. Daily, A. Vishnu, C. Siegel, T. Warfel, V. Amatya, Gossipgrad: Scalable
re-
deep learning using gossip communication based asynchronous gradient
descent (2018). arXiv:1803.05880.
[45] J. Wang, A. K. Sahu, Z. Yang, G. Joshi, S. Kar, Matcha: Speeding up
decentralized sgd via matching decomposition sampling (2019). arXiv:
1905.09435.
[46] A. Lalitha, O. C. Kilinc, T. Javidi, F. Koushanfar, Peer-to-peer federated
lP
learning on graphs (2019). arXiv:1901.11173.
[47] H. B. McMahan, E. Moore, D. Ramage, B. A. y Arcas, Federated learning
of deep networks using model averaging.
[48] S. Yang, B. Ren, X. Zhou, L. Liu, Parallel distributed logistic regression for
vertical federated learning without third-party coordinator, arXiv preprint
rna
arXiv:1911.09824.
[49] Y. Chen, J. Wang, C. Yu, W. Gao, X. Qin, Fedhealth: A federated trans-
fer learning framework for wearable healthcare, CoRR abs/1907.09173.
arXiv:1907.09173.
URL http://arxiv.org/abs/1907.09173
[50] Y. Liu, T. Chen, Q. Yang, Secure federated transfer learning (2018).
arXiv:1812.03337.
Jou
41
Journal Pre-proof
of
for fast personalization, in: 2019 IEEE Second International Conference
on Artificial Intelligence and Knowledge Engineering (AIKE), 2019, pp.
123–127.
[55] H. Lim, J. Kim, C. Kim, G. Hwang, H. Choi, Y. Han, Federated rein-
forcement learning for controlling multiple rotary inverted pendulums in
pro
edge computing environments, in: 2020 International Conference on Arti-
ficial Intelligence in Information and Communication (ICAIIC), 2020, pp.
463–464.
[56] B. Liu, L. Wang, M. Liu, Lifelong federated reinforcement learning:
A learning architecture for navigation in cloud robotic systems, IEEE
Robotics and Automation Letters 4 (4) (2019) 4555–4562.
[57] Fate framework from webank, web.
re-
URL https://fate.fedai.org
42
Journal Pre-proof
of
evaluation of federated learning algorithms, in: Proceedings of the Sec-
ond Workshop on Distributed Infrastructures for Deep Learning, DIDL
’18, ACM, New York, NY, USA, 2018, pp. 1–8. doi:10.1145/3286490.
3286559.
URL http://doi.acm.org/10.1145/3286490.3286559
pro
[66] K. Bonawitz, V. Ivanov, B. Kreuter, A. Marcedone, H. B. McMahan,
S. Patel, D. Ramage, A. Segal, K. Seth, Practical secure aggregation for
federated learning on user-held data (2016). arXiv:1611.04482.
[67] T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, V. Smith, Feder-
ated optimization in heterogeneous networks (2018). arXiv:1812.06127.
[68] Anonymous, Federated learning with matched averaging, in: Submitted
to International Conference on Learning Representations, 2020, under re-
view.
re-
URL https://openreview.net/forum?id=BkluqlSFDS
887–895. doi:10.1145/3097983.3098118.
URL https://doi.org/10.1145/3097983.3098118
[72] J. Ma, Q. Zhang, J. Lou, J. C. Ho, L. Xiong, X. Jiang, Privacy-preserving
tensor factorization for collaborative health data analysis, in: Proceedings
of the 28th ACM International Conference on Information and Knowledge
Management, 2019, pp. 1291–1300.
[73] M. G. Arivazhagan, V. Aggarwal, A. K. Singh, S. Choudhary, Federated
Jou
43
Journal Pre-proof
of
for mobile internet applications, World Wide Webdoi:10.1007/
s11280-019-00775-w.
[77] Y. Chen, X. Sun, Y. Jin, Communication-efficient federated deep learning
with layerwise asynchronous model update and temporally weighted ag-
gregation, IEEE Transactions on Neural Networks and Learning Systems
pro
(2019) 1–10.
[78] G. D. A. Chaum, David, Untraceable Electronic Mail, Return Addresses
and Digital Pseudonyms, Springer US, Boston, MA, 2003, pp. 211–219.
doi:10.1007/978-1-4615-0239-5_14.
URL https://doi.org/10.1007/978-1-4615-0239-5_14
[79] Y. Wang, Co-op: Cooperative machine learning from mobile devices.
re-
[80] T. Nishio, R. Yonetani, Client selection for federated learning with het-
erogeneous resources in mobile edge, ICC 2019 - 2019 IEEE International
Conference on Communications (ICC)doi:10.1109/icc.2019.8761315.
URL http://dx.doi.org/10.1109/ICC.2019.8761315
[81] E. Jeong, S. Oh, H. Kim, J. Park, M. Bennis, S.-L. Kim, Communication-
efficient on-device machine learning: Federated distillation and augmen-
tation under non-iid private data (2018). arXiv:1811.11479.
lP
[82] Tff:open source for federated learning from google, web.
URL https://www.tensorflow.org/federated/federated_learning
[83] G. Sannino, G. De Pietro, A deep learning approach for ecg-based heart-
beat classification for arrhythmia detection, Future Generation Computer
Systems 86 (2018) 446–455.
rna
44
Journal Pre-proof
of
A. Talwalkar, LEAF: A benchmark for federated settings, CoRR
abs/1812.01097. arXiv:1812.01097.
URL http://arxiv.org/abs/1812.01097
[90] Pysyft: Open source framework for federated learning, web.
URL https://github.com/OpenMined/PySyft
pro
[91] T. Ryffel, A. Trask, M. Dahl, B. Wagner, J. Mancuso, D. Rueckert,
J. Passerat-Palmbach, A generic framework for privacy preserving deep
learning (2018). arXiv:1811.04017.
[92] Z. Sun, P. Kairouz, A. T. Suresh, H. B. McMahan, Can you really back-
door federated learning? (2019). arXiv:1911.07963.
[93] The clara training framework, web.
URL https://developer.nvidia.com/clara
re-
[94] Paddlefl, web.
URL https://github.com/PaddlePaddle/PaddleFL
[95] Uberhorovod, web.
URL https://eng.uber.com/horovod/
[96] G. Ulm, E. Gustavsson, M. Jirstrand, Functional federated learning in
lP
erlang (ffl-erl), in: J. Silva (Ed.), Functional and Constraint Logic Pro-
gramming, Springer International Publishing, Cham, 2019, pp. 162–178.
[97] Federatd learning with crypten, web.
URL https://crypten.ai
[98] Owsap defination for vulnerability, web.
rna
URL https://www.owasp.org/index.php/Category:Vulnerability
[99] J. Men, G. Xu, Z. Han, Z. Sun, X. Zhou, W. Lian, X. Cheng, Finding
sands in the eyes: vulnerabilities discovery in iot with eufuzzer on human
machine interface, IEEE Access 7 (2019) 103751–103759.
[100] E. Bagdasaryan, A. Veit, Y. Hua, D. Estrin, V. Shmatikov, How to back-
door federated learning, in: International Conference on Artificial Intelli-
gence and Statistics, 2020, pp. 2938–2948.
Jou
[101] J. Feng, Q.-Z. Cai, Z.-H. Zhou, Learning to confuse: Generating training
time adversarial data with auto-encoder (2019). arXiv:1905.09027.
[102] L. Muñoz-González, B. Biggio, A. Demontis, A. Paudice, V. Wongras-
samee, E. C. Lupu, F. Roli, Towards poisoning of deep learning algorithms
with back-gradient optimization, Proceedings of the 10th ACM Workshop
on Artificial Intelligence and Security - AISec ’17doi:10.1145/3128572.
3140451.
URL http://dx.doi.org/10.1145/3128572.3140451
45
Journal Pre-proof
of
machines, arXiv preprint arXiv:1206.6389.
[104] A. N. Bhagoji, S. Chakraborty, P. Mittal, S. Calo, Analyzing federated
learning through an adversarial lens, in: International Conference on Ma-
chine Learning, 2019, pp. 634–643.
pro
[105] M. Fang, X. Cao, J. Jia, N. Z. Gong, Local model poisoning attacks to
byzantine-robust federated learning, arXiv preprint arXiv:1911.11815.
[106] A. Shafahi, W. R. Huang, M. Najibi, O. Suciu, C. Studer, T. Dumitras,
T. Goldstein, Poison frogs! targeted clean-label poisoning attacks on neu-
ral networks, in: Advances in Neural Information Processing Systems,
2018, pp. 6103–6113.
[107] M. Nasr, R. Shokri, A. Houmansadr, Comprehensive privacy analysis of
re-
deep learning: Passive and active white-box inference attacks against cen-
tralized and federated learning, in: 2019 IEEE Symposium on Security
and Privacy (SP), 2019, pp. 739–753.
[108] K. Liu, B. Dolan-Gavitt, S. Garg, Fine-pruning: Defending against back-
dooring attacks on deep neural networks, in: International Symposium on
Research in Attacks, Intrusions, and Defenses, Springer, 2018, pp. 273–
294.
lP
[109] Y. Jiang, S. Wang, B. J. Ko, W.-H. Lee, L. Tassiulas, Model pruning
enables efficient federated learning on edge devices (2019). arXiv:1909.
12326.
[110] C. Xie, K. Huang, P.-Y. Chen, B. Li, Dba: Distributed backdoor attacks
against federated learning, in: International Conference on Learning Rep-
rna
resentations, 2019.
[111] Y. Liu, S. Ma, Y. Aafer, W.-C. Lee, J. Zhai, W. Wang, X. Zhang, Tro-
janing attack on neural networks.
[112] M. Zou, Y. Shi, C. Wang, F. Li, W. Song, Y. Wang, Potrojan: pow-
erful neural-level trojan designs in deep learning models, arXiv preprint
arXiv:1802.03043.
[113] A. Koloskova, S. U. Stich, M. Jaggi, Decentralized stochastic optimization
Jou
46
Journal Pre-proof
of
learning using generative adversarial nets, in: 2019 18th IEEE Interna-
tional Conference On Trust, Security And Privacy In Computing And
Communications/13th IEEE International Conference On Big Data Sci-
ence And Engineering (TrustCom/BigDataSE), 2019, pp. 374–380.
[116] Generative adversarial networks, web.
pro
URL https://developers.google.com/machine-learning/gan
[117] B. Hitaj, G. Ateniese, F. Perez-Cruz, Deep models under the gan: In-
formation leakage from collaborative deep learning, in: Proceedings of
the 2017 ACM SIGSAC Conference on Computer and Communications
Security, CCS ’17, ACM, New York, NY, USA, 2017, pp. 603–618.
[118] J. Kone, H. B. McMahan, F. X. Yu, P. Richtrik, A. T. Suresh, D. Bacon,
Federated learning: Strategies for improving communication efficiency
re-
(2016). arXiv:1610.05492.
[119] L. WANG, W. WANG, B. LI, Cmfl: Mitigating communication overhead
for federated learning, in: 2019 IEEE 39th International Conference on
Distributed Computing Systems (ICDCS), 2019, pp. 954–964. doi:10.
1109/ICDCS.2019.00099.
[120] X. Yao, C. Huang, L. Sun, Two-stream federated learning: Reduce the
lP
communication costs, in: 2018 IEEE Visual Communications and Image
Processing (VCIP), 2018, pp. 1–4.
[121] J. Lin, M. Du, J. Liu, Free-riders in federated learning: Attacks and
defenses (2019). arXiv:1911.12560.
[122] B. Zong, Q. Song, M. R. Min, W. Cheng, C. Lumezanu, D. Cho, H. Chen,
rna
[125] Y. Chen, Y. Mao, H. Liang, S. Yu, Y. Wei, S. Leng, Data poison detection
schemes for distributed machine learning, IEEE Access 8 (2020) 7442–
7454.
[126] M. Li, D. G. Andersen, J. W. Park, A. J. Smola, A. Ahmed, V. Josifovski,
J. Long, E. J. Shekita, B.-Y. Su, Scaling distributed machine learning
with the parameter server, in: 11th {USENIX} Symposium on Operating
Systems Design and Implementation ({OSDI} 14), 2014, pp. 583–598.
47
Journal Pre-proof
of
rameter server for distributed machine learning.
[128] P. J. Taylor, T. Dargahi, A. Dehghantanha, R. M. Parizi, K.-K. R. Choo,
A systematic literature review of blockchain cyber security, Digital Com-
munications and Networks 6 (2) (2020) 147 – 156.
pro
[129] H. Kim, J. Park, M. Bennis, S. Kim, Blockchained on-device federated
learning, IEEE Communications Letters 24 (6) (2020) 1279–1283.
[130] U. Majeed, C. S. Hong, Flchain: Federated learning via mec-enabled
blockchain network, in: 2019 20th Asia-Pacific Network Operations and
Management Symposium (APNOMS), 2019, pp. 1–4. doi:10.23919/
APNOMS.2019.8892848.
[131] K. Salah, M. H. U. Rehman, N. Nizamuddin, A. Al-Fuqaha, Blockchain
re-
for ai: Review and open research challenges, IEEE Access 7 (2019) 10127–
10149. doi:10.1109/ACCESS.2018.2890507.
[132] Y. Zhao, J. Zhao, L. Jiang, R. Tan, D. Niyato, Mobile edge computing,
blockchain and reputation-based crowdsourcing iot federated learning: A
secure, decentralized and privacy-preserving system (2019). arXiv:1906.
10893.
lP
[133] L. U. Khan, N. H. Tran, S. R. Pandey, W. Saad, Z. Han, M. N. H. Nguyen,
C. S. Hong, Federated learning for edge networks: Resource optimization
and incentive mechanism (2019). arXiv:1911.05642.
[134] J. Weng, J. Weng, J. Zhang, M. Li, Y. Zhang, W. Luo, Deepchain: Au-
ditable and privacy-preserving deep learning with blockchain-based in-
centive, IEEE Transactions on Dependable and Secure Computing (2019)
rna
1–1doi:10.1109/TDSC.2019.2952332.
[135] J. Kang, Z. Xiong, D. Niyato, S. Xie, J. Zhang, Incentive mechanism for
reliable federated learning: A joint optimization approach to combining
reputation and contract theory, IEEE Internet of Things Journal.
[136] D. Cao, S. Chang, Z. Lin, G. Liu, D. Sun, Understanding distributed
poisoning attack in federated learning, in: 2019 IEEE 25th International
Conference on Parallel and Distributed Systems (ICPADS), 2019, pp. 233–
239.
Jou
[137] D. Li, J. Wang, Fedmd: Heterogenous federated learning via model dis-
tillation (2019). arXiv:1910.03581.
[138] P. Blanchard, E. M. El Mhamdi, R. Guerraoui, J. Stainer, Machine learn-
ing with adversaries: Byzantine tolerant gradient descent, in: I. Guyon,
U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan,
R. Garnett (Eds.), Advances in Neural Information Processing Systems
30, Curran Associates, Inc., 2017, pp. 119–129.
48
Journal Pre-proof
of
in collaborative deep learning systems, in: A uror: defending against
poisoning attacks in collaborative deep learning systems, 2016, pp. 508–
519. doi:10.1145/2991079.2991125.
[140] R. Ito, M. Tsukada, H. Matsutani, An on-device federated learn-
ing approach for cooperative anomaly detection, arXiv preprint
pro
arXiv:2002.12301.
[141] R. Colbaugh, K. Glass, Moving target defense for adaptive adversaries,
in: 2013 IEEE International Conference on Intelligence and Security In-
formatics, 2013, pp. 50–55. doi:10.1109/ISI.2013.6578785.
[142] G. F. Cretu-Ciocarlie, A. Stavrou, M. E. Locasto, S. J. Stolfo, A. D.
Keromytis, Casting out demons: Sanitizing training data for anomaly
sensors, 2008 IEEE Symposium on Security and Privacy (sp 2008) (2008)
81–95. re-
[143] M. Sabt, M. Achemlal, A. Bouabdallah, Trusted execution environment:
What it is, and what it is not, in: 2015 IEEE Trustcom/BigDataSE/ISPA,
Vol. 1, 2015, pp. 57–64. doi:10.1109/Trustcom.2015.357.
[144] F. Mo, H. Haddadi, Efficient and private federated learning using tee.
lP
[145] C. Fung, C. J. M. Yoon, I. Beschastnikh, Mitigating sybils in federated
learning poisoning, CoRR abs/1808.04866. arXiv:1808.04866.
URL http://arxiv.org/abs/1808.04866
[146] V. Smith, C.-K. Chiang, M. Sanjabi, A. S. Talwalkar, Federated multi-
task learning, in: Advances in Neural Information Processing Systems,
2017, pp. 4424–4434.
rna
49
Journal Pre-proof
of
nipulating machine learning: Poisoning attacks and countermeasures for
regression learning, in: 2018 IEEE Symposium on Security and Privacy
(SP), IEEE, 2018, pp. 19–35.
[154] U. national cyber, National cyber leap year summit 2009 cochairs report
(2009 (accessed Jun 21, 2020)).
pro
URL https://www.nitrd.gov/nitrdgroups/index.php?title=File:
National_Cyber_Leap_Year_Summit_2009_CoChairs_Report.pdf
[155] F. C. Research, D. Program, Nitrd csia iwg cybersecurity game-change
research & development recommendations (2014 (accessed Jun 21,
2020)).
URL https://www.nitrd.gov/pubs/CSIA_IWG_%20Cybersecurity_
%20GameChange_RD_%20Recommendations_20100513.pdf
re-
[156] R. Li, F. Ma, W. Jiang, J. Gao, Federated multitask learning, in: 2019
IEEE International Conference on Big Data (Big Data), 2019, pp. 215–
220.
[157] T. Yu, T. Li, Y. Sun, S. Nanda, V. Smith, V. Sekar, S. Seshan, Learning
context-aware policies from multiple smart homes via federated multi-task
learning, in: 2020 IEEE/ACM Fifth International Conference on Internet-
of-Things Design and Implementation (IoTDI), 2020, pp. 104–115.
lP
[158] S. Caldas, V. Smith, A. Talwalkar, Federated kernelized multi-task learn-
ing.
[159] F. Sattler, K.-R. Müller, W. Samek, Clustered federated learning: Model-
agnostic distributed multi-task optimization under privacy constraints
(2019). arXiv:1910.01991.
rna
[162] Y. Shen, S. Sanghavi, Learning with bad training data via iterative
trimmed loss minimization, in: International Conference on Machine
Learning, 2019, pp. 5739–5748.
[163] B. Tran, J. Li, A. Madry, Spectral signatures in backdoor attacks, in:
Advances in Neural Information Processing Systems, 2018, pp. 8000–8010.
[164] P. W. Koh, J. Steinhardt, P. Liang, Stronger data poisoning attacks break
data sanitization defenses, ArXiv abs/1811.00741.
50
Journal Pre-proof
[165] Y. Zhao, J. Chen, J. Zhang, D. Wu, J. Teng, S. Yu, PDGAN: A Novel Poi-
of
soning Defense Method in Federated Learning Using Generative Adversar-
ial Network, 2020, pp. 595–609. doi:10.1007/978-3-030-38991-8_39.
[166] K. Cheng, T. Fan, Y. Jin, Y. Liu, T. Chen, Q. Yang, Secureboost: A
lossless federated learning framework, CoRR abs/1901.08755. arXiv:
pro
1901.08755.
URL http://arxiv.org/abs/1901.08755
[167] S. Truex, L. Liu, M. Gursoy, L. Yu, W. Wei, Demystifying membership
inference attacks in machine learning as a service, IEEE Transactions on
Services Computing PP (2019) 1–1. doi:10.1109/TSC.2019.2897554.
preserving federated deep learning, in: ICC 2019 - 2019 IEEE Inter-
national Conference on Communications (ICC), 2019, pp. 1–6. doi:
10.1109/ICC.2019.8761267.
[173] M. Hao, H. Li, X. Luo, G. Xu, H. Yang, S. Liu, Efficient and privacy-
enhanced federated learning for industrial artificial intelligence, IEEE
Transactions on Industrial Informatics (2019) 1–1.
[174] C. Dwork, Differential Privacy, Springer US, Boston, MA, 2011, pp. 338–
340.
Jou
51
Journal Pre-proof
of
intelligence, in: 2019 IEEE 39th International Conference on Distributed
Computing Systems (ICDCS), 2019, pp. 1601–1609. doi:10.1109/ICDCS.
2019.00159.
[178] S. Truex, N. Baracaldo, A. Anwar, T. Steinke, H. Ludwig, R. Zhang,
Y. Zhou, A hybrid approach to privacy-preserving federated learning, in:
pro
Proceedings of the 12th ACM Workshop on Artificial Intelligence and
Security, AISec’19, Association for Computing Machinery, New York, NY,
USA, 2019, p. 1–11. doi:10.1145/3338501.3357370.
URL https://doi.org/10.1145/3338501.3357370
1701.07875.
[185] I. Gulrajani, F. Ahmed, M. Arjovsky, V. Dumoulin, A. C. Courville,
Improved training of wasserstein gans, in: I. Guyon, U. V. Luxburg,
S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, R. Garnett (Eds.),
Advances in Neural Information Processing Systems 30, Curran Asso-
ciates, Inc., 2017, pp. 5767–5777.
URL http://papers.nips.cc/paper/7159-improved-training-of-wasserstein-gans.
pdf
Jou
52
Journal Pre-proof
doi:https://doi.org/10.1016/j.procs.2020.01.039.
of
URL http://www.sciencedirect.com/science/article/pii/
S1877050920300478
[188] R. Shokri, M. Stronati, C. Song, V. Shmatikov, Membership inference
attacks against machine learning models, in: 2017 IEEE Symposium on
Security and Privacy (SP), 2017, pp. 3–18.
pro
[189] A. Salem, Y. Zhang, M. Humbert, P. Berrang, M. Fritz, M. Backes, Ml-
leaks: Model and data independent membership inference attacks and
defenses on machine learning models, arXiv preprint arXiv:1806.01246.
[190] X. Bao, C. Su, Y. Xiong, W. Huang, Y. Hu, Flchain: A blockchain for
auditable federated learning with trust and incentive, in: 2019 5th Inter-
national Conference on Big Data Computing and Communications (BIG-
COM), 2019, pp. 151–159.
re-
[191] E. Bagdasaryan, O. Poursaeed, V. Shmatikov, Differential privacy has
disparate impact on model accuracy, in: H. Wallach, H. Larochelle,
A. Beygelzimer, F. d'Alché-Buc, E. Fox, R. Garnett (Eds.), Advances
in Neural Information Processing Systems 32, Curran Associates, Inc.,
2019, pp. 15479–15488.
[192] J. Zhang, B. Chen, S. Yu, H. Deng, Pefl: A privacy-enhanced federated
lP
learning scheme for big data analytics, in: 2019 IEEE Global Communi-
cations Conference (GLOBECOM), IEEE, 2019, pp. 1–6.
[193] K. Wei, J. Li, M. Ding, C. Ma, H. H. Yang, F. Farokhi, S. Jin, T. Q. S.
Quek, H. Vincent Poor, Federated learning with differential privacy: Al-
gorithms and performance analysis, IEEE Transactions on Information
Forensics and Security 15 (2020) 3454–3469.
rna
53
Journal Pre-proof
[198] Q. Li, Z. Wen, B. He, Federated learning systems: Vision, hype and reality
of
for data privacy and protection, CoRR abs/1907.09693. arXiv:1907.
09693.
URL http://arxiv.org/abs/1907.09693
[199] Q. Yang, Y. Liu, T. Chen, Y. Tong, Federated machine learning: Concept
and applications (2019). arXiv:1902.04885.
pro
[200] W. Y. B. Lim, N. C. Luong, D. T. Hoang, Y. Jiao, Y. Liang, Q. Yang,
D. Niyato, C. Miao, Federated learning in mobile edge networks: A com-
prehensive survey, IEEE Communications Surveys Tutorials (2020) 1–1.
Parizi, An ensemble of deep recurrent neural networks for detecting iot cy-
ber attacks using network traffic, IEEE Internet of Things Journal (2020)
1–1.
54
Journal Pre-proof
of
blockchain: A secure blockchain-enabled packet parser for software defined
networking, Computers & Security 88 (2020) 101629. doi:https://doi.
org/10.1016/j.cose.2019.101629.
[211] A. Yazdinejad, R. M. Parizi, A. Dehghantanha, K. R. Choo, Blockchain-
enabled authentication handover with efficient privacy protection in sdn-
pro
based 5g networks, IEEE Transactions on Network Science and Engineer-
ing (2019) 1–1doi:10.1109/TNSE.2019.2937481.
[212] E. Nyaletey, R. M. Parizi, Q. Zhang, K.-K. R. Choo, Blockipfs–blockchain-
enabled interplanetary file system for forensic and trusted data trace-
ability, in: 2nd IEEE International Conference on Blockchain (IEEE
Blockchain-2019), 2019.
[213] T. Nishio, R. Yonetani, Client selection for federated learning with hetero-
re-
geneous resources in mobile edge, in: ICC 2019 - 2019 IEEE International
Conference on Communications (ICC), 2019, pp. 1–7.
55
Journal Pre-proof
of
Identifying and evaluating privacy threats, their mitigation techniques,
and the trade-offs cost associated with privacy-preserving techniques in
the federated learning environments.
pro
to enhance the security and privacy of the federated learning implementation.
re-
lP
rna
Jou
Journal Pre-proof
of
Applications. She has a professional certification on Machine Learning to her credit.
pro
security researcher with an entrepreneurial spirit. He is a senior member of IEEE,
IEEE Blockchain Community, and ACM. Prior to joining KSU, he was a faculty at
New York Institute of Technology. He received a Ph.D. in Software Engineering in
2012 and M.Sc. and B.Sc. degrees in Software Engineering and Computer Science
respectively in 2008 and 2005. His research interests are R\&D in decentralized AI, blockchain
systems, smart contracts, and emerging issues in the practice of secure software-run world
applications. re-
Seyedamin Pouriyeh is an Assistant Professor of Information Technology at
Kennesaw State University, GA, USA. He received an M.Sc. in Information
Technology Engineering from Shiraz University, and his Ph.D. in Computer
Science from the University of Georgia in 2009 and 2018 respectively. His
primary research interests span Federated Machine Learning, Blockchain, and
lP
Cyber Security.
University, and B.S., M.S. degrees from Heilongjiang University. His current research
focuses on Cyber Security & Privacy, Federated Learning, and IoT.
Dr. Gautam Srivastava was awarded his B.Sc. degree from Briar Clif
University in U.S.A. in the year 2004, followed by his M.Sc. and Ph.D.
degrees from the University of Victoria in Victoria, British Columbia,
Canada in the years 2006 and 2011, respectively. He then taught for 3
years at the University of Victoria in the Department of Computer
of
Science, where he was regarded as one of the top undergraduate
professors in the Computer Science Course Instruction at the University.
From there in the year 2014, he joined a tenure-track position at Brandon
University in Brandon, Manitoba, Canada, where he currently is active in
pro
various professional and scholarly activities. He was promoted to the rank Associate Professor in
January 2018. Dr. G, as he is popularly known, is active in research in the field of Data Mining
and Big Data. In his 8-year academic career, he has published a total of 43 papers in high-impact
conferences in many countries and in high-status journals (SCI, SCIE) and has also delivered
invited guest lectures on Big Data, Cloud Computing, Internet of Things, and Cryptography at
many Taiwanese and Czech universities. He is an Editor of several international scientific
research journals. He currently has active research projects with other academics in Taiwan,
re-
Singapore, Canada, Czech Republic, Poland and U.S.A. He is constantly looking for collaboration
opportunities with foreign professors and students. Assoc. Prof. Gautam Srivastava received
*Best Oral Presenter Award* in FSDM 2017 which was held at the National Dong Hwa University
(NDHU) in Shoufeng (Hualien County) in Taiwan (Republic of China) on November 24-27, 2017.
lP
rna
Jou
Journal Pre-proof
of
pro
re-
lP
rna
Jou
Journal Pre-proof
Declaration of interests
☒ The authors declare that they have no known competing financial interests or personal relationships
that could have appeared to influence the work reported in this paper.
of
☐The authors declare the following financial interests/personal relationships which may be considered
as potential competing interests:
pro
re-
lP
rna
Jou