Jsan 12 00013
Jsan 12 00013
Sensor and
Actuator Networks
Article
Building Trusted Federated Learning: Key Technologies and
Challenges
Depeng Chen 1,2 , Xiao Jiang 1,2 , Hong Zhong 1, * and Jie Cui 1,2
1 School of Computer Science and Technology, Anhui University, Hefei 230601, China
2 Institute of Artificial Intelligence, Hefei Comprehensive National Science Center, Hefei 230026, China
* Correspondence: [email protected]
Abstract: Federated learning (FL) provides convenience for cross-domain machine learning appli-
cations and has been widely studied. However, the original FL is still vulnerable to poisoning and
inference attacks, which will hinder the landing application of FL. Therefore, it is essential to design a
trustworthy federation learning (TFL) to eliminate users’ anxiety. In this paper, we aim to provide
a well-researched picture of the security and privacy issues in FL that can bridge the gap to TFL.
Firstly, we define the desired goals and critical requirements of TFL, observe the FL model from the
perspective of the adversaries and extrapolate the roles and capabilities of potential adversaries back-
ward. Subsequently, we summarize the current mainstream attack and defense means and analyze
the characteristics of the different methods. Based on a priori knowledge, we propose directions for
realizing the future of TFL that deserve attention.
Keywords: trustworthy federated learning; machine learning; security and defense; privacy protection
1. Introduction
Artificial intelligence technology represented by machine learning strongly drives the
development of various industries. Machine learning (ML) is a paradigm that learns from
past experience and makes accurate predictions about new problems. ML gives machines
the ability to learn with little or even without human intervention. In recent years, ML
Citation: Chen, D.; Jiang, X.; Zhong,
algorithms, represented by deep learning (DL), have achieved great achievement in areas
H.; Cui, J. Building Trusted Federated
such as image recognition and natural language processing. However, researchers find
Learning: Key Technologies and
that high-quality predictive models rely on high-quality training data, yet individuals or
Challenges. J. Sens. Actuator Netw.
2023, 12, 13. https://doi.org/
groups are often reluctant to contribute data due to privacy concerns, resulting in data silos.
10.3390/jsan12010013
At the same time, sufficient attention has been paid to the misuse of privacy, and privacy
laws have further regulated access to data.
Academic Editor: Mingjun Xiao Federated learning (FL) [1] provides an excellent idea for solving these problems.
Received: 13 December 2022 Unlike distributed machine learning, in federated learning, users update the model rather
Revised: 17 January 2023 than their data to obtain a better global model. It assures that data can be utilized without
Accepted: 18 January 2023 leaving the local area, thus dispelling users’ privacy anxiety. Although FL has been partially
Published: 6 February 2023 applied in practice, such as Google, using it to predict the subsequent input of the user’s
keyboard, we found that the current level of FL is still insufficient to meet its security
requirements. Means such as poisoning attacks, and inference attacks, still affect the
usability of FL, especially in combination with highly sensitive information areas such
Copyright: © 2023 by the authors. as medicine and finance. The large-scale application of FL has been hampered by these
Licensee MDPI, Basel, Switzerland. problems, and researchers have had to redesign the model to achieve user trustworthiness.
This article is an open access article Therefore, trustworthy federated learning (TFL) [2], in combination with safety solutions,
distributed under the terms and
deserves to be discussed further.
conditions of the Creative Commons
Unlike traditional FL, TFL’s goal is to eliminate users’ concerns about the security
Attribution (CC BY) license (https://
and privacy of the model system and ensure the credibility of the model framework.
creativecommons.org/licenses/by/
Typically, researchers choose to use security algorithms [3] or secure architecture [4,5], such
4.0/).
as blockchain technology, to achieve TFL. However, the current research lacks a systematic
definition of TFL’s requirements. TFL imposes more stringent safety requirements on FL
systems, which should have the following basic principles:
• High Confidentiality: Confidentiality is reflected in the fact that malicious adversaries
cannot steal sensitive information in FL.
• High Integrity: Integrity is reflected in the fact that private data cannot be maliciously
modified without authorization during training.
• High Availability: The model system is required to provide access by authorized users
and be used on demand. The model also needs to have a usable accuracy rate as well
as efficiency. The cost of trustworthiness cannot be a significant loss of accuracy and a
high rate of loss of efficiency.
• Strong Robustness: In addition to following the information security fundamentals, FL
should have sufficient resistance in the face of complex scenarios or unknown attacks.
• Provable Security: The security protocols and methods must be rigorously secure
based on specific mathematical assumptions.
In response to the above requirements, we survey the current status of FL and look
forward to the next more promising development direction of TFL. Analytical work on
TFL has been partially studied, but we will look at the threats faced by FL from some
new perspectives.
1.1. Contributions
Although similar studies have been conducted to investigate the threat of FL, these
efforts still need to provide a more comprehensive summary of existing technologies
and a clear indication of future research directions. Our work provides a comprehensive
overview of FL, including its definition, threats, and potential future research. This paper
will facilitate the construction of usable TFL paradigms and their rapid application to actual
production. Our main contributions are as follows:
• We thoroughly investigate the development mapping and critical technologies of FL
and meticulously analyze the existing FL research content.
• We assess the threats to FL from an adversary perspective. Furthermore, we sum-
marize mainstream FL-specific attacks from the perspective of security threats and
privacy threats.
• We summarise and abstract the approaches to privacy protection in FL and evaluate
their strengths and weaknesses. Based on this, we provide some valuable prospects
for building TFL.
2. Federated Learning
Similar to the traditional distributed machine learning paradigm, FL also utilizes
the assistance of distributed clients for more complex model training. The target of FL’s
communication shifts from data to the model. The optimized global model is obtained
through the aggregation of multiple local models, which ensures the availability of the
global model while making the client’s data visible only to itself, eliminating its distrust
of the server or external attackers, thus allowing better quality data to participate in the
training [6–8]. In addition, FL has likewise been extensively studied for joint learning
of heterogeneous data [9,10], which makes it a good prospect for cross-domain machine
learning. For example, information from different devices (e.g., images and text) can be
used for training at the same time to capture more feature information and provide a
J. Sens. Actuator Netw. 2023, 12, 13 3 of 18
more valuable output model. The work in Ref. [11] uses FL in 6G networks for resource
recommendation and scheduling-based propagation analysis.
The traditional FL model is composed of a server and a number of clients. The server
aggregates the local model, broadcasts the global model, and ultimately outputs a highly
available predictive model. The client obtains the global model, updates it with local data,
and finally uploads it to the server. However, this structure requires establishing a mutual
trust mechanism between the client and the server, which will pose a severe threat to the
system if the server is hijacked or malicious. We will discuss the details in Section 3.
2.1. Classification of FL
FL is in a rapid state of development, and various techniques and methods are being
used to enhance its applicability of FL. In order to cope with more complex requirements,
FL with different complex morphologies is proposed [7]. Some research works want to
apply machine learning models to more significant distributed scenarios, requiring FL to
consider communication and data aggregation costs. For example, Ref. [12] focuses on
using FL in mobile edge computing. However, some other works require more secure
participation in training, which requires FL to pay more attention to security and privacy
protection. In ref. [13], the authors focus more on security threats and classify FL in terms
of data feature distribution. Other than that, Ref. [14] also classifies FL with technologies
from different perspectives. Therefore, classifying FL according to different perspectives is
an essential first step in understanding and optimizing FL design.
2.1.1. Centralized/Multi-Center/Decentralized FL
Although FL has been implemented for decentralized training, centralized FL still
requires a central server to complete the accusation of aggregation and broadcasting, which
we call centralized FL [1,15–18]. The single-server design ensures that the model’s rights
are centralized in the hands of the server, which helps to manage the whole training process
and avoid errors. For example, Google’s Gboard for Android keyboard is based on this
architecture. However, a centralized server tends to occur a single point of failure, which
might destroy the whole FL system. To release this security dependency, a decentralized FL
was proposed. It attempts to reduce or even eliminate the server’s control over the global
model. As demonstrated in Ref. [19], authors proposed two asynchronous peer-to-peer
algorithms for the novel setting of decentralized collaborative learning of personalized
models. This approach removes the server directly to achieve complete decentralization.
However, the time and communication cost of this approach is often huge. Multi-center
FL does not require a centralized server, but multiple decentralized edge servers need to
manage model updates. This weakens the impact of servers becoming malicious nodes
on the global model while ensuring the utility of the model. To solve the above problems,
Ref. [20] learns multiple global models from the data, simultaneously derives the best
match between users and centers, and proposes an optimization approach, federated SEM,
to eliminate the multi-center FL problem effectively.
3. Threats in FL
In this section, we will introduce the mainstream attacks confronted in FL. Before dis-
cussing what threats FL faces, we first introduce a new perspective to analyze where these
threats may originate. In the second and third parts, we analyze the possible attacks on FL
from the point of view of security and privacy, respectively. The analysis of the source of
threats and other sections have been partially studied in previous work [13,14,24–26].
3. External attackers: An external attacker aiming at sabotage might hijack the server or
bring it down directly, thus completely disrupting the training. External entities that
eavesdrop on server and client communication channels also threaten clients’ privacy
considerably. Homomorphic encryption and differential privacy can limit its access to
accurate information, and the combination of the trusted execution environment’s FL
can shield it from threats.
4. Collusion: Multiple malicious adversaries can collude to launch a joint attack. In prac-
tice, a conspiracy attack requires only a tiny amount of secrecy to be divulged by an
internal adversary to undermine the availability of most security protocols. For ex-
ample, HE and SMC-based security schemes rely on the absolute security of keys.
They synchronize and upload the colluded malicious parameters to the server for
aggregation and perform iterative attacks to disrupt the performance of the model [29].
Furthermore, dishonest clients and servers can conspire to steal confidential infor-
mation (e.g., private keys), posing a threat to the partially privacy-preserving FL
model [12].
In Table 1, we analyze and compare the impact of different malicious entities on the
model. The security concern is determined by whether the model can converge adequately.
We simply divide the threats confronted by our system from low to high. Since the server
has too much information, it is involved in a high threat level of attacks, while the threat to
the client depends on the number of attacks involved. Privacy is measured by the threat to
the client’s private data. Although a single client has limited information and low threat,
collusion with the server threatens other users’ information. The notable exception is
the collusion attack, which is more complex to analyze because of the collusion between
different entities [30]. However, because collusion attacks usually involve servers, we set
the threat level as high.
clean-label attacks. The former tends to misclassify by injecting desired target labels into
training datasets. The typical dirty-label attack is a label-flipping attack [35]; it reverses the
label of a feature-invariant sample, thus forcing the model to recognize it as another class.
In Figure 1, malicious adversaries generate poisonous training samples by label flipping and
eventually mislead the global model to generate incorrect classifications. Unlike dirty-label
attacks, clean-label attacks correctly classify poisoned labels during training. However, the
classification models will classify it into the wrong class. Clean label attacks are more
insidious than the former, as most resistance methods based on distribution differences
have little impact on them.
$WWDFN0RGHO
3UHGLFW
0HPEHU,QIHUHQFH
$WWDFN
*XLGDQFH
ĂĂ
6KDGRZ 6KDGRZ
0RGHO 0RGHON
ILQHWXQH
7UDLQ
/DEHO)OLS3LQJ
$WWDFN
6KDGRZ 6KDGRZ
3RVLRQHG /DEHO /DEHO/DEHO 6HW 6HW
Unlike data poisoning, model poisoning usually requires sophisticated technical skills
and high computational resources. However, it is also relatively more destructive to
models. Model poisoning aims to make the model misclassify for selective inputs with
high confidence. The work in [36] was carried out by an adversary controlling a small
number of malicious agents to cause the global model to misclassify a set of chosen inputs
with high confidence. It is a common method to misjudge the model by noise. In the work
of [37], noise with different intensities and distributions was used to detect the pixel space
of opposite images.
enough revenue and negatively affect FL training. Typically, fewer free riders are less
harmful, but this is unfair to other clients. Currently, the use of FL models based on
contribution value estimation can alleviate this part of the problem.
3.3. Privacy in FL
Attacks against user privacy undermine the confidentiality of the FL model. Al-
though FL needs to share model parameters instead of sharing local data, there are still
ways to steal user local information.
3.3.3. GANs
GANs have been hugely successful in the image field. It can generate a large amount
of high-quality fake data through gaming methods. Therefore, it is an enhancement from
J. Sens. Actuator Netw. 2023, 12, 13 8 of 18
both an offensive and a defensive point of view. On the one hand, techniques based on
GANs can enhance the ability to poison and infer attacks. The fake data generated by
GANs facilitates poisoning attacks. The work in [52] achieves over 80% accuracy in both the
poisoning and the main task by generating data through GAN. The work in [53] considers
the use of GAN to generate enriched attack data for the shadow model, which in turn
improves the accuracy of member inference attacks to 98%. Due to the nature of GANs,
the system cannot predict all possible threats based on them. Therefore, it is more difficult
to prevent attacks based on GANs. On the other hand, mechanisms combined with GAN
can also improve the robustness of the FL model [54]. The work in [55] shares the client’s
generator with the server to aggregate the client’s shared knowledge and thus improve the
performance of each client’s local network.
Privacy Model In- Using leaked information to reverse model analysis to Map In-
Client Confidentiality [49]
version obtain private information version
Random
The attacker obtains a high-value training
GANs weights Client Confidentiality [52]
model with low-value data.
attack
4. Defense
4.1. Defense Mechanism
Based on our analysis of security and privacy issues in federated learning above,
two main perspectives are worth considering to improve the security of FL: FL needs to
identify and deal with possible security threats at any stage of training. In addition, FL
should ensure mutual trust between all entities, which helps to attract more quality data to
participate in training.
For the first problem, we usually use some proactive defenses. These methods are
expected to detect and eliminate threats as they arise. This is typically cost-effective, but it
is limited in the number of threats it can handle. For the second problem, the key problem
is to keep sensitive information from being transmitted directly. The usual method is to
encrypt sensitive information or use a secure transmission channel. Such approaches tend
to be reactive, and data are not monitored once processed.
clients. By testing the outlier degree of data points, the server can effectively reduce the
damage of poisoned data to the global model, a common means to actively defend against
poisoning attacks. For example, the work in [15] designed an outlier data point detection
mechanism that can effectively eliminate tag reversal and backdoor-based poisoning attacks.
Li et al. [56] use a pre-trained anomaly detection model to test whether users deviate from
the FL training regulations. In addition, by saving incremental updates in the blockchain
distributed ledger, the server can detect and audit the updates of the model [57]. Another
aspect of anomaly detection methods such as BAFFLE [58] on the client side allows the
detection to be decentralized to the client, with the server simply analyzing the results of
the participant’s determination. At the same time, anomaly detection methods based on
participant parameter distributions and energy anomalies can be constructed to cope with
free-rider attacks.
4.1.2. Blockchain
Blockchain is based on a peer-to-peer network. Blockchain ensures secure storage and
data traceability through a combination of chain, tree, and graph structures. In addition,
the blockchain achieves tamper-evident data through the consensus mechanism of proof of
work (POW). Blockchain and FL are complementary to each other. Blockchain is a natural
fit for development alongside FL as an inherently secure distributed system. Combined
with FL, we can make all its data copied, shared, and distributed on multiple servers. As in
Figure 2, FL can build a trusted third party and complete some trusted operations on the
chain, thus reducing the trust anxiety on the server.
The blockchain’s decentralization can weaken the server’s authority, while the dis-
tributed ledger provides secure verification for FL [59]. In addition to its verifiable nature,
the blockchain can enhance FL’s fairness. The transparent and verifiable distribution of
incentives can be improved through smart contracts, thereby achieving that all customers
receive incentives that meet their values [60]. Ref. [61] uses blockchain to facilitate transpar-
ent processes and enforce regulations. The server is independent of blockchain computing,
enabling a trust boundary with the user.
J. Sens. Actuator Netw. 2023, 12, 13 10 of 18
Figure 3. We experimentally compared the performance of different DP-FL schemes on the MNIST
dataset under the same privacy budget.
Here, M and C denote the plaintext space and ciphertext space, respectively, denotes
the operator. Operations on ciphertext can be overloaded as addition and multiplication.
J. Sens. Actuator Netw. 2023, 12, 13 11 of 18
6*;
3 3XOO
$JJUHJDWLRQ
6HJPHQWV
'HFUSW\ 1
H
W
3
Z
,QWUDJURXS R
U
*UDGLHQW
N
4.1.7. Hybrid
On the one hand, in the face of complex security needs, sometimes maintaining data
localization alone often fails to achieve sufficient privacy guarantees. The work in [76]
combined DP and SMC in the database using DP fuzzy secrets and used a secret sharing
mechanism to slice and dice the restructuring of parameters and answers for computational
and output privacy. Inspired by the hybrid methodology, Refs. [76,77] combine it with
FL. In other areas, Ref. [78] introduces a novel strategy that combines differential privacy
methods and homomorphic encryption techniques to achieve the best of both worlds.
Ref. [64] has designed an efficient and secure aggregation scheme and uses distributed DP
for privacy enhancement.
On the other hand, there are inherent limitations to a single-defense approach. For ex-
ample, differential privacy has an inherent loss of accuracy, and HE and SMC perform
poorly in terms of efficiency. In combination with SMC, DP in [77] can achieve a small
increase in noise while guaranteeing a predefined trust rate. This reduces the negative
impact of DP on the utility of the model. However, the underlying security mechanism
of [77] is additive homomorphic encryption and is therefore accompanied by a longer
training time and higher transmission costs. Ref. [79] improved security policy using
function encryption and designed HybridAlpha to achieve shorter time and communica-
tion costs. In Ref. [80], HE and TEE were used jointly to achieve an accurate analysis of
genomic data. This combined solution provided a good compromise regarding efficiency
and computational support for safe statistical analysis.
Table 3. Cont.
Homomorphic
high high low high low high [66]
Encryption
Secure
Multiparty high high medium high low medium [68]
Computing
Trusted
Execution En- medium high high medium low high [73,75]
vironments
Regarding system security protection, AD, TEE, and DS cannot work against internal
malicious nodes. The related means based on perturbation and cryptography hide the in-
termediate variables of the computation with solid theoretical security. Regarding accuracy
impact, DP introduces noise to mask critical information, which affects the final model
convergence accuracy. This loss of accuracy is often unacceptable under LDP. In terms of
efficiency performance, the DP-based FL has the lowest time consumption thanks to its
streamlined algorithm. The intermediate information masking method represented by DP
has good generalization for different types of FL. In contrast, schemes such as AD need to
be designed specifically for different kinds of FL and have certain bureaus.
Homomorphic encryption is usually inefficient in the face of high-dimensional vectors
in FL, which take a lot of time to encrypt and decrypt. AD and DS exhibit weak robustness
to new attacks and need to be dynamically updated on time. The scalability of FL is
reflected in the performance in complex scenarios such as large-scale node distribution and
unexpected user dropouts. The HE and SMC participating nodes are also computational
and heavily burdened with computation and communication. It is worth noting that HE
and SMC are cryptographically provably secure, which is significant for the interpretability
of the construction of TFL. With a large number of level nodes, training efficiency is signifi-
cantly reduced. At the same time, limited local computing resources limit the use of TEE.
However, directly executing the training process in the TEE environment will significantly
decline performance. Take Intel SGX as an example; it only supports CPU operation, which
limits the model’s efficiency (CNN, DNN) and relies on GPU training. At the same time,
when the memory exceeds the limit, it will induce a lot of paging overhead.
5. Future Research
The continuing fire of federal learning research, new forms of attack, and scenario
demands have raised the bar even higher. However, the investigation into TFL is still in its
infancy. In this section, the main existing threats and means of defense in FL are combined
to suggest future directions worthy of attention.
tivity analysis. As can be seen, the vast majority of the work is a metric for the refinement
of specific privacy paradigms, but it does measure the means of the global model.
5.4. Decentralization
A centralized FL can give the server too much power to create a trust crisis. How-
ever, an utterly decentralized server would also be inconvenient to manage and audit. It
is worth thinking about effectively spreading the risk while ensuring the model can be
handled safely. Blockchain can help decentralize FL [60], where the server no longer acts as
the core of auditing and verification but only performs aggregation algorithms and where
the associated auditing and verification can be achieved through smart contracts on the
blockchain. However, Block-FL suffers from both efficiency and expense problems.
6. Conclusions
FL facilitates the free flow of data and offers the possibility of machine learning cross-
domain applications. However, it is necessary to remove user trust anxiety and facilitate
the commercial deployment of TFL. However, research on TFL is still in its infancy. In this
article, we clearly define trustworthy federated learning. By summarizing and analyzing
the security and privacy threats faced by FL, we hope to provide new research perspectives
J. Sens. Actuator Netw. 2023, 12, 13 15 of 18
to community researchers. Finally, we provide some helpful research directions for the
top-level design of TFL. TFL is an enhanced framework designed for market needs, and our
research aims to provide some references for its design.
References
1. McMahan, H.; Moore, E.; Ramage, D.; Hampson, S.; Aguera y Arcas, B. Communication-Efficient Learning of Deep Networks
from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, PMLR, MA,
Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. Volume 54.
2. Yang, Z.; Shi, Y.; Zhou, Y.; Wang, Z.; Yang, K. Trustworthy Federated Learning via Blockchain. IEEE Internet Things J. 2023, 10,
92–109. [CrossRef]
3. Lin, X.; Wu, J.; Li, J.; Zheng, X.; Li, G. Friend-as-learner: Socially-driven trustworthy and efficient wireless federated edge learning.
IEEE Trans. Mob. Comput. 2023, 22, 269–283. [CrossRef]
4. Bugshan, N.; Khalil, I.; Rahman, M.S.; Atiquzzaman, M.; Yi, X.; Badsha, S. Toward Trustworthy and Privacy-Preserving Federated
Deep Learning Service Framework for Industrial Internet of Things. IEEE Trans. Ind. Inform. 2022, 19, 1535–1547. [CrossRef]
5. Zhang, Q.; Ding, Q.; Zhu, J.; Li, D. Blockchain empowered reliable federated learning by worker selection: A trustworthy
reputation evaluation method. In Proceedings of the 2021 IEEE Wireless Communications and Networking Conference Workshops
(WCNCW), Nanjing, China, 29 March 2021; IEEE: Piscataway, NJ, USA, 2021.
6. Yang, Q.; Liu, Y.; Cheng, Y.; Kang, Y.; Chen, T.; Yu, H. Federated learning. Synth. Lect. Artif. Intell. Mach. Learn. 2019, 13, 1–207 .
7. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Roselander, J. Towards federated learning at scale:
System design. In Proceedings of Machine Learning and Systems, Standford, CA, USA, 31 March–2 April 2019; pp. 374–388.
8. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. Acm Trans. Intell. Syst. Technol. 2019,
10, 1–19. [CrossRef]
9. Nishio, T.; Yonetani, R. Client selection for federated learning with heterogeneous resources in mobile edge. In Proceedings of the
ICC 2019-2019 IEEE international conference on communications (ICC), Shanghai, China, 20–24 May 2019; IEEE: Piscataway, NJ,
USA; pp. 1–7.
10. Xu, C.; Qu, Y.; Xiang, Y.; Gao, L. Asynchronous federated learning on heterogeneous devices: A survey. arXiv 2022,
arXiv:2109.04269.
11. Ahmed, S.T.; Kumar, V.V.; Singh, K.K.; Singh, A.; Muthukumaran, V.; Gupta, D. 6G enabled federated learning for secure IoMT
resource recommendation and propagation analysis. Comput. Electr. Eng. 2022, 102, 108210. [CrossRef]
12. Lim, W.Y.B.; Luong, N.C.; Hoang, D.T.; Jiao, Y.; Liang, Y.C.; Yang, Q.; Miao, C. Federated learning in mobile edge networks: A
comprehensive survey. IEEE Commun. Surv. Tutor. 2020, 22, 2031–2063. [CrossRef]
13. Lyu, L.; Yu, H.; Yang, Q. Threats to federated learning. Lect. Notes Comput. Sci. 2020, 12500, 3–16.
14. Mothukuri, V.; Parizi, R.M.; Pouriyeh, S.; Huang, Y.; Dehghantanha, A.; Srivastava, G. A survey on security and privacy of
federated learning. Future Gener. Comput. Syst. 2021, 115, 619–640. [CrossRef]
15. Liu, X.; Li, H.; Xu, G.; Chen, Z.; Huang, X.; Lu, R. Privacy-enhanced federated learning against poisoning adversaries. IEEE Trans.
Inf. Forensics Secur. 2021, 16, 4574–4588. [CrossRef]
16. Yue, S.; Ren, J.; Xin, J.; Zhang, D.; Zhang, Y.; Zhuang, W. Efficient federated meta-learning over multi-access wireless networks.
IEEE J. Sel. Areas Commun. 2022, 40, 1556–1570. [CrossRef]
17. Liu, R.; Cao, Y.; Chen, H.; Guo, R.; Yoshikawa, M. Flame: Differentially private federated learning in the shuffle model. AAAI
Conf. Artif. Intell. 2021, 35, 8688–8696. [CrossRef]
18. Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res.
2021, 5, 1–19. [CrossRef] [PubMed]
19. Vanhaesebrouck, P.; Bellet, A.; Tommasi, M. Decentralized collaborative learning of personalized models over networks. In
Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 509–517.
20. Xie, M.; Long, G.; Shen, T.; Zhou, T.; Wang, X.; Jiang, J.; Zhang, C. Multi-center federated learning. arXiv 2020. arXiv:2005.01026.
21. Li, T.; Sahu, A.K.; Zaheer, M.; Sanjabi, M.; Talwalkar, A.; Smith, V. Federated Optimization in Heterogeneous Networks. Mach.
Learn. Syst. 2020, 2, 429–450.
22. Zhang, Q.; Gu, B.; Deng, C.; Huang, H. Secure bilevel asynchronous vertical federated learning with backward updating. AAAI
Conf. Artif. Intell. 2021, 35, 10896–10904. [CrossRef]
J. Sens. Actuator Netw. 2023, 12, 13 16 of 18
23. Liu, Y.; Kang, Y.; Xing, C.; Chen, T.; Yang, Q. A secure federated transfer learning framework. IEEE Intell. Syst. 2020, 35, 70–82.
[CrossRef]
24. Li, Q.; Wen, Z.; Wu, Z.; Hu, S.; Wang, N.; Li, Y.; He, B. A survey on federated learning systems: Vision, hype and reality for data
privacy and protection. IEEE Trans. Knowl. Data Eng. 2021. [CrossRef]
25. Jere, M.S.; Farnan, T.; Koushanfar, F. A taxonomy of attacks on federated learning. IEEE Secur. Priv. 2020, 19, 20–28. [CrossRef]
26. Lyu, L.; Yu, H.; Ma, X.; Sun, L.; Zhao, J.; Yang, Q.; Yu, P.S. Privacy and robustness in federated learning: Attacks and defenses.
arXiv 2020, arXiv:2012.06337.
27. Girgis, A.; Data, D.; Diggavi, S.; Kairouz, P.; Suresh, A.T. Shuffled model of differential privacy in federated learning. In
Proceedings of the International Conference on Artificial Intelligence and Statistics, Online, 13–15 April 2021; pp. 2521–2529.
28. Bagdasaryan, E.; Veit, A.; Hua, Y.; Shmatikov, V. How to backdoor federated learning. In Proceedings of the International
Conference on Artificial Intelligence and Statistics,Online, 26–28 August 2020; pp. 2938–2948.
29. Taheri, R.; Shojafar, M.; Alazab, M.; Tafazolli, R. FED-IIoT: A robust federated malware detection architecture in industrial IoT.
IEEE Trans. Ind. Inform. 2020, 17, 8442–8452. [CrossRef]
30. Ranjan, P.; Corò, F.; Gupta, A.; Das, S.K. Leveraging Spanning Tree to Detect Colluding Attackers in Federated Learning. In
Proceedings of the IEEE INFOCOM 2022-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS),
New York, NY, USA, 2–5 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–2.
31. Li, Y.; Chen, C.; Liu, N.; Huang, H.; Zheng, Z.; Yan, Q. A blockchain-based decentralized federated learning framework with
committee consensus. IEEE Netw. 2020, 35, 234–241. [CrossRef]
32. Cao, D.; Chang, S.; Lin, Z.; Liu, G.; Sun, D. Understanding distributed poisoning attack in federated learning. In Proceedings of
the 2019 IEEE 25th International Conference on Parallel and Distributed Systems (ICPADS), Tianjin, China, 4–6 December 2019;
IEEE: Piscataway, NJ, USA, 2019; pp. 233–239.
33. Hu, H.; Salcic, Z.; Sun, L.; Dobbie, G.; Yu, P.S.; Zhang, X. Membership inference attacks on machine learning: A survey. ACM
Comput. Surv. 2021, 54, 1–37. [CrossRef]
34. Xiao, X.; Tang, Z.; Li, C.; Xiao, B.; Li, K. SCA: Sybil-based Collusion Attacks of IIoT Data Poisoning in Federated Learning. IEEE
Trans. Ind. Inform. 2022. [CrossRef]
35. Fung, C.; Yoon, C.J.; Beschastnikh, I. Mitigating sybils in federated learning poisoning. arXiv 2018, arXiv:1808.04866.
36. Bhagoji, A.N.; Chakraborty, S.; Mittal, P.; Calo, S. Analyzing federated learning through an adversarial lens. In Proceedings of the
International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 634–643.
37. Tabacof, P.; Valle, E. Exploring the space of adversarial images. In Proceedings of the 2016 International Joint Conference on
Neural Networks (IJCNN), Vancouver, BC, Canada, 14–19 July 2016; IEEE: Pisataway, NJ, USA, 2016; pp. 426–433.
38. Wang, H.; Sreenivasan, K.; Rajput, S.; Vishwakarma, H.; Agarwal, S.; Sohn, J.Y.; Papailiopoulos, D. Attack of the tails: Yes, you
really can backdoor federated learning. Adv. Neural Inf. Process. Syst. 2020, 33, 16070–16084.
39. Sun, Z.; Kairouz, P.; Suresh, A.T.; McMahan, H.B. Can you really backdoor federated learning? arXiv 2019, arXiv:1911.07963.
40. Lin, J.; Du, M.; Liu, J. Free-riders in federated learning: Attacks and defenses. arXiv 2019, arXiv:1911.12560.
41. Fraboni, Y.; Vidal, R.; Lorenzi, M. Free-rider attacks on model aggregation in federated learning. In Proceedings of the International
Conference on Artificial Intelligence and Statistics, Online, 13–15 April 2021; pp. 1846–1854.
42. Nasr, M.; Shokri, R.; Houmansadr, A. Comprehensive privacy analysis of deep learning: Passive and active white-box inference
attacks against centralized and federated learning. In Proceedings of the 2019 IEEE symposium on security and privacy (SP) , San
Francisco, CA, USA, 19–23 May 2019: IEEE: Piscataway, NJ, USA, 2019; pp. 739–753.
43. Choquette-Choo, C.A.; Tramer, F.; Carlini, N.; Papernot, N. Label-only membership inference attacks. In Proceedings of the
International Conference on Machine Learning, Online, 18–24 July 2021; pp. 1964–1974.
44. Gao, J.; Hou, B.; Guo, X.; Liu, Z.; Zhang, Y.; Chen, K.; Li, J. Secure aggregation is insecure: Category inference attack on federated
learning. IEEE Trans. Dependable Secur. Comput. 2021, 20, 147–160. [CrossRef]
45. Luo, X.; Wu, Y.; Xiao, X.; Ooi, B.C. Feature inference attack on model predictions in vertical federated learning. In Proceedings of
the 2021 IEEE 37th International Conference on Data Engineering (ICDE), Chania, Greece, 19–22 April 2021; pp. 181–192.
46. Fu, C.; Zhang, X.; Ji, S.; Chen, J.; Wu, J.; Guo, S.; Wang, T. Label inference attacks against vertical federated learning. In Proceedings
of the 31st USENIX Security Symposium (USENIX Security 22), Boston, MA, USA, 10–12 August 2022.
47. Li, Z.; Zhang, Y. Membership leakage in label-only exposures. In Proceedings of the 2021 ACM SIGSAC Conference on Computer
and Communications Security, Online, Republic of Korea, 15–19 November 2021; ACM: New York, NY, USA, 2021; pp. 880–895.
48. Fredrikson, M.; Jha, S.; Ristenpart, T. Model inversion attacks that exploit confidence information and basic countermeasures. In
Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October
2015; ACM: New York, NY, USA, 2015; pp. 1322–1333.
49. Vepakomma, P.; Singh, A.; Gupta, O.; Raskar, R. NoPeek: Information leakage reduction to share activations in distributed
deep learning. In Proceedings of the 2020 International Conference on Data Mining Workshops (ICDMW), Sorrento, Italy, 17–20
November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 933–942.
50. He, Z.; Zhang, T.; Lee, R.B. Attacking and protecting data privacy in edge–cloud collaborative inference systems. IEEE Internet
Things J. 2020, 8, 9706–9716. [CrossRef]
J. Sens. Actuator Netw. 2023, 12, 13 17 of 18
51. Li, J.; Rakin, A.S.; Chen, X.; He, Z.; Fan, D.; Chakrabarti, C. ResSFL: A Resistance Transfer Framework for Defending Model
Inversion Attack in Split Federated Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern
Recognition, Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA, 2022; pp. 10194–10202.
52. Zhang, J.; Chen, J.; Wu, D.; Chen, B.; Yu, S. Poisoning attack in federated learning using generative adversarial nets. In Proceedings
of the 2019 18th IEEE International Conference On Trust, Security And Privacy In Computing And Communications/13th IEEE
International Conference On Big Data Science And Engineering (TrustCom/BigDataSE), Rotorua, New Zealand, 5–8 August 2019;
IEEE: Piscataway, NJ , USA, 2019; pp. 374–380.
53. Zhang, J.; Zhang, J.; Chen, J.; Yu, S. Gan enhanced membership inference: A passive local attack in federated learning. In
Proceedings of the ICC 2020-2020 IEEE International Conference on Communications (ICC), Dublin, Ireland, 7–11 June 2020;
IEEE: Piscataway, NJ, USA, 2020; pp. 1–6.
54. Ghonima, R. Implementation of GANs Using Federated Learning. In Proceedings of the 2021 Tenth International Conference on
Intelligent Computing and Information Systems (ICICIS), Cairo, Egypt, 5–7 December 2021; IEEE: Piscataway, NJ, USA, 2021;
pp. 142–148.
55. Wu, Y.; Kang, Y.; Luo, J.; He, Y.; Yang, Q. Fedcg: Leverage conditional gan for protecting privacy and maintaining competitive
performance in federated learning. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence,
Vienna, Austria, 23–29 July 2021; pp. 2334–2340.
56. Li, S.; Cheng, Y.; Liu, Y.; Wang, W.; Chen, T. Abnormal client behavior detection in federated learning. arXiv 2019, arXiv:1910.09933.
57. Preuveneers, D.; Rimmer, V.; Tsingenopoulos, I.; Spooren, J.; Joosen, W.; Ilie-Zudor, E. Chained anomaly detection models for
federated learning: An intrusion detection case study. Appl. Sci. 2018, 8, 2663. [CrossRef]
58. Andreina, S.; Marson, G.A.; Möllering, H.; Karame, G. Baffle: Backdoor detection via feedback-based federated learning. In
Proceedings of the 2021 IEEE 41st International Conference on Distributed Computing Systems (ICDCS), Washington, DC, USA,
7–10 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 852–863.
59. Kim, H.; Park, J.; Bennis, M.; Kim, S.L. Blockchained on-device federated learning. IEEE Commun. Lett. 2019, 24, 1279–1283.
[CrossRef]
60. Rückel, T.; Sedlmeir, J.; Hofmann, P. Fairness, integrity, and privacy in a scalable blockchain-based federated learning system.
Comput. Netw. 2022, 202, 108621. [CrossRef]
61. Miao, Y.; Liu, Z.; Li, H.; Choo, K.K.R.; Deng, R.H. Privacy-Preserving Byzantine-Robust Federated Learning via Blockchain
Systems. IEEE Trans. Inf. Forensics Secur. 2022, 17, 2848–2861. [CrossRef]
62. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In
Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October
2016; pp. 308–318.
63. Mironov, I. Rényi differential privacy. In Proceedings of the 2017 IEEE 30th Computer Security Foundations Symposium (CSF),
Santa Barbara, CA, USA, 21–25 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 263–275.
64. Kairouz, P.; Liu, Z.; Steinke, T. The distributed discrete gaussian mechanism for federated learning with secure aggregation. In
Proceedings of the International Conference on Machine Learning, Xiamen, China, 8–11 November 2021; pp. 5201–5212.
65. Paillier, P. Public-key cryptosystems based on composite degree residuosity classes. In Proceedings of the International Con-
ference on the Theory and Applications of Cryptographic Techniques, Prague, Czech Republic, 2–6 May 1999; Springer:
Berlin/Heidelberg, Germany, 1999; pp. 223–238
66. Zhang, C.; Li, S.; Xia, J.; Wang, W.; Yan, F.; Liu, Y. BatchCrypt: Efficient Homomorphic Encryption for Cross-Silo Federated
Learning. In Proceedings of the 2020 USENIX annual technical conference (USENIX ATC 20), Boston, MA, USA, 15–17 July 2020;
pp. 493–506.
67. Chai, D.; Wang, L.; Chen, K.; Yang, Q. Secure federated matrix factorization. IEEE Intell. Syst. 2020, 36, 11–20. [CrossRef]
68. Mugunthan, V.; Polychroniadou, A.; Byrd, D.; Balch, T.H. Smpai: Secure multi-party computation for federated learning. In
Proceedings of the NeurIPS 2019 Workshop on Robust AI in Financial Services, Vancouver, BC, Canada, 9–14 December 2019;
MIT Press: Cambridge, MA, USA, 2019.
69. Li, C.; Li, G.; Varshney, P.K. Communication-efficient federated learning based on compressed sensing. IEEE Internet Things J.
2021, 8, 15531–15541. [CrossRef]
70. Hauck, E.; Loss, J. Efficient and universally composable protocols for oblivious transfer from the CDH assumption. Cryptology
2017, 1011.
71. Wei, Z.; Pei, Q.; Zhang, N.; Liu, X.; Wu, C.; Taherkordi, A. Lightweight Federated Learning for Large-scale IoT Devices with
Privacy Guarantee. IEEE Internet Things J. 2021. [CrossRef]
72. Xu, G.; Li, H.; Liu, S.; Yang, K.; Lin, X. Verifynet: Secure and verifiable federated learning. IEEE Trans. Inf. Forensics Secur. 2019, 15,
911–926. [CrossRef]
73. Mondal, A.; More, Y.; Rooparaghunath, R.H.; Gupta, D. Poster: FLATEE: Federated Learning Across Trusted Execution
Environments. In Proceedings of the 2021 IEEE European Symposium on Security and Privacy (EuroS&P), Vienna, Austria, 6–10
September 2021; pp. 707–709.
74. Mondal, A.; More, Y.; Rooparaghunath, R.H.; Gupta, D. Flatee: Federated Learning Across Trusted Execution Environments.
arXiv 2021, arXiv:2111.06867.
J. Sens. Actuator Netw. 2023, 12, 13 18 of 18
75. Zhang, Y.; Wang, Z.; Cao, J.; Hou, R.; Meng, D. ShuffleFL: Gradient-preserving federated learning using trusted execution
environment. In Proceedings of the 18th ACM International Conference on Computing Frontiers, Online, 11–13 May 2021; ACM:
New York, NY, USA, 2021; pp. 161–168.
76. Pettai, M.; Laud, P. Combining differential privacy and secure multiparty computation. In Proceedings of the 31st Annual
Computer Security Applications Conference, Los Angeles, CA, USA, 7–11 December 2015; ACM: New York, NY, USA, 2015;
pp. 421–430.
77. Truex, S.; Baracaldo, N.; Anwar, A.; Steinke, T.; Ludwig, H.; Zhang, R.; Zhou, Y. A hybrid approach to privacy-preserving
federated learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, London, UK, 15 November
2019; ACM: New York, NY, USA, 2019; pp. 1–11.
78. Kim, M.; Lee, J.; Ohno-Machado, L.; Jiang, X. Secure and differentially private logistic regression for horizontally distributed data.
IEEE Trans. Inf. Forensics Secur. 2019, 15, 695–710. [CrossRef]
79. Xu, R.; Baracaldo, N.; Zhou, Y.; Anwar, A.; Ludwig, H. Hybridalpha: An efficient approach for privacy-preserving federated
learning. In Proceedings of the 12th ACM Workshop on Artificial Intelligence and Security, London, UK, 15 November 2019;
ACM: New York, NY, USA, 2019; pp. 13–23.
80. Sadat, M.N.; Al Aziz, M.M.; Mohammed, N.; Chen, F.; Jiang, X.; Wang, S. Safety: Secure gwas in federated environment through a
hybrid solution. IEEE/ACM Trans. Comput. Biol. Bioinform. 2018, 16, 93–102. [CrossRef] [PubMed]
81. Shen, Y.; Sanghavi, S. Learning with bad training data via iterative trimmed loss minimization. In Proceedings of the International
Conference on Machine Learning , Long Beach, CA, USA, 9–15 June 2019; pp. 5739–5748.
82. Wagner, I.; Eckhoff, D. Technical privacy metrics: A systematic survey. ACM Comput. Surv. 2018, 51, 1–38. [CrossRef]
83. Majeed, I.A.; Kaushik, S.; Bardhan, A.; Tadi, V.S.K.; Min, H.K.; Kumaraguru, K.; Muni, R.D. Comparative assessment of federated
and centralized machine learning. arXiv 2022, arXiv:2202.01529.
84. Koh, P.W.; Liang, P. Understanding black-box predictions via influence functions. In Proceeding of the International Conference
on Machine Learning, Sydney, Australia, 7–9 August 2017; pp. 1885–1894.
85. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the
22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August
2016; ACM: New York, NY, USA, 2016; pp. 1135–1144.
86. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat.
Mach. Intell. 2019, 1, 206–215. [CrossRef]
87. Li, X.; Huang, K.; Yang, W.; Wang, S.; Zhang, Z. On the convergence of fedavg on non-iid data. arXiv 2019, arXiv:1907.02189.
88. Zhang, X.; Chen, X.; Hong, M.; Wu, S.; Yi, J. Understanding Clipping for Federated Learning: Convergence and Client-Level
Differential Privacy. In Proceedings of the International Conference on Machine Learning, PMLR, MA, Baltimore, MD, USA,
17–23 July 2022; pp. 26048–26067.
89. Kim, M.; Günlü, O.; Schaefer, R.F. Federated learning with local differential privacy: Trade-offs between privacy, utility, and
communication. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics Speech and Signal
Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2011; IEEE: Piscataway, NJ, USA, 2021; pp. 2650–2654.
90. Balle, B.; Bell, J.; Gascón, A.; Nissim, K. The privacy blanket of the shuffle model. In Proceedings of the Annual International
Cryptology Conference, Santa Barbara, CA, USA, 18–22 August 2019; Springer: Cham, Switzerland, 2019; pp. 638–667.
91. Li, Z.; Kovalev, D.; Qian, X.; Richtárik, P. Acceleration for compressed gradient descent in distributed and federated optimization.
In Proceedings of the 37th International Conference on Machine Learning, Online, 13–18 July 2020; pp. 5895–5904.
92. Cheng, A.; Wang, P.; Zhang, X.S.; Cheng, J. Differentially Private Federated Learning with Local Regularization and Sparsification.
In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020;
IEEE: Piscataway, NJ, USA, 2022; pp. 10122–10131.
93. Yazdinejad, A.; Parizi, R.M.; Dehghantanha, A.; Choo, K.K.R. Blockchain-enabled authentication handover with efficient privacy
protection in SDN-based 5G networks. IEEE Trans. Netw. Sci. Eng. 2019, 8, 1120–1132. [CrossRef]
94. Li, Y.; Tao, X.; Zhang, X.; Liu, J.; Xu, J. Privacy-preserved federated learning for autonomous driving. IEEE Trans. Intell. Transp.
Syst. 2021, 23, 8423–8434. [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.