Role of Federated Learning in Edge Computing A Sur
Role of Federated Learning in Edge Computing A Sur
net/publication/374654594
CITATION READS
1 555
5 authors, including:
SEE PROFILE
All content following this page was uploaded by Sambit Kumar Mishra on 28 March 2024.
ABSTRACT
This paper explores various approaches to enhance federated learning (FL) through the utilization of edge
computing. Three techniques, namely Edge-Fed, hybrid federated learning at edge devices, and cluster federated learning,
are investigated. The Edge-Fed approach implements the computational and communication challenges faced by mobile
devices in FL by offloading calculations to edge servers. It introduces a network architecture comprising a central cloud
server, an edge server, and IoT devices, enabling local aggregations and reducing global communication frequency. Edge-
Fed offers benefits such as reduced computational costs, faster training, and decreased bandwidth requirements. Hybrid
federated learning at edge devices aims to optimize FL in multi-access edge computing (MAEC) systems. Cluster
federated learning introduces a cluster-based hierarchical aggregation system to enhance FL performance. The paper
explores the applications of these techniques in various domains, including smart cities, vehicular networks, healthcare,
cybersecurity, natural language processing, autonomous vehicles and smart homes. The combination of edge computing
(EC) and federated learning (FL) is a promising technique gaining popularity across many applications. EC brings cloud
computing services closer to data sources, further enhancing FL. The integration of FL and EC offers potential benefits
in terms of collaborative learning.
Keywords: cloud computing; edge computing; federated learning; hybrid federated learning; cluster federated learning;
asynchronous federated learning; multi-tasking federated learning (MTFL); multi access edge computing (MAEC);
vehicular edge networks (VEN); mobile edge computing (MEC)
ARTICLE INFO
1. Introduction
Received: 7 May 2023
Accepted: 31 July 2023 A modern form of machine learning that deals with the issues of
Available online: 11 October 2023 data privacy and decentralization is federated learning. Traditional
COPYRIGHT machine learning environments frequently collect and store data in
Copyright © 2023 by author(s). centralized servers, causing issues with security and privacy. A
Journal of Autonomous Intelligence is solution is provided by federated learning, which enables machine
published by Frontier Scientific Publishing.
This work is licensed under the Creative learning models to be trained directly on distributed edge devices such
Commons Attribution-NonCommercial 4.0 as mobile phones, tablets, or Internet of Things (IoT) devices without
International License (CC BY-NC 4.0).
https://creativecommons.org/licenses/by- the need for raw data to leave the devices. Federated learning greatly
nc/4.0/ improves privacy protection by keeping data local and decentralized.
The models are trained locally on the devices utilizing the relevant data
instead of transferring the data to a central server for analysis. Only
model updates or combined data are then shared with a central server.
By ensuring that critical data stays on the devices, this method reduces
the possibility of hacking or unauthorized access. Applications that
deal with sensitive or personal data, like those in healthcare, banking,
or personalized recommendations, have a lot of commitment to
federated learning. While protecting data privacy and control, it
enables organizations to take use of the collective understanding of
1
distributed devices. Developments in many fields are made possible through federated learning, which uses
the power of machine learning while protecting privacy[1]. Coming to edge computing, edge computing
emerged as an evolutionary paradigm that deals with the limitations of traditional cloud-centric architectures
in the context of rapidly increasing data volumes and real-time applications. Rather than depending entirely
on centralized cloud servers, edge computing brings computational resources and data storage closer to the site
of data generation, which is often at the edge of the network. Edge computing offers better processing, less
latency, and better reaction times by placing edge devices, such as edge servers, gateways or IoT devices, at
the network edge. Edge computing has a major beneficial effect on real-time applications like self-driving cars,
virtual reality, augmented reality, and industrial IoT because it reduces the time and bandwidth needed to
transport data from a location to a centralized cloud. By distributing computational tasks across edge devices
and reducing the load on centralized cloud servers, edge computing also provides scalability benefits.
Organizations without their own data centers can use cloud resources efficiently and cheaply due to this
decentralized architecture[1,2].
We observed from many studies that, as a result of cloud computing the way we live, work, and learn has
greatly changed but not in all cases, for example consider cisco internet business solutions group, once said
that by the year 2020 the usage of 50 billion things that are connected to the internet will be formed but their
assumption had become false and yet present the number of IoT devices connected to the internet are 14.4
billion because certain IoT applications need very quick responses, some of which may contain private data,
and others of which may generate a lot of data, placing a tremendous burden on networks[3]. Therefore, cloud
computing is not effective enough to support these applications for this kind of network. We envision that the
edge of the network is transitioning from a data consumer to a data producer as a result of the pull from IoT
and the push from cloud services. Here, the data that the cloud is consuming is very high such that the
processing of the data is also becoming a huge task for the cloud. In this type of situation, the capacity of the
cloud will not be suitable for heavy computations. Take Amazon as an example, in the year 2022, the company
has an increasing number of job applicants, but it is difficult to accommodate them all. As a result, the
employees have been let go. The same thing is happening in the cloud environment, where the number of
devices is growing, and data processing is getting more and more challenging.
We researched the benefits of edge computing over cloud computing as well as several edge computing
applications on Internet of Things (IoT) devices. We also discussed Edge-Fed and Fed-Avg, the advantages of
Edge-Fed over Fed-Avg and Edge-Fed’s future possibilities. We worked on cluster federated learning,
dynamic federated learning[4], multitasking in federated learning[5] and hybrid federated learning at edge
devices. Multi-tasking federated learning for predicting traffic. Asynchronous federated learning[6],
collaborative federated learning at IoT devices, collaborative federated learning in healthcare for COVID-19
diagnosis, collaborative data sharing in vehicle systems, and wireless communications performance for
collaborative federated learning[7]. We have collaborated on a table that covers things that are taken into
consideration from various federated learning in edge computing sectors, such as energy consumption
management in federated learning in edge systems and how caching and offloading take place in federated
learning in mobile edge computing[8].
The rest of the paper is organized as follows. We presented the different applications of federated learning.
The usage of federated learning over edge computing in different areas. In section 2, we presented Edge-Fed
and its applications. In sections 3 and 4, we presented the hybrid FL, clustered FL, and dynamic FL. In section
5, the asynchronous FL is presented. In section 6, federated SGD is discussed. In section 7, multi-tasking FL
is discussed. In section 8, collaborative FL over different applications in edge computing is presented. A few
case studies are also discussed for all sections mentioned above. In section 9, energy consumption in FL is
2
discussed. In section 10, caching and offloading in FL over mobile edge computing are discussed. In section
11, FL in wireless communication is discussed.
There are different types of techniques in federated learning. As edge computing is an emerging
technology, our main theme of the research is to relate usage of federated learning in edge computing. So that
the edge technology can be more efficient in different applications. Data privacy and security has been
improved with federated learning, because federated learning enables training of the models locally without
sharing raw data to the central server. Edge devices often operate in dynamic and unreliable network conditions.
Federated learning can handle intermittent connectivity and device failures by allowing training to continue
on available devices.
We used below mentioned techniques of federated learning because these techniques are widely used in
different applications and these techniques will be helpful for enhanced data privacy, reduced communication
overhead, lower latency, improved scalability, resilience, personalization and energy efficiency. These
improvements make federated learning an attractive approach for deploying machine learning models in edge
computing scenarios.
The techniques we included in this paper are Edge-Fed, cluster-dynamic cluster FL, hybrid FL,
asynchronous FL, federated SGD and multi-tasking in FL, collaborative federated learning over different
applications of edge computing.
2. Edge-Fed
Actually, mobile devices have to perform lots of calculations if they use the federated averaging (Fed-
Avg) algorithm. Federated averaging (Fed-Avg) is an algorithm that is efficient in communication for
distributed training with a larger number of clients. In Fed-Avg, a central server is used to communicate
between the clients, and clients keep their data locally for privacy protection, but in Fed-Avg the distance
between the central server and the clients will be more where that much bandwidth will take time[2].
To solve this problem, we have Edge-Fed where, Edge-Fed is a network with the central cloud server,
edge server, and IoT devices, where the central cloud server all the global aggregations will take place, in the
edge server, it takes some of the workloads from the IoT devices and finally, in IoT devices, local aggregations
happen. In this section, the efficient usage of federated learning in edge computing is elaborated, which is also
shown in Figure 1.
Between the edge server and the central server, bandwidth is less than the bandwidth between the clients
and the edge server. Edge-Fed can decrease the needed global communication frequency to reach a satisfying
accuracy. Consequently, the global communication cost can be reduced compared with Fed-Avg. Edge-Fed
has advantages in different bandwidth scenarios from mobile clients to the edge server. By offloading part of
the calculations, the computational cost of the mobile devices and the global communication expense can be
simultaneously reduced as compared to Fed-Avg. For optimizing the federated learning based on edge
computing, we can overcome the large computational cost in edge devices while performing Fed-Avg.
Some experiments were done (a division of the process of local updates to be completed by both mobile
devices and the edge server and in between the edge servers and the central server, a global aggregation process
was conducted). Those results show that the total computational and communication cost of edge devices are
simultaneously reduced than Fed-Avg. If you want to do many calculations, then it may take more time. To
reduce this type of time complexity the researchers found edge computing which collaborated with federated
learning and formed edge federated learning. Each device will process autonomously with the help of Edge-
Fed. The edge server now collects all outputs from mobile devices in order to increase learning effectiveness
3
and decrease global communication frequency. The computation costs and computations from mobile clients
to the edge server will decrease as a result of this Edge-Fed.
Some of the primary benefits of Edge-Fed are as follows: the local updates will happen in both IoT devices
and edge servers. Where IoT devices can focus more on low layers and edges, services can do more
computational tasks with the required resources. IoT devices and edge servers make up the two categories of
local updates for this model. While services can do more computational activities with the necessary resources,
IoT devices may concentrate more on low layers and edges. So that the training needed for mobile devices will
be low and faster. The bandwidth of clients on the edge server is very high so, by using Edge-Fed we can
reduce bandwidth. By Edge-Fed we can also apply some guidelines in which the model can run accurately
with less resource cost.
IoT helps the present world in many ways. IoT stands for the Internet of Things, which is the networking
of physical things equipped with sensors, software and other technologies for establishing connections and
exchanging data with other devices through the internet. These days, IoT devices are employed more in public
safety equipment and by farmers to monitor their crops and do surveys, among other uses. Low latency and
power consumption are required for IoT devices to carry out duties like monitoring and uploading sensor data,
among others[3].
Generally, IoT devices are weak due to security issues. For IoT, device authentication is required to ensure
that the connected devices on the IoT are trusted. For that unique identifier is needed. In order to optimize
mobile edge computing and communication, Deep reinforcement learning (DRL) and federated learning (FL)
frameworks were combined with mobile edge systems. Energy consumption and execution delays are some
advantages of this partnership. Only the local information is required for this, the channel state information
can be known afterward.
We came across some of the assumptions as follows: IoT devices are typically thought of as being
extremely powerful gadgets with the capacity to train their own DRL agents on their own. IoT devices may
not be as powerful in the near future, they may only be able to compute using minimal neural networks.
Researchers were inspired to develop a new type of computing known as edge computing, which aids in
processing data at the network’s edges, as a result of the Internet of Things (IoT) explosive growth and the
enormous popularity of its rich cloud services. We will examine a variety of case studies in this, ranging from
the cloud to smart cities and smart homes, as well as collaborative edge to broaden the idea of edge computing.
Some of the case studies explain cloud offloading, smart city, and video analytics where they have used
edge computing and federated learning for better results.
4
1) Akamai: Akamai is a leader in content delivery networks and edge computing. Akamai has a network of
servers deployed at the edge of the internet, providing advanced services like streaming video and web
acceleration. Akamai’s edge computing platform provides the scalability, security, and performance
necessary to serve the world’s most heavily trafficked websites.
2) Fastly: Fastly is an edge computing platform that enables organizations to build, deploy, and scale
applications at the edge of the internet. Fastly’s edge network provides customers with faster page load
times, lower latency, and improved reliability. Fastly’s edge computing platform is powered by its
proprietary varnish cache technology, which allows customers to quickly deliver content and applications
without relying on the traditional data center model.
5
hybrid federated learning to address its needs. They used a private cloud infrastructure to deploy a
federated learning system. The federated model was trained using customer data from two of the
company’s stores. The first store was used as the main training set and the second store was used as a test
set. The model was trained using a federated learning algorithm and the data was anonymized before
being sent to the cloud. This enabled the company to keep customer data secure and private. The company
was able to successfully train its model and obtain better results than using traditional centralized training.
This helped them increase customer satisfaction and profits.
● A large financial institution was looking for a way to securely share customer data with its partners
without compromising privacy. The company wanted to use machine learning to gain insights from the
data but wanted to ensure that the data remained private and secure. The company decided to use hybrid
federated learning to address its needs. They used a private cloud infrastructure to deploy a federated
learning system. The federated model was trained using customer data from the financial institution and
its partners. The model was trained using a federated learning algorithm and the data was anonymized
before being sent to the cloud. This enabled the company to keep customer data secure and private while
still obtaining insights from the data. The company was able to successfully train their model and was
able to gain insights from the data without compromising privacy. This helped them increase profits and
customer satisfaction.
6
3.1.3. Natural language processing (NLP)
Scenario:
NLP models often require large amounts of text data for training, which may come from various sources,
such as social media, news articles and private documents. In situations where the data sources cannot be
centrally accessed due to privacy concerns, hybrid federated learning can be used to collaboratively train NLP
models across multiple data owners while maintaining data privacy.
Details:
Data owners, such as different organizations or individuals, act as local clients and train NLP models
using their respective text datasets. By combining knowledge from diverse data sources, the global NLP model
becomes more versatile and accurate. Importantly, the raw text data remains localized, ensuring data privacy
and confidentiality.
Algorithms Accuracy (while training CNN Accuracy (while training CNN Observations
performed model over CIFER-10 dataset) model over MNIST dataset)
Cluster-based FL 35% 74.2% Effectively deals with data
(CFL) imbalance.
Dynamic CFL 51% 95.9% Handles the node failure and
(DCFL) maintains the training process well.
Some of the case studies that include cluster federated learning for better results are as follows:
1. Cluster federated learning on electronic health records: a case study of anemia diagnosis.
2. Cluster federated learning on mobile applications: a case study of predicting user engagement.
7
healthcare institutions. Devices within each cluster, such as hospital servers or patient monitoring devices, can
collaborate to train models specific to their cluster while adhering to data privacy regulations. This enables
knowledge sharing, disease detection, treatment recommendation systems, and improved healthcare outcomes
across different clusters.
4.1.3. Recommendation systems
Recommendation systems require large amounts of data to be trained on. Cluster federated learning can
be used to train recommendation systems on multiple datasets without having to share the underlying data.
8
1) Synchronous scheme: using this approach, data is gathered from all local servers and transferred to the
training data’s global servers. The parameter server will then divide the work into smaller tasks and
deliver information about each edge node’s updated nodes for the specified amount of time. However,
there will be drawbacks such as issues with bandwidth and training time.
2) Asynchronous scheme: with this approach, data is gathered from a few local servers and transferred to
the training data’s worldwide servers. The parameter server will then divide the work into smaller tasks
and deliver information about each edge node’s updated nodes for the specified amount of time. We also
require a lot of time and bandwidth for this.
Communication-efficient asynchronous federated learning (CE-AFL) approach is employed to get around
this. This approach collects data from a specific number of edge nodes’ local servers and sends it to the training
data’s global servers. The work of a specific number of edge nodes will be combined by the parameter server,
which will then update the local server in the order of arrival time. In edge computing with limited resources,
federated learning enables the training of global models over over-dispersed datasets. The ideal outcome of
training data from various users and devices is enhanced model representation and generalization[6].
For optimizing this federated learning, we use two solutions like gradient descent, stochastic gradient
descent.
The case study of asynchronous federated learning for mobile applications is described below:
● A corporation that has to access client data from various sources, such as mobile applications, online
browsers and other connected devices, may be the subject of a federated learning case study. The
corporation wants to use the information gathered from various sources to create a model that can
precisely predict client behavior. The business would need to create an asynchronous federated learning
system in order to achieve this goal. Multiple models of various sources of client data would be used in
this system. The findings would be synced among the models after each model had been trained on a
portion of the data. By doing so, the model would be able to use data from different sources while learning
patterns that might be particular to each data source. The models can be used to forecast customer
behavior once they have been trained. After that, the business can utilize this data to decide how to
improve customer service. This might entail adjustments to goods or services, marketing plans or
customer service techniques. The organization can acquire insights from several sources of consumer data
by using asynchronous federated learning without having to move the data between sources. This keeps
customers’ personal information private while still giving the business useful information.
9
5.1.3. Real-time adaptation and personalization
Asynchronous federated learning enables real-time adaptation and personalization of models. Devices
can update their local models based on their individual data and preferences, without waiting for
synchronization with other devices. This enables applications like personalized news recommendations,
adaptive chatbots and real-time user behavior modeling, where timely updates and personalized experiences
are critical.
6. Federated SGD
These days, the corporate community is paying close attention to machine learning, which the company
will handle the data. The data is gathered from many sources, and ML applies operations to it. However, the
problem is that the collected data should be moved to a single location where the ML can execute its duty as
stated above, but due to security concerns, which prevented anybody from authorizing the move, the
company’s reputation and their clients’ data were damaged. The solution is PPML (privacy preserving machine
learning), but because ML algorithms in telecommunications are data-hungry, we can’t take the data into a
single spot due to a number of problems.
Due to the aforementioned factors, the FL framework was applied in this communication. Tensor flow
and pytorch were used to create this framework.
The optical networks are becoming their own distinct entities and ML is probably going to play a different
role in this. However, there are a lot of difficulties to be aware of throughout this era of change. The
inaccessibility of the original data that was collected makes it difficult to create ML-based solutions. While
the FL framework fixes the ML issues in optical networks.
Federated SGD (stochastic gradient descent) is a distributed machine learning technique that allows
multiple devices to collaborate in training a machine learning model. This technique enables devices that are
not connected to a central server to train a model without needing to share their local data. Federated SGD can
be used to train models on data that is distributed across multiple devices, such as smartphones, IoT devices,
and edge devices. This allows for distributed training in a secure manner and can be used to train models that
would otherwise require a central server.
Some of the case studies that include for better results are as follows:
● The first case study concerns a study carried out in the UK that employed federated SGD to lower the
price of optical network planning. In order to learn a model of the network topology and evaluate it to
improve the arrangement of the optical links, the project employed federated SGD. This decreased the
cost of network setup and maintenance and allowed the team to plan the network quickly and precisely.
● The second case study focuses on a study that was carried out in the US and employed federated SGD to
lower optical network’s power requirements. In order to reduce power consumption, the project optimized
the location of optical links and components using federated SGD, which was used to learn a model of
the optical network. As a result, the team was able to considerably lower the network’s overall power
consumption. These case examples demonstrate that federated SGD can be used to plan effectively and
precisely.
10
institutions collaborate to train models on patient data while preserving privacy. Each institution performs local
updates using stochastic gradient descent on its own patient data, contributing to the collective learning process.
Fed-SGD can be used for disease prediction, treatment recommendation, clinical decision support, and medical
research while adhering to privacy regulations.
6.1.2. Finance
Fed-SGD can be applied to federated financial systems to train models on distributed financial data while
respecting privacy and security. Each financial institution performs local updates using stochastic gradient
descent on its transactional data, allowing for collaborative learning without sharing sensitive customer
information. Fed-SGD can be used for fraud detection, risk assessment, personalized financial services and
regulatory compliance in a privacy-preserving manner.
6.1.3. Distributed machine learning platforms
Fed-SGD can be integrated into distributed machine learning platforms that span multiple devices or
cloud servers. Each computing node performs local stochastic gradient descent updates, and the model updates
are aggregated to create a global model. Fed-SGD enables distributed training, efficient utilization of
computing resources and scalability for large-scale machine learning applications.
11
keeping data local, the privacy of users’ activities and preferences was preserved. The shared model
learned from the distributed data, enabling it to make accurate predictions about activities, optimize
energy consumption, and identify anomalous events within the smart home ecosystem.
Rounds of training are conducted here until a termination condition is satisfied. The server, which is seen
in the above image, starts these rounds.
The processes involved in using the MTFL algorithm in edge computing are listed below.
The server selects a group of clients to participate in the round from its whole database and sends them a
task request. The clients now get the request from the server and send back an accepted message in response
based on the user preferences. This suggests that a user may set up their gadget such that it only participates
in the FL when it is fully charged and linked to a Wi-Fi network. Clients who accept the request then signal
the server with an accept message. The server now sends the global model to its receptive clients. The clients
modify the local copy of the global model with their own fixes. We use patches from batch-normalization
layers in this. The clients now carry out local training before conserving their patches for the following round.
A subset of clients that agreed to participate in the rounds submit their non-private model and optimizer settings
to the server after waiting a predetermined amount of time. The server then saves the aggregate to create a
single global model that is saved on the server after sorting all models by average. After that, the server may
start a fresh round.
12
7.1.1. Traffic prediction by multi-task federated learning
In this, we did the traffic system prediction of the traffic jam and the speed detection of the vehicle for
the upgrade of the intelligence transportation systems. Because it helps the citizens when choosing their raid
and both to improve fuel economy and to reduce air pollution. These will benefit the populace and are essential
to cutting-edge traffic management systems[5]. An inductive loop may monitor travel speed by measuring how
inductance varies over time in the corresponding traffic lights and Internet of Things devices and this
information can be utilized to estimate traffic speed. They’ll utilize a lot of IoT devices to gather an unrivaled
amount of traffic data from the actual world. This enormous amount of data may lead to an increase in studies
in this area. They can be taken into an entity into two categories: the ARIMA model, and the Kalman filtering
models.
When dealing with traffic simply in regular changes, for instance during rush hours or peak hours in a
city, we may provide reports that are worthwhile. The value will differ from the real values during peak hours
due to the unusual nature of the traffic and the parametric approach’s projections. We may fit the data and do
ML-like non-parametric techniques by gathering the data and performing the above-mentioned mathematical
procedures. Using data gathered by the signal and its arriving and departing traffic signals, a long short-term
memory (LSTM), recurrent neural network (RNN) can anticipate the speed and filling of a traffic jam. To
capture the latent traffic evolution patterns inside the underlying traffic network, we shall employ the
convolution neural network (CNN). This project seeks to use cutting-edge ML, where we may utilize a single
task learning approach, sometimes referred to as the (STL). However, in actuality, there will be a variety of
elements at play, including changes in the weather as well as several events like camps and protests. A
significant responsibility or load for STL is laying the new roads and modifying the existing ones. We can
presume that the local data or traffic signal may have a memory limit where there will be issues with bandwidth
and storage. To address this, they provide multi-task federated learning (FL), which will improve the traffic
conditions as previously described. We first employ the acquired traffic data’s field values, which are impacted
by their immediate neighbors, and build a divisive hierarchical clustering (DIANA), which separates the traffic
data at each signal into a collection of clusters. The federated learning will then do or train them at each local
data set without doing so on the global or central cluster for each data cluster scattered across signals. In this
situation, the proposed multi-task FL architecture can safeguard privacy and lower transmission costs because
no local data is exchanged among stations. On the map where the plan was carried out, we applied the multi-
horizon speed prediction from the multi-task federated learning framework. The dependency graph will be
used to create them in this road map and the modified A* method will be used to identify the cheapest and
fastest route possible. To provide the most accurate traffic predictions in a variety of anomalous traffic
scenarios, we additionally take into account the traffic and route map.
13
browsing history to identify similarities between them and other users. The data is then used to group
users into cohorts, or groups of users, who share similar interests and behaviors. This allows Google to
target ads more accurately and efficiently.
14
Vehicular edge layer (VEL): in the VEL, vehicle edge devices are directly connected to neighboring
MAECS (multi access edge computing servers) over wireless cellular networks, and they produce data traffic
from different user defined VEN services and applications.
Advantages of using federated learning for MAEC-empowered IoV:
Reduction in network bandwidth. Massive VS data is not transmitted to the data center for training since
federated learning can decentralize to learn a deep model at different MAECSs. The use of network bandwidth,
energy consumption, and data transfer could all be significantly decreased via federated learning. Privacy
protection. Users’ sensitive information is shielded from the risk of hacking because less data must be sent to
the data center and their privacy is somewhat secured by low latency. Each local participant (edge server or
device) in federated learning takes decisions in real time, allowing the deep models to be consistently trained
and updated while lowering transmission latency.
The results for the approach are as follows, the convergence of the proposed algorithm is faster than the
centralized scheme. The latency of the proposed algorithm and centralized scheme are similar which means
both methods give optimal results.
8.1.3. Wireless communications for collaborative federated learning
In order to use FL over IoT networks in practice, edge devices must repeatedly send their trained ML
models to a central controller over wireless links. Only some devices can use FL due to constrained wireless
resources, such as those found in an IoT.
The learning process is hampered by wireless channel faults and delays that are transmitted from IoT
devices to a central controller (such as a base station). In order to enhance FL performance, it is required to
take into account wireless network optimization[10].
Performance of collaborative federated learning over wireless networks:
Loss function: CFL training’s objective is to locate an ML model that depends on the local FL models of
all involved devices, which minimizes the loss function. These models encounter transmission delays and
mistakes when sent over wireless networks, which might have an adverse effect on the loss function during
training. Due to limited energy only a few data samples were involved, and the loss function increased.
Convergence time: the time it takes for each device to train its local FL model plus the number of iterations
needed for FL convergence makes up the CFL convergence time.
For each CFL iteration, there are a predetermined amount of local FL model changes.
Energy consumption: the amount of energy used by CFL depends on a number of factors, including the
size of the FL model data, the distance between the BS and the devices, the needed convergence time, and the
goal loss function value.
Reliability: the wireless channel conditions affect CFL’s reliability. Each device’s transmit power
improves as a result, which reduces the number of inaccurate local FL models and raises CFL reliability.
15
problem is split into multiple more manageable problems due to the relationship between the decision
components. The strategy for controlling computing resources is then created by ensuring the energy budget
of mobile users. The phrasing of the dataset offloading and uplink resource management problem as a GNEP
(generalized nash equilibrium problem) also leads to the conclusion that a GNE exists. The dataset offloading
and uplink bandwidth allocation issues are resolved to decrease overall time consumption. To accomplish this,
the energy-aware resource management algorithm is recommended. The proposed MAEC-enabled FL model’s
overall time consumption is competitively lower than that of the traditional FL technique, according to
extensive simulations that employ the recommended resource management algorithm. The performance
parameter energy consumption is considered by many researchers.
Due to the limited energy and computational capabilities of mobile devices, the efficiency of model
training to meet the aim of local energy reduction is always in jeopardy. Multi-access edge computing
(MAEC)-enabled FL eliminates the trade-off between the model performance and the energy use of mobile
devices by allowing users to send a portion of their local dataset to an edge server for the model training. Due
to the edge server’s enormous processing capability, the amount of time needed for model training is
minuscule[11].
Table 2. Observations of different factors between computing resources and offloaded data.
Increase in Loss Time consumption Energy consumption
Computing resources allocation Decreases Decreases Increases
Offloaded data set size Decreases Increases Decreases
16
The results are as follows, the proposed MAEC-enabled FL performs better than the traditional FL due
to the offloading and collaborative training of the local datasets at the edge server. The MAEC-enabled FL can
function with less time consumption than the traditional FL thanks to the proposed resource management
technique. Although there will ultimately be an energy cap for mobile users, the energy use of cell-center users
fluctuates more than that of cell-edge users.
17
Federated learning (FL): federated learning is a machine learning technique that trains an algorithm using
a number of distributed edge devices or servers that each maintain their own local data samples without sharing
them. This strategy is distinct from more traditional decentralized approaches, which usually assume that local
data samples be distributed equally and traditional centralized machine learning techniques, where all local
datasets are uploaded to a single server.
Environment (Env): there are two types of environments in which we conclude they are known as
● Heterogeneous environment (het): using equipment and operating system software from several suppliers.
Computers, operating systems, and databases from many suppliers are often used by businesses.
● Homogenous environment (hom): using software and hardware from the same supplier.
Throughput (TP): throughput is the amount of data that actually moves over a certain period of time. It
may also be described as the maximum quantity of traffic that a website or application can manage.
Scalability (S): scalability is the capacity of a system to modify its cost and performance in response to
changes in application and system processing demands.
Computational power (CP): computing power is a computer’s capacity to do a task quickly and accurately.
Hardware (HWR): data on hardware capacity reveals the communication and computing power of various
devices.
Bandwidth (BND): to ensure minimum transmission delay for training traffic, edge computing introduces
bandwidth slicing to allow federated learning.
Communication cost (CC): the communication cost of a task is the size of the input to the task and can
also be in bytes.
Loss (L): loss of the data happened during different communication rounds.
Table 3. Factors that are considered from different areas of federated learning in edge computing.
References A L T E R EC FL Env TP S CP HWR BND CC L
[1] ✔ ✔ ✔ ✔ ✔ Het ✔ ✔ ✔ ✔ ✔ ✔
[2] ✔ ✔ ✔ ✔ Het ✔ ✔ ✔
[3] ✔ ✔ Het ✔ ✔ ✔ ✔ ✔
[4] ✔ ✔ ✔ ✔
[5] ✔ ✔ ✔ ✔ ✔ ✔
[6] ✔ ✔ ✔ ✔ Het ✔
[7] ✔ ✔
[8] ✔ ✔ ✔
[9] ✔
[10] ✔ ✔ ✔ ✔
[11] ✔ ✔ ✔ ✔ ✔
[13] ✔ ✔ ✔ ✔ ✔ ✔ Het ✔ ✔ ✔
[14] ✔ ✔ ✔ ✔ ✔ Het ✔ ✔ ✔ ✔
[15] ✔ ✔ ✔ ✔ Het ✔
[16] ✔ ✔
[17] ✔ ✔ ✔ ✔
[18] ✔ ✔ ✔ ✔ ✔
18
Table 3. (Continued).
References A L T E R EC FL Env TP S CP HWR BND CC L
[19] ✔ ✔ ✔ ✔ ✔ ✔ ✔ Both ✔ ✔ ✔
[20] ✔ ✔ ✔
[21] ✔
19
traditional federated learning, as participants can continue training even when others are offline. In order to
ensure that the model is able to learn from multiple tasks, it needs to be able to identify and isolate the features
that are relevant to each task. This is accomplished by introducing regularization terms into the model, which
penalize the model for using features that are irrelevant to the task. Overall, collaborative FL provides an
efficient and secure way to share and learn from distributed data without compromising data privacy and
security[38,41]. It is particularly useful in scenarios where data is sensitive or distributed across multiple devices.
Additionally, it has been shown to improve the accuracy of machine learning models in federated settings.
Finally, from the complete study of literature, we found that federated learning in edge computing gives
better results than traditional works.
20
In future, we will consider containerization which allows developers to create and deploy applications
faster and with more security. Security framework for managing cyber security risks. Action in current edge
computing programming model to the edge federated learning for better results.
Author contributions
Material preparation, data collection and analysis, SKM, NSK, BR, B and LT; writing—original draft
preparation, NSK, BR, B and LT; writing—review and editing, SKM. All authors have read and agreed to the
published version of the manuscript.
Conflict of interest
The authors declare no conflict of interest.
References
1. Abreha HG, Hayajneh M, Serhani MA. Federated learning in edge computing: A systematic survey. Sensors 2022;
22(2): 450. doi: 10.3390/s22020450
2. Ye Y, Li S, Liu F, et al. Edge-Fed: Optimized federated learning based on edge computing. IEEE Access 2020; 8:
209191–209198. doi: 10.1109/ACCESS.2020.3038287
3. Shi W, Cao J, Zhang Q, et al. Edge computing: Vision and challenges. IEEE Internet of Things Journal 2016;
3(5): 637–646. doi: 10.1109/JIOT.2016.2579198
4. Wang Z, Xu H, Liu J, et al. Resource-efficient federated learning with hierarchical aggregation in edge computing.
In: Proceedings of the IEEE INFOCOM 2021 Conference on Computer Communications; 10–13 May 2021;
Vancouver, Canada. pp. 1–10.
5. Zeng T, Guo J, Kim KJ, et al. Multi-task federated learning for traffic prediction and its application to route
planning. In: Proceedings of the 2021 IEEE Intelligent Vehicles Symposium (IV); 11–17 July 2021; Nagoya,
Japan. pp. 451–457.
6. Liu J, Xu H, Xu Y, et al. Communication-efficient asynchronous federated learning in resource-constrained edge
computing. Computer Networks 2021; 199: 108429. doi: 10.1016/j.comnet.2021.108429
7. Qayyum A, Ahmad K, Ahsan MA, et al. Collaborative federated learning for healthcare: Multi-modal covid-19
diagnosis at the edge. IEEE Open Journal of the Computer Society 2022; 3: 172–184. doi:
10.1109/OJCS.2022.3206407
8. Wu W, He L, Lin W, Mao R. Accelerating federated learning over reliability-agnostic clients in mobile edge
computing systems. IEEE Transactions on Parallel and Distributed Systems 2020; 32(7): 1539–1551. doi:
10.1109/TPDS.2020.3040867
9. Li X, Cheng L, Sun C, et al. Federated-learning-empowered collaborative data sharing for vehicular edge
networks. IEEE Network 2021; 35(3): 116–124. doi: 10.1109/MNET.011.2000558
10. Chen M, Poor HV, Saad W, Cui S. Wireless communications for collaborative federated learning. IEEE
Communications Magazine 2020; 58(12): 48–54. doi: 10.1109/MCOM.001.2000397
11. Zaw CW, Pandey SR, Kim K, Hong CS. Energy-aware resource management for federated learning in multi-
access edge computing systems. IEEE Access 2021; 9: 34938–34950. doi: 10.1109/ACCESS.2021.3055523
12. Wang X, Han Y, Wang C, et al. In-edge AI: Intelligentizing mobile edge computing, caching and communication
by federated learning. IEEE Network 2019; 33(5): 156–165. doi: 10.1109/MNET.2019.1800286
13. Niknam S, Dhillon HS, Reed JH. Federated learning for wireless communications: Motivation, opportunities, and
challenges. IEEE Communications Magazine 2020; 58(6): 46–51. doi: 10.1109/MCOM.001.1900461
14. Xia Q, Ye W, Tao Z, et al. A survey of federated learning for edge computing: Research problems and solutions.
High-Confidence Computing 2021; 1(1): 100008. doi: 10.1016/j.hcc.2021.100008
15. Ren J, Wang H, Hou T, et al. Federated learning-based computation offloading optimization in edge computing-
supported Internet of Things. IEEE Access 2019; 7: 69194–69201. doi: 10.1109/ACCESS.2019.2919736
16. Fantacci R, Picano B. Federated learning framework for mobile edge computing networks. CAAI Transactions on
Intelligence Technology 2020; 5(1): 15–21. doi: 10.1049/trit.2019.0049
17. Wang X, Wang S, Wang Y, et al. Distributed task scheduling for wireless powered mobile edge computing: A
federated-learning-enabled framework. IEEE Network 2021; 35(6): 27–33. doi: 10.1109/MNET.201.2100179
18. Shariati B, Safari P, Mitrovska A, et al. Demonstration of federated learning over edge-computing enabled metro
optical networks. In: Proceedings of the 2020 European Conference on Optical Communications (ECOC); 6–10
December 2020; Brussels, Belgium. pp. 1–4.
19. Suomalainen J, Juhola A, Shahabuddin S, et al. Machine learning threatens 5G security. IEEE Access 2020; 8:
190822–190842. doi: 10.1109/ACCESS.2020.3031966
21
20. Guo H, Huang W, Liu J, Wang Y. Inter-server collaborative federated learning for ultra-dense edge computing.
IEEE Transactions on Wireless Communications 2021; 21(7): 5191–5203. doi: 10.1109/TWC.2021.3137843
21. Hsu RH, Wang YC, Fan CI, et al. A privacy-preserving federated learning system for Android malware detection
based on edge computing. In: Proceedings of the 2020 15th Asia Joint Conference on Information Security
(AsiaJCIS); 20–21 August 2020; Taiwan, China. pp. 128–136.
22. Mishra SK, Sahoo B, Parida PP. Load balancing in cloud computing: A big picture. Journal of King Saud
University-Computer and Information Sciences 2020; 32(2): 149–158. doi: 10.1016/j.jksuci.2018.01.003
23. Huang X, Li P, Yu R, et al. Fedparking: A federated learning based parking space estimation with parked vehicle
assisted edge computing. IEEE Transactions on Vehicular Technology 2021; 70(9): 9355–9368. doi:
10.1109/TVT.2021.3098170
24. Zhang J, Zhao Y, Wang J, Chen B. FedMEC: Improving efficiency of differentially private federated learning via
mobile edge computing. Mobile Networks and Applications 2020; 25(6): 2421–2433. doi: 10.1007/s11036-020-
01586-4
25. Mishra SK, Mishra S, Alsayat A, et al. Energy-aware task allocation for multi-cloud networks. IEEE Access 2020;
8: 178825–178834. doi: 10.1109/ACCESS.2020.3026875
26. Feng C, Zhao Z, Wang Y, et al. On the design of federated learning in the mobile edge computing systems. IEEE
Transactions on Communications 2021; 69(9): 5902–5916. doi: 10.1109/TCOMM.2021.3087125
27. Kumar Swain C, Routray P, Kumar Mishra S, Alwabel A. Predictive VM consolidation for latency sensitive tasks
in heterogeneous cloud. In: Advances in Distributed Computing and Machine Learning. Springer Singapore; 2023.
28. Chen N, Li Y, Liu X, Zhang Z. A mutual information based federated learning framework for edge computing
networks. Computer Communications 2021; 176: 23–30. doi: 10.1016/j.comcom.2021.05.013
29. Mishra SK, Puthal D, Rodrigues JJ, et al. Sustainable service allocation using a metaheuristic technique in a fog
server for industrial applications. IEEE Transactions on Industrial Informatics 2018; 14(10): 4497–4506. doi:
10.1109/TII.2018.2791619
30. Yu R, Li P. Toward resource-efficient federated learning in mobile edge computing. IEEE Network 2021; 35(1):
148–155. doi: 10.1109/MNET.011.2000295
31. Kumar Mishra S, Kumar Sahoo S, Kumar Swain C, et al. CS-based energy-efficient service allocation in cloud. In:
Advances in Distributed Computing and Machine Learning. Springer, Singapore; 2023.
32. Zhang J, Chen B, Cheng X, et al. Poisongan: Generative poisoning attacks against federated learning in edge
computing systems. IEEE Internet of Things Journal 2021; 8(5): 3310–3322. doi: 10.1109/JIOT.2020.3023126
33. Zhang W, Wang X, Zhou P, et al. Client selection for federated learning with non-IID data in mobile edge
computing. IEEE Access 2021; 9: 24462–24474. doi: 10.1109/ACCESS.2021.3056919
34. Mishra SK, Puthal D, Sahoo B, et al. Energy-efficient VM-placement in cloud data center. Sustainable
Computing: Informatics and Systems 2018; 20: 48–55. doi: 10.1016/j.suscom.2018.01.002
35. Mishra SK, Sindhu K, Teja MS, et al. Applications of federated learning in computing technologies. Convergence
of Cloud with AI for Big Data Analytics: Foundations and Innovation 2023; 107–120. doi:
10.1002/9781119905233.ch6
36. Fan S, Zhang H, Zeng Y, Cai W. Hybrid blockchain-based resource trading system for federated learning in edge
computing. IEEE Internet of Things Journal 2021; 8(4): 2252–2264. doi: 10.1109/JIOT.2020.3028101
37. Mishra SK, Puthal D, Sahoo B, et al. Energy-efficient deployment of edge dataenters for mobile clouds in
sustainable IoT. IEEE Access 2018; 6: 56587–56597. doi: 10.1109/ACCESS.2018.2872722
38. Makkar A, Ghosh U, Rawat DB, Abawajy JH. FedLearnSP: Preserving privacy and security using federated
learning and edge computing. IEEE Consumer Electronics Magazine 2022; 11(2): 21–27. doi:
10.1109/MCE.2020.3048926
39. Xiao H, Zhao J, Pei Q, et al. Vehicle selection and resource optimization for federated learning in vehicular edge
computing. IEEE Transactions on Intelligent Transportation Systems 2021; 23(8): 11073–11087. doi:
10.1109/TITS.2021.3099597
40. Rowe WST, Waterhouse RB. Edge-Fed patch antennas with reduced spurious radiation. IEEE Transactions on
Antennas and Propagation 2005; 53(5): 1785–1790. doi: 10.1109/TAP.2005.846797
41. Elmezughi AS, Rowe WST, Waterhouse RB. Further investigations into Edge-Fed cavity backed patches. In:
Proceedings of the 2007 IEEE Antennas and Propagation Society International Symposium; 9–15 June 2007;
Honolulu, USA. pp. 920–923.
22