6G Whitepaper WS
6G Whitepaper WS
1
Centre for Wireless Communications, University of Oulu, Finland, (emails:{samad.ali, nandana.rajatheva,
hamid.shiri, kai.mei}@oulu.fi).
2
Wireless@VT, Bradley Department of Electrical and Computer Engineering, Virginia Tech, USA, (email:
[email protected]).
3
Electronics and Telecommunications Research Institute (ETRI), Daejeon, South Korea (email:
[email protected]).
4
InterDigital, Inc., USA, (email: [email protected])
5
TU Dortmund University, Germany, (emails: {benjamin.sliwa, christian.wietfeld}@tu-dortmund.de)
6
Blekinge Institute of Technology, Sweden, (emails: {hans-jurgen.zepernickm, thi.my.chinh.chu}@bth.se)
7
VTT Technical Research Center, Finland, (emails: {ijaz.ahmad, jyrki.huusko}@vtt.fi)
8
Biomimetics and Intelligent Systems (BISG), University of Oulu, Finland, (email: [email protected])
9
Fraunhofer Institute for Integrated Circuits IIS, Germany, (email: [email protected])
10
IIT Indore, India, (emai: [email protected])
11
University of Quebec, Montreal, Canada, (email:[email protected])
12
IIT Hyderabad, India, (emai: [email protected])
13
Macquarie University, Australia, (email: [email protected])
14
Warwick Institute for the Science of Cities, UK, (email: [email protected])
15
Capobianco - Business Innovation Management, Pordenone, Italy, (email: [email protected])
16
ZTE Corporation, China, (email:[email protected])
17
Empirical Software Engineering in Software, Systems and Services (M3S), University of Oulu, (emails:
{maelick.claes, teemu.karvonen}@oulu.fi)
18
Princeton Univeristy, USA, (email: [email protected])
19
Ericson Research, Sweden, (email: [email protected])
20
Prontominds O, Estonia, (email: [email protected])
This paper is a compilation of ideas presented by various entities at the 6G Wireless Summit 2020. It does
not reflect an agreement on all the included technology aspects by all the involved entities.
1
Contents
1 Abstract 3
2 Introduction 4
8 Standardization Activities 26
2
1 Abstract
The focus of this white paper is on machine learning (ML) in wireless communications. 6G
wireless communication networks will be the backbone of the digital transformation of societies
by providing ubiquitous, reliable, and near-instant wireless connectivity for humans and machines.
Recent advances in ML research has led enable a wide range of novel technologies such as self-
driving vehicles and voice assistants. Such innovation is possible as a result of the availability
of advanced ML models, large datasets, and high computational power. On the other hand, the
ever-increasing demand for connectivity will require a lot of innovation in 6G wireless networks,
and ML tools will play a major role in solving problems in the wireless domain. In this paper, we
provide an overview of the vision of how ML will impact the wireless communication systems. We
first give an overview of the ML methods that have the highest potential to be used in wireless
networks. Then, we discuss the problems that can be solved by using ML in various layers of
the network such as the physical layer, medium access layer, and application layer. Zero-touch
optimization of wireless networks using ML is another interesting aspect that is discussed in this
paper. Finally, at the end of each section, important research questions that the section aims to
answer are presented.
3
2 Introduction
Todays technological aspirations will represent tomorrows reality with technologies such as holo-
graphic telepresence, eHealth and wellness applications, pervasive connectivity in smart envi-
ronments, industry 4.0 and massive robotics, massive unmanned mobility in three dimensions,
augmented reality (AR) and virtual reality (VR) to name a few. Each of them is expected to
require more effective and efficient wireless communications than ever before and 6G wireless
networks must provide broadband, near-instant, and reliable connectivity to enable massive data
exchange at different frequencies and by using a large variety of technologies. Moreover, the evo-
lution of technologies are towards more intelligent device in the internet of things (IoT) which will
require a more reliable, efficient, resilient and secure connectivity. When the connected objects
are more intelligent it becomes difficult deal with their complexity by using the communication
network in a static, simplistic and rigid manner. The same need will likely emerge for other
traditional services such as phone calls or a video streaming, where the wireless communication
network will no longer just provide a connection between two or more people, but will bring the
need to properly authenticate both parties, guarantee the security of data fluxes and recognizing
possible abnormal behaviors and events. Data exchange will be, in practice, much more than just
pure data exchange and will become the exchange of information, knowledge, experience, and
also past, present, and possibly future properties of the data. What we can easily anticipate is
the fact that larger and larger amounts of data will be transferred through the future wireless
communication networks and more added value applications and services will heavily depend on
such data exchanges. Machine learning (ML) will represent a basic functionality to guarantee
the efficiency of future wireless communication networks and, at the same time, can represent
the enabling technology for several added-value applications and services. ML on the wireless
communication nodes can enable several advanced services and quality of service functionalities
for the proposed applications.
Current wireless networks heavily rely on mathematical models that define the structure of the
communication system. Such mathematical models often do not present the systems accurately.
Moreover, there are no mathematical models for some of the building blocks of wireless networks
and devices and as a result, modeling of such blocks becomes challenging. On the other hand,
the optimization of wireless networks also requires heavy mathematical solutions that are often
not efficient in terms of computational time and complexity, and, also consume a lot of energy.
The above mentioned mathematical models and solutions will most likely fall short in improv-
ing the capacity and performance of wireless networks that are expected to meet the stringent
requirements that will be set by 6G applications [1]. ML, therefore, will play a crucial role in
6G wireless networks as it is capable of modeling systems that can not be presented by a math-
ematical equation. Moreover, it is expected that ML tools can be used to replace heuristic or
brute-force algorithms to optimize certain localized tasks. Meanwhile, it is envisioned that ML
will enable real-time analysis and automated zero touch operation and control in 6G networks.
Such intelligence will rely on the availability of data timely streamed from wireless devices, espe-
cially in extreme applications, such as real-time video monitoring and extended reality (XR). To
fully leverage these capabilities, the network should support ML-native agents that can be freely
placed and moved to the required network locations.
Furthermore, additional ML actions or predictions could be performed by mobile devices and
reported to the network to assist in decision making in resource management, making mobile
devices an integral part of the infrastructure resource. 6G networks are expected to employ ML
agents for multiple functions, including optimization of the radio interface, adaptive beamforming
strategies, network management, and orchestration. Such functionality will require data from
different domains and sources in the network. This poses additional requirements on the efficiency
of data transfer to avoid the transmission and storage of massive amounts of data that may never
be utilized over network management interfaces.
4
Figure 1: The role of ML in 6G networks.
ML algorithms should be deployed and trained at different levels of the network: management
layer, core, radio base stations, and as well as in mobile devices, possibly with the assistance of
the network itself (e.g., via configuration and/or device programmability). These new paradigms
may drive the need for ML-native and data-driven network architecture, as network functions
in the network and management domains may require data from different sources. Meanwhile,
physical-layer algorithms (e.g., link adaptation), as well as higher layer algorithms (e.g., mobility)
can be optimized with ML agents deployed in a controlled and predictable manner. Currently,
such algorithms tend to be deployed statically, whereas allowing them to change dynamically
would open up for enhanced performance and utilization. Moreover, allowing also configurations
of the network to be automatized reduces the need for expensive hands-on human work.
The white paper provides a vision for the role of ML in wireless communications by discussing
the various network problems that can utilize learning methods. A detailed look at the problems
at different layers of the communications protocol stack is provided. Moreover, ML in security of
wireless networks as well as standardization activities are also discussed.
5
3 Machine Learning Overview
ML models are computing systems that are used to learn the characteristics of a system that
can not be presented by an explicit mathematical model. These models are used in tasks such
as classification, regression, and interactions of an intelligent agent with an environment. Once
a model learns the characteristics of a system, which is known as a trained model, then it can
efficiently perform the task using some basic arithmetic calculations. ML spans three paradigms
which are known as a) supervised learning: where the model is learned by presenting input samples
and their associated outputs, b) unsupervised learning, in which, there are no output labels and
the model learns to classify samples of the input, and, c) reinforcement learning, where an agent
interacts with an environment and learns to map any input to an action. A general overview of
some of ML methods is provided in the following.
6
However, there has been a lot of work based on approximation methods such as variational Bayes,
Expectation Propagation, and sampling approaches based on Markov chain Monte Carlo to be
able scale these techniques to distributed big data challenges of wireless communication systems.
a. Each mobile device uses its collected data to train its local FL model and sends the trained
local FL model to the data center.
b. The data center integrates the local FL models to generated the global FL model and
broadcasts it back to all mobile devices.
c. Steps b. and c. are repeated until find the optimal FL models to minimize the FL loss
functions.
7
From the FL training process, we can see that mobile devices must transmit the training
parameters over wireless links. Hence, imperfect wireless transmission, dynamic wireless channels,
and limited wireless resource (e.g., bandwidth) will significantly affect the performance of FL. In
consequence, a number of the existing works such as [7] and [8] have studied the optimization
of wireless networks for the implementation of FL. Meanwhile, since FL enables mobile devices
to collaboratively train a shared ML model without data transmission, it has been studied for
solving wireless communication problems such as intrusion detection [9], orientation and mobility
prediction, and extreme event prediction.
• Why deep reinforcement learning will be one of the major components of automation of 6G
wireless networks?
• How can the goal for open data access be brought together with business-oriented mobile
network operator interests?
8
4 ML at the Physical Layer
In recent years, ML has begun to penetrate into all walks of life, including the field of wireless
communication. The physical layer of traditional wireless communication is generally designed
based on mathematical models, and several major modules are modeled and optimized separately.
This design method can adapt to the fast time-varying characteristics of the physical layer, but
often some nonlinear factors in the physical layer cannot be modeled. The research and attempt to
use ML in the physical layer of wireless communication have been carried out in recent years [11],
and some progress has been made, so it is necessary to integrate ML into the physical layer of 6G
wireless systems. There are several levels to integrate ML into 6G wireless communication system.
The first level is ML for some special functions. We should first consider using ML to replace some
functions that are not well solved at present. For example, interference detection and mitigation,
uplink and downlink reciprocity in FDD, channel prediction and so on. These problems still
exist in the current communication systems due to lack of accurate models or non-linearity. The
second level is to update the existing discrete modules. The traditional design of each module
is generally based on a linear model, so once the system encounters strong non-linear factors,
the performance of the system will decline sharply. The third level is the joint optimization of
the modules in the physical layer. As mentioned above, the traditional design of physical layer is
divided into modules and optimized separately. For example, decoding, modulation and waveform
are designed separately in the traditional design. Once the three are considered together, the
complexity of the receiver is often too high to be optimized as a whole. However, for ML, it
is not necessary to design all kinds of coding schemes carefully, nor to think about all kinds of
constellations. The best end-to-end mapping mode can be obtained automatically by learning.
Which modules in the physical layer use ML for joint optimization is a future direction to be
worth exploring. The fourth level is the integration of ML and the existing model based methods.
Although the traditional model-based method is sometimes over idealized, it can describe the
main characteristics of a process after all. If some existing model features are used for reference
in the design process of ML and input into ML as additional information, it is likely to overcome
some inherent defects of ML, such as the need for huge training data, under fitting or over fitting,
slow convergence, etc.
The above discussion provides an overview of how ML will be used in the physical layer of the
communication systems. In the following, we provide a detailed view of some of the problems in
the physical layer that can benefit from ML methods.
9
far remains tens of bits (16 or 32 by Polar Code). In other words, it is difficult to compare/predict
whether the code length actually can be learned, and how much of the benefits will be in terms of
computational complexity and time compared to the currently commercialized state-of-the-art.
The difficulty of increasing code length is that the range of codes to learn increases exponentially
as the length of code increases (e.g., to learn length-n codeword, the number of 2n cases must
be learned). There are several attempts to overcome this problem (e.g., how to learn about all
zero-codeword), but there are no clear findings yet. However, in order to be applied to actual
systems, it should be reviewed not only for performance but also for the time spent in decoding
and other aspects. For example, it is necessary to graph an optimal scheme and a sub-optimal
scheme in terms of both performance and time. In addition, these graphs may vary not only
over time but also by a service using channel coding or by parameters that a system considers
important (such as computational complexity).
4.1.2 Synchronization
In general, all devices must go through time/frequency and cell synchronization procedure without
exception, so synchronization that meets system requirements is the starting point for all stan-
dards, including 4G long-term evolution (LTE) and 5G new radio (NR). Accordingly, it is pivotal
to have synchronization technology that meets system requirement on synchronization accuracy,
even in the worst radio channel environment, the fastest mobile environment, and the highest
carrier frequency offset (CFO) environment. An end-to-end auto-encoder (AE) based communi-
cation system is likely to achieve global optimal performance, with the possibility of implementing
the communication system as an end-to-end deep neural network, including transmitter, channel
model, synchronization using sync signal (SS) as reference, and receiver [11]. However, in the
presence of sampling timing offset (STO) and sampling frequency offset (SFO) between transmit-
ter and receiver, it is still too early to perform both synchronization and signal detection/decoding
with only one auto-encoder (AE) based deep learning neural network. Recent research has shown
that deep learning technologies using SS separately from signal detection/decoding [14],[15], for-
ward error correction (FEC) based synchronization [16], and classification based synchronization
[17] were introduced.
4.1.3 Positioning
Current mobile positioning technology has identified the location of users in indoor or outdoor
environments based on various signals received from mobile devices or wireless channels using a
mathematical approach. However, a fatal problem with the mathematical approach is that if there
are non-line-of-sight (NLoS) multi-paths, there are high positional errors. As a means of solving
this problem, most of recent ML methods have been deep neural networks. To date, the deep
learning technology applied to the position technology is mostly based on indoor dimensions, and
existing fingerprint methods are characterized by learning from deep learning model and applying
it. Received signal strength (RSS), channel state information (CSI), or Hybrid information are
used as input data for the fingerprint based deep learning. The learning and experiment results of
most deep learning-based positioning technologies were used to build learning data and measure
performance in ideal environments, such as fixed experimental environments. There is no guaran-
tee that deep learning models, which represent the best performance in space A, will also perform
well in space B. Therefore, it is necessary to develop a learning model that is less sensitive to
changes in the environment and represents good performance, or to make the best learning model
for each environment. In real-world environments, input data may not be as perfect as they were
in learning (e.g., RSS information missing, if the bulb is turned off when using a light sensor,
temperature changes, environmental changes by people/things not considered, etc.). Therefore,
it is necessary to develop and analyze a learning network model that can operate when the input
10
data value changes. Most positioning systems have been carried out taking into account only
one target environment. However, it is expected that the actual system will have interference
effects in an environment with multiple people or multiple devices. Therefore, the performance
difference of given deep learning-based location-level technology between the experimental and
the actual environment must be analyzed in the course of the research. Through this analysis,
a deep learning technology with large positioning difference should facilitate evolution by means
of adaptive technology (e.g., combined with signal processing technology) to adapt to the actual
environment. On the other hand, a deep learning technology without significant positioning dif-
ference in itself should facilitate evolution through coupling with online learning in addition to
offline learning.
4.1.5 Beamforming
At the physical level, intelligent beamforming and smart antenna solutions can also greatly con-
tribute at guaranteeing performance, stability of the throughput, reduce sensitivity to interfer-
ences, extend coverage, enable highly mobile applications, and reduce energy consumption. We
already witnessed the evolution of antenna technologies from relatively dumb and basic anten-
nas to more advanced active antennas that include progressively more and more intelligence to
map the knowledge of the environment and guarantee the optimization of the radio links. This
evolution is already an integral part of 5G communication and will be boosted further with 6G
communication where all elements in the communication chain will be characterized by some
level of intelligence or at least capacity to operate in a optimal manner following some degree of
training. Again at this level ML (and more specifically deep learning) can represent the optimal
solution to support adaptive and real time massive MIMO beamforming, follow mobility patterns
to capture structural information of the radio channels, coordinate beams with neighbor base
stations, properly allocate power, adapt emission patters for mobile devices, exploit beamforming
for added value services. Dedicated hardware, other than dedicated algorithms, can help imple-
ment efficient machine learning solution to support a new generation of intelligent beamforming
and smart antennas.
11
of spectrum sensing for cognitive radios, optimal beamforming formulated as a sum rate maxi-
mization problem under a total power constraint to name only a few. This type of problems may
be solved using dual decomposition techniques that require iterative algorithms which in turn
often cannot be computed in real time due to high computational load. To alleviate the high
computational complexity and resulting latency associated with existing iterative algorithms,
heuristic solutions have been proposed for some physical layer problems such as beamforming
design. Although heuristic solutions can be obtained with low computational delay, this benefit
comes at the expense of performance loss. On the other hand, deep learning techniques have great
potential to find solutions to those problems in real time while maintaining good performance
and reducing computational delay. As such, deep learning is a powerful technique for designing,
enhancing, and optimizing one or multiple functions in the physical layer for 6G. This includes
CNNs for signal classification and DNNs for channel estimation and signal detection.
Recent research on physical layer optimization that exploit ML includes a deep learning frame-
work for optimization of multi-input multi-output downlink beamforming [19]. The CNN-based
solution takes expert knowledge into account such as uplink-downlink duality as well as the known
structure of the optimal solutions. The proposed beamforming neural network (BNN) is shown
to achieve a good trade-off between performance and computational complexity. Open questions
in this context include providing solutions for imperfect CSI and multi-cell scenarios.
In case that joint optimization of functional blocks at the physical layer is considered and the
channels are too complex for modeling, deep learning models are the best solutions for achieving
performance improvement. Conventionally, the channel estimation based on the pilot estimation
and the signal detection based on channel estimation are executed separately one after the other.
In [20], by considering the channel as a black box, a fully connected DNN with five layers is
implemented for joint channel estimation and detection. The received signals corresponding to
both the transmit signals and the pilots are taken as inputs of the DNN to recover the transmit
signals as outputs. This DNN has been shown to be more robust to the number of pilots than
conventional methods and is able to address complicated channel distortions. Future directions in
physical layer optimization with ML center around the paradigm of an autoencoder that has been
introduced in [21] aiming at a deep learning-based end-to-end physical layer architecture. In this
approach, transmitter and receiver components are jointly optimized in the presence of a given
channel. Autoencoders of the deep learning network for building the end-to-end physical layer
modules consider designing a communication system as an end-to-end reconstruction optimization
task. The autoencoder would jointly learn transmitter and receiver realizations without the need
of expert knowledge and modules. Given the complexity related to building end-to-end physical
layers, it is currently more feasible to exploit deep learning techniques for designing, enhancing,
and optimizing one or multiple functions in the physical layer for 6G.
12
Simulation
The first stage of development of a wireless modem typically is a software simulation of the
physical layer transmitter and receiver. The air interface is simulated with a channel model that
tries to recreate real world conditions such as noise, fading, multipath, Doppler spread and path
loss. Various aspects of the receiver can be implemented in an ANN which are talked about in this
paper. At this point, ML will take place in ANNs where the number of nodes, layers, connections,
activation functions and back propagation loss functions all need to be flexible while the network
trains. During this initial stage, the many parameters and characteristics of the ANN will need to
be identified with trade-offs between performance and physical resources. Even though training
of an ANN is not performed in real-time, performance considerations are still important since
there are practical limitations how long simulations can run. Offloading ML algorithms from a
Central Processing Unit (CPU) to a Graphics Processing Unit (GPU) can increase performance
by 10 to 100 times [22]. In addition specific ANN accelerators can improve performance even
more but are not always suited to supporting back-propagation required for training[23].
In order to train an ANN, many different channel models need to be employed and run in a
Monte-Carlo style simulation with multiple trials. Each trial run with a different random seed
can be fairly complex to generate and take hours to run since the model simulates impairments
at the symbol rate. How well the ANN will model real world conditions depends upon the quality
and diversity of the channel models. For illustrative purposes, if we have 30 channel models, each
is run 20 times with randomized data and the simulation takes 8 hours to run that would result
in 200 days of run time. This shows that these simulations would need to run in parallel on a
high end grid or cloud based engine. Also it is obvious that we want to reduce simulation time
by offloading the ANN to specialized hardware. One big task during simulation is to identify
the structure and features of the neural network. If we want to compare performance of several
activation functions or vary the numbers of connected nodes in each layer, we can see that the
computing resources required in the simulation stage is vast.
A big part of the design with any ML algorithm in the physical layer is to determine what
the inputs to the ANN are. Filtered outputs such as channel estimator data, FFT output, pilot
symbols or possible uniquely filtered data are all candidates as inputs to the ANN. Raw I/Q
samples would likely overwhelm any reasonably sized ANN and cause convergence to take way
too long if at all possible. Hooks into the processing stack are required to bring out any raw
data that is required as input to the ANN. Also outputs such as BLER, BER, SINR and CRC
validation will need to be fed back into the loss function.
Prototyping
After simulation, a prototype platform typically will be developed utilizing a Field Programmable
Gate Array (FPGA) as the main processing engine [24]. It is desirable to be able to run the
platform real-time or at least at a scaled down rate such as 1/2 or 1/4 of the real-time sampling
rate. We want to be able to transmit and receive over the air in the band of interest so that we
are not limited to training with predefined channel models as in the simulation stage. In this
case, ANNs can be trained over a wide set of conditions that include varying distance, rural or
urban environments, speed and weather. It is important to be careful so that when training in
one scenario, the ANN doesn’t ”forget” previous scenarios. For example the system may adapt
well to a rural environment but after then training in an urban environment, the performance in
a rural environment may suffer [25].
There are IP cores that can be synthesized into an FPGA to implement a DNN [26]. These
cores such as Xilinx’s Deep Learning Processor Unit (DPU) are highly configurable allowing
the user to allocate resources such as DSP slices, block RAM, UltraRAM, and convolutional
architecture. However these settings only allow choosing from a fixed set of possible architectures
so an extremely efficient design to fit just what is required is not possible. Also there are now
13
Impact
CSI-based Positioning
RSS-based Positioning
SS-based
synchronization
Classification-based
synchronization
Uncertainty
FEC-based synchronization
chips such as the Xilinx Versal [27] where there are up to 400 inference engines built into the
1
FPGA. This will allow for a lot of flexibility and speed in the design.
There is also an open-source framework for accelerating Deep Neural Networks on FPGAs
called DnnWeaver (dnnweaver.org). The framework lets a developer specify the ANN architecture
from a high level and then the tool automatically generates Verilog code. It is also platform
independent so it is not specific to one manufacturer over another.
With the end goal of an efficient ASIC, after acceptable performance is found, the ANN has
to be analyzed for optimization. It has been shown [23] that reducing the number of bits in
fixed point multipliers, even from 32 bits to 16 can result in only a very small performance loss
but use almost 1/8th the power and area of the die. Even quantization to 8 bits can result
in little inference loss [28]. Weights that are close to zero can be pruned so that memory is
saved in addition to computational resources as shown in [28] with minimum accuracy loss. The
assumption is that the number of nodes and layers in an ANN would not change significantly
when porting the design to an Application Specific Integrated Circuit (ASIC).
Product Phase
Any final product with an ANN to facilitate physical layer processing will have to place hard
limits on the number of nodes, layers and bits in fixed point MAC operations. Once the design is
ported to an ASIC, it will be assumed that a fully trained ANN will be imported to the design.
However there has to still be some flexibility in updating the network as well so that weights and
some connection information can be updated through software downloads.
Design considerations have to be made regarding which inputs and outputs will be available
to/from the ANN. Allowing the ANN to reside on a separate co-processor, requiring moving data
off chip can take up more than the available timeline. Any ANN would have to be treated the
same as any physical layer processing block where the data is readily available and the neural net
is part of the processing chain.
14
Performance degradation due to offline learning and actual mismatch in the wireless channel
environment is expected to be relatively higher for positioning items. Because channel coding
assumes radio channel compensation, the effect of radio channels is reduced in the form of colored
noise. And synchronization is not to perform any channel estimation itself but to perform a
correlation greater than the synchronization signal length for a given radio channel, which may
be less affected by environmental inconsistencies. On the other hand, positioning is based on
the fact that it is directly related to the nature of radio channels and therefore more likely to be
affected by environmental inconsistencies.
MOLA
• Channel coding
• Synchronization
• Positioning
SOLO
Secondly, as shown in Fig. 3, we outline the research direction regarding2 deep learning-
driven PHY-layer technologies. From now until x years21 , it is expected to have two directions
as follows. The first is Multi-Offline Learning and Adaptation (MOLA), which will perform
offline learning on a number of channel models in advance, file-up to the system, and monitor the
actual radio channel characteristics to apply the appropriate offline learning to the system. The
second is Single Offline Learning and Online Learning (SOLO), which identifies the performance
sensitivity of each radio channel characteristic factor, and applies offline learning based on the
least sensitive factors to the actual system, and online learning to adapt to the actual radio channel
characteristics online. Next, it is expected that after y years, it will be used either MOLA or
SOLO depending on the radio channel situation. The classification criteria in this figure are as
follows. The MOLA is expected to take a long time as it will require vast amounts of data bases
and memory, but is expected to be used effectively in some wireless channel environments. In
addition, in radio-channel environments that are not covered by MOLA, SOLO is expected to be
applied in a semi-optimal manner, but this is a prediction and should not be ruled out that if the
SOLO can cover the MOLA in both performance and implementation terms, it can eventually go
in the form of SOLO.
In this section, we have tried to answer the following research questions:
15
• How does deep learning-based physical layer optimization perform under imperfect CSI,
channel correlation, and other imperfections?
• How can deep learning-based physical layer optimization be combined with the intelligence
in upper layers?
• How to reduce the training workload of deep learning models for physical layer optimization
by using the recent advancement of deep learning such as domain adaptation and transfer
learning?
• How to reduce training data by applying the advancement of generative adversarial networks
to generate artificial data in the physical layer?
16
5 ML at the Medium Access Control Layer
Medium access control (MAC) layer of a cellular networks performs tasks such as user selection,
user pairing for MIMO systems, resource allocation, modulation and coding scheme selection,
power control of uplink transmissions and random access and handover control. Several heuristic
algorithms are in place currently to address these problems considering the complexity of the
problem. There are no optimal solutions for these problems in real environments. ML tools must
be leveraged to significantly enhance the MAC scheduler in order to provide significant gains
in real environments. While optimal solutions are not available, significant thought must go on
how to train ML models for these problems. Hence, reinforcement learning frameworks are most
sutiable for problems in which the network can adapt to varying users conditions such as channel
conditions and learn the optimal strategies. For example, a scheduler must learn to predict the
buffer traffic characteristics, speed and channel variations over time and use these predictions
to make intelligent scheduling decisions. Care must be taken because the state-action space can
grow large very quickly in such a situation. Intelligent deep reinforcement learning algorithms
that can deal with combinatorial actions spaces and multi-agent environments must be explored
for these problems. In the following, we provide some use cases for which ML can be used in
MAC layer communications.
17
directions on the predictive resource allocation for MTC are presented in [35]. However, there are
many open problems to be solved. The first is to study various types of data traffic originating
from MTC and to find the right tools to solve the source traffic prediction problem. For example,
event-driven traffic prediction requires sophisticated ML tools for event detection and traffic
prediction. Second is optimal predictive resource allocation using online decision-making tools in
various systems such as Non-Orthogonal Multiple Access (NOMA), massive MIMO, and cell-free
massive MIMO. Therefore, it is clear that ML plays a big role in enabling predictive resource
allocation mechanisms in future MTC networks.
18
have interference from neighboring cell uplink users with low transmit power and therefore, cor-
responding downlink power can be adjusted. ML based algorithms can drive such techniques in a
proactive manner as the entire concept is based on traffic pattern and network activity data and
would result in increased system performance.
In this section, we have tried to answer the following research questions:
19
6 ML for Security of Wireless Networks
This section give a brief discussion on the role of ML in security of the future wireless systems.
First, a general road-map towards 6G security is provided which is then followed by security
aspects in wireless medium.
20
nodes and may potentially stop important communication. There are two aspects for wireless
security that must be studied - defense and attack. Defensive mechanisms include cryptography
and the likes, while attacks refer to mechanisms where proactively an attack such as jamming or
eavesdropping is performed to secure the future transmissions. Such security-related studies not
only allow for the analysis of the system vulnerabilities but also enable to undermine an enemy
system capabilities.
The fast pace of research in the field of ML can potentially enable every device to possess some
sort of intelligence that can be used either in a positive or in a negative manner. If such capabilities
exist with the malicious devices i.e., the ones that want to intentionally cause interference, then
it is a threat to the security of the various devices that co-exist in the same environment. It is
thus highly important that devices be intelligent to know everything about the adversary so as
to limit the effectiveness of attacks.
Typically, such problems have been addressed via game theory or optimization frameworks.
While they give good insights, they often assume static environments, or static action space for an
adversary etc which may not be the case when an adversary himself possesses the ML capabilities.
Therefore, we must study these systems both from attack [37] and defense [38] perspective. From
an attack perspective, we need to design ML models that can learn the environment in real-time
and stop the adversary from communicating or interfering with the required network. From a
defense perspective, we need to design a communication system that is robust against any kind
of attacks and adversarial ML mechanisms can be used for the same to design robust techniques.
In this section, we have tried to answer the following research questions:
• What is the role of Machine Learning in 6G Security (beyond ML-based security in 5G)?
• What aspects of security, physical layer, mac layer, network layer can be addressed via
Machine Learning?
• What and where does machine learning in security find use cases? Defense applications as
an example?
21
7 ML at the Application Layer
ML solutions directly embedded on the wireless communication nodes at the lower layers, with
advanced features such as context awareness, performance optimization, and multi-agent rein-
forcement learning, will enable more reliable and stable per user and per-application data rate,
peak data rate, air-interface latency, spectrum efficiency, energy efficiency. At the same time
Embedded ML solutions on the wireless communication nodes at the transport layer or the ap-
plication layer, with sensor fusion techniques and with the capacity to run ML as a service, will
improve experience sharing, remote control capacity, seamless connectivity, and services.
Context-aware systems, in particular, provide the capacity to implement services that maxi-
mize the applications safety while minimizing the applications explicit interaction with the envi-
ronment. In general, a set of rules has to be specified for the possible contextual configurations
and each rule is assigned to a specific service in the context-aware system. This is a common
problem to determine and limit the set of possible context configurations. Instead of a rule-based
approach, ML can be used to predict all possible and meaningful context configurations. This
ML approach can use the previous choice about the service and can adapt itself by a new choice
about the service from user/application feedback information. A variety of ML techniques can
help to develop general-purpose context-aware applications without needing to define a priori
rules and elements of the context. This context-aware application can provide service proactively
by using different types of learning algorithms in an operational environment that can be smart
and continuously changing. The user preferences (choice for services) may also change over time
so that an adaptive learning algorithm would certainly be preferable. The middleware layer plays
a vital role in the context-aware system. The middleware is responsible for context modeling,
context reasoning and controlling sensors and data sources, appliances and devices based on the
decision from the context-aware application layer.
Making ML available as a service on wireless communication nodes will flexibility and power
to the communication networks. Four key trends are making ML more accessible to users and
companies: (1) improved processing power, (2) reduced costs for data storage and processing, (3)
expanding data availability, and (4) improved techniques, like the emergence of cloud-based deep
learning solutions. Hybrid cloud and fog computing might likely further extend such accessibility
by making ML available as a service for users and applications in the application layer of wireless
communication nodes.
22
Figure 4: ML enhancing 6G network performance management.
could enhance the performance of coverage, throughput, QoS prediction, automatic network con-
figuration, power control, operation, maintenance, fault management, power-saving, and beam
management. Fig. 4 shows ML enhancing 6G network performance management aspects.
23
ability, safety, and energy consumption of the UAV control are improved. This way of control
where both the ML training and the communications benefit from each other is extensively stud-
ied in [45]. Other possible ML and communication co-design use-cases such as [46] intelligently
utilize communications resources with the help of predictions provided by the ML, and [47] solve
a distributed ML problem in a communication efficient way. Based on these research examples,
considering communication and ML/control can provide many advantages. However, the control
and communications co-design is still a challenging issue that needs to be addressed further in
6G.
24
Figure 5: Opportunistic data transfer in vehicular networks.
a data-driven monitor, extract data, learn, predict cycles in development of the systems and
services. Consequently, one of the first steps is that existing engineering tools, methods and
processes should evaluated based on their adaptivity to above described ML-driven development
loop. This gives an overacrching understanding of the magnitude of changes and investments that
are required in industry domains.
In parallel, with data science and ML gaining in popularity, software problems specific to
the field also became apparent. Systems relying heavily on ML not only share the same issues
that other software systems encounter, but have additional long-term shortcomings that can
incur high maintenance costs for real-world usage. In the past years, DataOps, a movement
inspired by DevOps, has emerged to better deal with those problems specific to data science and
ML [49]. This movement aims at providing development processes and software technologies to
improve quality when delivering data-based solutions in a fast-changing world. To provide ML
solutions at a large scale for wireless systems, 6G will have to embrace development practices
from Agile software development, DevOps, and DataOps. Moreover, movements like DevOps and
DataOps are relatively new and in an ever-evolving state. Thus, because networking and wireless
communications have their specificities and requirements of their own, people with an active role
in the development of 6G might also have to take an active role in these movements.
In this section, we have tried to answer the following research questions:
• How will ML enable, enhance and automate the network performance management for 6G
Mobile networks?
• How will ML will enable, enhance and automate the 6G mobile network optimization?
• What existing software development practices, processes, and technologies will be needed
in order to incorporate ML in large scale real-world networking and wireless communication
technologies?
• What are specificities of 6G that will require to adapt existing or create new Agile, DevOps,
or DataOps practices, processes, and technologies?
25
Task categories
Level
Execution Awareness Analysis Decision Intent translation
L0 Manual operating network Human Human Human Human Human
L1 Assisted operating network Human & Network system Human & Network system Human Human Human
L2 Preliminary autonomous network Network system Human & Network system Human & Network system Human Human
L3 Intermediate autonomous network Network system Network system Human & Network system Human & Network system Human
L4 Advanced autonomous network Network system Network system Network system Network system Human & Network system
L5 Fully autonomous network Network system Network system Network system Network system Network system
8 Standardization Activities
Various standardization bodies like 3GPP, and International Telecommunication Union (ITU),
but also the 5GAA (5G Automotive Association) have started evaluating ML in 5G and future
networks. From a standardization perspective, the ML models and algorithms will not be stan-
dardized [50]. Other bodies such as the ORAN alliance have started defining open interfaces
in order to exchange relevant information between various parts of the protocol stack. Specifi-
cally, they have defined entities’ names as a real-time intelligent controller and a non-real-time
intelligent controller. Non-real time RIC is one where the training for the ML models happens
using the data captured by lower layers. This learning happens very slowly and hence the word
non-realtime.
This learned model is fed into the real-time RIC which uses this model on real-time data and
makes real-time decisions in an online fashion. Such systems can be deployed in core networks or
in RAN based on the type of data that can be collected.
The discussion of introducing ML capabilities in the 3GPP RAN is still in the preliminary
stage in the standardization. The autonomous network is an important topic for RAN considering
the complexity of future networks. Six levels of automation are proposed for the RAN. Level zero
(L0) starts with a manual operating network and ends with L5 at fully autonomous networks
with no human involvement at any stage. The levels are summarized in the Table 1 along with
the tasks [51]. Additionally, it is also required to define
• data required by the ML algorithms either reported by the user equipment (UE) or collected
from an NG-RAN node, and
• outputs generated by the algorithms to be delivered to the network including the network
functions and core network.
Also, if the UE has the capability to support at least a part of ML inference on board then it be-
comes relevant to study how the ML-enabled UE obtains an updated ML model and intermediate
output based on dynamic environment changes and application. It is unfeasible to pre-load all
possible models on-board because of limited storage space in the UEs. Therefore, the ML model
downloading or transfer learning is needed. ITU-T Rec. Y.3172 defines a technology-agnostic
logical architecture model for the high-level machine learning requirements such as interfaces,
support for heterogeneous data sources, machine learning mechanisms in future networks. The
actual underlay network technology (e.g., 4G, 5G, 6G, IEEE 802.11) is virtually mirrored by a
digital twin referred to as closed-loop subsystem – which is utilized to safely explore the outcomes
of different machine learning-enabled acting options.
Acknowledgement
This draft white paper has been written by an international expert group, led by the Finnish 6G
Flagship program (6gflagship.com) at the University of Oulu, within a series of twelve 6G white
papers to be published in their final format in June 2020.
26
References
[1] W. Saad, M. Bennis, and M. Chen, “A vision of 6G wireless systems: Applications, trends, technologies, and
open research problems,” IEEE Network, pp. 1–9, 2019.
[2] A. Zappone, M. D. Renzo, and M. Debbah, “Wireless networks design in the era of deep learning: Model-
based, AI-based, or both?” IEEE Communications Magazine, 2020.
[3] Z. Ghahramani, “Probabilistic machine learning and artificial intelligence,” Nature, vol. 521, no. 7553, pp.
452–459, May 2015.
[4] R. Mitra, F. Miramirkhani, V. Bhatia, and M. Uysal, “Mixture-kernel based post-distortion in RKHS for
time-varying VLC channels,” IEEE Transactions on Vehicular Technology, vol. 68, no. 2, pp. 1564–1577, Feb
2019.
[5] R. Mitra, S. Jain, and V. Bhatia, “Least minimum symbol error rate based post-distortion for VLC using
random fourier features,” IEEE Communications Letters, vol. 24, no. 4, pp. 830–834, 2020.
[6] M. Chen, U. Challita, W. Saad, C. Yin, and M. Debbah, “Artificial neural networks-based machine learning
for wireless networks: A tutorial,” IEEE Communications Surveys and Tutorials, vol. 21, no. 4, pp. 3039–3071,
Fourthquarter 2019.
[7] M. Chen, Z. Yang, W. Saad, C. Yin, H. V. Poor, and S. Cui, “A joint learning and communications framework
for federated learning over wireless networks,” arXiv preprint arXiv:1909.07972, 2019.
[8] M. Chen, H. V. Poor, W. Saad, and S. Cui, “Convergence time optimization for federated learning over
wireless networks,” arXiv preprint arXiv:2001.07845, 2020.
[9] A. Ferdowsi and W. Saad, “Generative adversarial networks for distributed intrusion detection in the Internet
of Things,” arXiv preprint arXiv:1906.00567, 2019.
[10] A. Taleb Zadeh Kasgari, W. Saad, M. Mozaffari, and H. V. Poor, “Experienced deep reinforcement learning
with generative adversarial networks (GANs) for model-free ultra reliable low latency communication,” arXiv,
pp. arXiv–1911, 2019.
[11] D. Gndz, P. de Kerret, N. D. Sidiropoulos, D. Gesbert, C. R. Murthy, and M. van der Schaar, “Machine
learning in the air,” IEEE Journal on Selected Areas in Communications, vol. 37, no. 10, pp. 2184–2199,
2019.
[12] E. Nachmani, Y. Be’ery, and D. Burshtein, “Learning to decode linear codes using deep learning,” in 2016
54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Sep. 2016, pp.
341–346.
[13] A. Askri and G. R. Othman, “DNN assisted sphere decoder,” in 2019 IEEE International Symposium on
Information Theory (ISIT), July 2019, pp. 1172–1176.
[14] H. Wu, Z. Sun, and X. Zhou, “Deep learning-based frame and timing synchronization for end-to-end commu-
nications,” in Journal of Physics: Conference Series, vol. 1169, no. 1. IOP Publishing, 2019, p. 012060.
[15] J. Schmitz, C. von Lengerke, N. Airee, A. Behboodi, and R. Mathar, “A deep learning wireless transceiver with
fully learned modulation and synchronization,” in 2019 IEEE International Conference on Communications
Workshops (ICC Workshops), May 2019, pp. 1–6.
[16] T. A. Chadov, S. D. Erokhin, and A. I. Tikhonyuk, “Machine learning approach on synchronization for FEC
enabled channels,” in 2018Systems of Signal Synchronization, Generating and Processing in Telecommunica-
tions (SYNCHROINFO), July 2018, pp. 1–4.
[17] A. Li, Y. Ma, S. Xue, N. Yi, R. Tafazolli, and T. E. Dodgson, “Unsupervised deep learning for blind mul-
tiuser frequency synchronization in OFDMA uplink,” in ICC 2019 - 2019 IEEE International Conference on
Communications (ICC), May 2019, pp. 1–6.
[18] H. Ye, G. Y. Li, and B. Juang, “Power of deep learning for channel estimation and signal detection in OFDM
systems,” IEEE Wireless Communications Letters, vol. 7, no. 1, pp. 114–117, Feb 2018.
[19] W. Xia, G. Zheng, Y. Zhu, J. Zhang, J. Wang, and A. P. Petropulu, “A deep learning framework for
optimization of MISO downlink beamforming,” IEEE Transactions on Communications, vol. 68, no. 3, pp.
1866–1880, 2020.
[20] H. Ye, G. Y. Li, and B. Juang, “Power of deep learning for channel estimation and signal detection in OFDM
systems,” IEEE Wireless Communications Letters, vol. 7, no. 1, pp. 114–117, 2018.
27
[21] T. OShea and J. Hoydis, “An introduction to deep learning for the physical layer,” IEEE Transactions on
Cognitive Communications and Networking, vol. 3, no. 4, pp. 563–575, 2017.
[22] A. Kayid, Y. Khaled, and M. Elmahdy, Performance of CPUs/GPUs for Deep Learning workloads. The
German University in Cairo, 2018, no. 10.13140/RG.2.2.22603.54563.
[23] T. Chen, N. S. Z. Du, J. Wang, C. W. Y. Chen, and O. Temam, “Diannao: a small-footprint high-throughput
accelerator for ubiquitous machine-learning,” ASPLOS, 2014.
[24] A. Shawahna, S. M. Sait, and A. El-Maleh, “FPGA-based accelerators of deep learning networks for learning
and classification: A review,” IEEE Access, vol. 7, pp. 7823–785, 2019.
[25] J. Kirkpatrick, R. Pascanu, N. Rabinowitz, J. Veness, G. Desjardins, A. A. Rusu, K. Milan, J. Quan, T. Ra-
malho, A. Grabska-Barwinska et al., “Overcoming catastrophic forgetting in neural networks,” arXiv, 2016.
[26] Xilinx. (2019, March) DPU IP product guide. [Online]. Available: https://www.xilinx.com/support/
documentation/ip documentation/dpu/v1 1/pg338-dpu.pdf
[27] (2019, 7) Xilinx ships heterogeneous chips for AI, 5G. EETimes. [Online]. Available: https:
//www.eetimes.com/xilinx-ships-heterogeneous-chips-for-ai-5g/
[28] Y. Chen, Y. Xie, L. Song, F. Chen, and T. Tang, “A survey of accelerator architectures for deep neural
networks,” Engineering, 2020. [Online]. Available: https://doi.org/10.1016/j.eng.2020.01.007
[29] M. Chen, O. Semiari, W. Saad, X. Liu, and C. Yin, “Federated echo state learning for minimizing breaks
in presence in wireless virtual reality networks,” IEEE Transactions on Wireless Communications, vol. 19,
no. 1, pp. 177–191, Jan 2020.
[30] M. Chen, W. Saad, and C. Yin, “Echo-liquid state deep learning for 360◦ content transmission and caching in
wireless VR networks with cellular-connected UAVs,” IEEE Transactions on Communications, vol. 67, no. 9,
pp. 6386–6400, Sep. 2019.
[31] M. Chen, W. Saad, C. Yin, and M. Debbah, “Data correlation-aware resource management in wireless virtual
reality (VR): An echo state transfer learning approach,” IEEE Transactions on Communications, vol. 67,
no. 6, pp. 4267–4280, June 2019.
[32] M. Chen, W. Saad, and C. Yin, “Virtual reality over wireless networks: Quality-of-service model and learning-
based resource management,” IEEE Transactions on Communications, vol. 66, no. 11, pp. 5621–5635, Nov
2018.
[33] C. Hoymann, D. Astely, M. Stattin, G. Wikstrom, J.-F. Cheng, A. Hoglund, M. Frenne, R. Blasco, J. Huschke,
and F. Gunnarsson, “LTE release 14 outlook,” IEEE Communications Magazine, vol. 54, no. 6, pp. 44–49,
2016.
[34] S. Ali, N. Rajatheva, and W. Saad, “Fast uplink grant for machine type communications: Challenges and
opportunities,” IEEE Communications Magazine, vol. 57, no. 3, pp. 97–103, March 2019.
[35] S. Ali, A. Ferdowsi, W. Saad, and N. Rajatheva, “Sleeping multi-armed bandits for fast uplink grant alloca-
tion in machine type communications,” in Proc. IEEE Global Communications Conference (GLOBECOM),
Workshop on Ultra-High Speed, Low Latency and Massive Connectivity Communication for 5G/B5G, Abu
Dhabi, UAE, Dec 2018, pp. 1–6.
[36] P. Sharma, H. Liu, H. Wang, and S. Zhang, “Securing wireless communications of connected vehicles with
artificial intelligence,” in 2017 IEEE International Symposium on Technologies for Homeland Security (HST),
April 2017, pp. 1–7.
[37] S. Amuru, C. Tekin, M. van der Schaar, and R. M. Buehrer, “Jamming banditsa novel learning method for
optimal jamming,” IEEE Transactions on Wireless Communications, vol. 15, no. 4, pp. 2792–2808, 2015.
[38] T. Erpek, Y. E. Sagduyu, and Y. Shi, “Deep learning for launching and mitigating wireless jamming attacks,”
IEEE Transactions on Cognitive Communications and Networking, vol. 5, no. 1, pp. 2–14, 2018.
[39] J. Lam and R. Abbas, “Machine learning based anomaly detection for 5G networks,” arXiv preprint
arXiv:2003.03474, 2020.
[40] M. Mozaffari, W. Saad, M. Bennis, and M. Debbah, “Efficient deployment of multiple unmanned aerial
vehicles for optimal wireless coverage,” IEEE Communications Letters, vol. 20, no. 8, pp. 1647–1650, Aug.
2016.
[41] H. Shiri, J. Park, and M. Bennis, “Massive autonomous UAV path planning: A neural network based mean-
field game theoretic approach,” in 2019 IEEE Global Communications Conference (GLOBECOM), 2019, pp.
1–6.
28
[42] J. Park, S. Samarakoon, H. Shiri, M. K. Abdel-Aziz, T. Nishio, A. Elgabli, and M. Bennis, “Extreme URLLC:
Vision, challenges, and key enablers,” 2020.
[43] H. Shiri, J. Park, and M. Bennis, “Remote UAV online path planning via neural network based opportunistic
control,” IEEE Wireless Communications Letters, pp. 1–1, 2020.
[44] H. Shiri, J. Park, and M. Bennis, “Communication-efficient massive UAV online path control: Federated
learning meets mean-field game theory,” 2020.
[45] J. Park, S. Samarakoon, M. Bennis, and M. Debbah, “Wireless network intelligence at the edge,” Proceedings
of the IEEE, vol. 107, no. 11, pp. 2204–2239, 2019.
[46] M. S. Elbamby, C. Perfecto, M. Bennis, and K. Doppler, “Edge computing meets millimeter-wave enabled
VR: Paving the way to cutting the cord,” 2018 IEEE Wireless Communications and Networking Conference
(WCNC), Apr 2018. [Online]. Available: http://dx.doi.org/10.1109/WCNC.2018.8377419
[47] A. Elgabli, J. Park, A. S. Bedi, M. Bennis, and V. Aggarwal, “GADMM: Fast and communication efficient
framework for distributed machine learning,” 2019.
[48] B. Sliwa, R. Falkenberg, T. Liebig, N. Piatkowski, and C. Wietfeld, “Boosting vehicle-to-cloud communication
by machine learning-enabled context prediction,” IEEE Transactions on Intelligent Transportation Systems,
Jul 2019.
[49] J. Ereth, “Dataops-towards a definition.” in LWDA, 2018, pp. 104–112.
[50] 3rd Generation Partnership Project, “SA1 study item description:”study on AI/ML model transfer in 5GS”,”
2019.
[51] 3rd Generation Partnership Project (3GPP), “3GPP TR 28.810:” study on concept, requirements and solu-
tions for levels of autonomous network”; release 16, v0.2.0,” 2019.
29