0% found this document useful (0 votes)
165 views10 pages

Green Ai For Iiot: Energy Efficient Intelligent Edge Computing For Industrial Internet of Things

This document discusses using intelligent edge computing to improve the energy efficiency of AI-driven applications in industrial IoT (IIoT) environments. It proposes an intelligent edge computing framework with heterogeneous hardware that can offload AI tasks from servers to nearby edge devices. It also describes an algorithm to optimize scheduling of different AI tasks across various computing resources to enhance energy efficiency. Simulation results show the proposed online scheduling strategy reduces energy consumption by over 20% compared to static scheduling and over 40% compared to a first-in, first-out strategy.

Uploaded by

Bishmay
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
165 views10 pages

Green Ai For Iiot: Energy Efficient Intelligent Edge Computing For Industrial Internet of Things

This document discusses using intelligent edge computing to improve the energy efficiency of AI-driven applications in industrial IoT (IIoT) environments. It proposes an intelligent edge computing framework with heterogeneous hardware that can offload AI tasks from servers to nearby edge devices. It also describes an algorithm to optimize scheduling of different AI tasks across various computing resources to enhance energy efficiency. Simulation results show the proposed online scheduling strategy reduces energy consumption by over 20% compared to static scheduling and over 40% compared to a first-in, first-out strategy.

Uploaded by

Bishmay
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO.

1, MARCH 2022 79

Green AI for IIoT: Energy Efficient Intelligent Edge


Computing for Industrial Internet of Things
Sha Zhu, Student Member, IEEE, Kaoru Ota , Member, IEEE, and Mianxiong Dong , Member, IEEE

Abstract—Artificial Intelligence (AI) technology is a huge infrastructure and energy consumption [9], [10]. Large cloud
opportunity for the Industrial Internet of Things (IIoT) in the providers also have good energy efficiency due to the high
fourth industrial revolution (Industry 4.0). However, most AI- utilization of resources and mature scheduling strategies on
driven applications need high-end servers to process complex AI
tasks, bringing high energy consumption to IIoT environments. homogeneous architectures [11]. However, in IIoT environ-
In this article, we introduce intelligent edge computing, emerg- ments, the cloud-based structure for general AI applications
ing technology to reduce energy consumption in processing AI loses its effectiveness due to inconsistency between the long
tasks, to build green AI computing for IIoT applications. We first distance from the cloud and high real-time requirements of
propose an intelligent edge computing framework with a hetero- IIoT applications [12]. By bringing AI computing near to
geneous architecture to offload most AI tasks from servers. To
enhance the energy efficiency of various computing resources, we the location where it is needed, intelligent edge computing
propose a novel algorithm to optimize the scheduling for differ- becomes an excellent opportunity for AI-driven IIoT because
ent AI tasks. In the performance evaluation, we build a small the data transfer time is reduced [13]. Some intelligent edge
testbed to show the AI-driven IIoT applications’ energy efficiency nodes have adequate capability to process complex AI tasks,
with intelligent edge computing. Meanwhile, extensive simulation including computer vision and natural language processing to
results show that the proposed online scheduling strategy con-
sumes less than 80% energy of the static scheduling and 70% of support AI-driven IIoT applications [14]. Meanwhile, unlike
the first-in, first-out (FIFO) strategy in most settings. homogeneous structure in most cloud platforms, it is possible
to customize intelligent edge computing with heterogeneous
Index Terms—Green computing, intelligent edge, industrial
Internet of Things (IIoT), artificial intelligence (AI). hardware, which has good energy efficiency in processing
specific AI tasks [15], [16].
This paper tries to introduce intelligent edge computing into
I. I NTRODUCTION the IIoT scenario to support energy-efficient AI-driven IIoT
applications. The most challenging problem in IIoT environ-
RTIFICIAL intelligence (AI) driven Industrial Internet of
A Things (IIoT) will play a prominent role in supporting
high automation manufacturing [1], [2]. For example, different
ments is to guarantee the processing latency, which is a matter
of life and death to most industrial applications [17], [18]. In
our design, a framework organizes all heterogeneous resources
object recognition models have been deployed in automated and then finds the most feasible hardware to handle differ-
product assembly, factory unattended monitoring, automated ent AI tasks in the required processing time. Since existing
quality inspections, and other core processes of industrial pro- AI tasks are developed in several major AI libraries such as
duction [3], [4]. Since most AI tasks are compute-bound, PyTorch or TensorFlow, the framework can connect various AI
it is necessary to deploy high-end infrastructure to sup- tasks and different AI hardware without modification. Another
port AI-driven IIoT in finishing each task in a short time problem is to schedule heterogeneous hardware for better
interval [5], [6]. However, high-performance computing hard- energy efficiency in processing different AI tasks in intelli-
ware, including processors, graphics processing units (GPUs), gent edge [19]. Because special AI hardware has better energy
and tensor processing units (TPUs), usually has high energy efficiency in processing specific tasks, the proposed frame-
consumption in processing AI tasks [7]. As one of the most work’s scheduling strategy leverages the task type and latency
efficient computing devices for AI tasks, a typical NVIDIA requirement to minimize energy consumption and guarantee
RTX GPU has 400 watts peak power consumption [8]. processing latency.
In general AI-driven applications, users apply cloud to We also implement the demonstration system in a small
process AI tasks to reduce the additional cost from both testbed with different edge devices to test the process
latency of several basic AI tasks in the proposed frame-
Manuscript received March 16, 2021; revised June 17, 2021; accepted
July 16, 2021. Date of publication August 20, 2021; date of current version work. Meanwhile, the scheduling strategy in the intelligent
February 16, 2022. This work was supported in part by JSPS KAKENHI edge is evaluated in large-scale scenarios by extensive simula-
under Grant JP19K20250, Grant JP20F20080, and Grant JP20H04174; and tions. The simulation results show that the online scheduling
in part by Leading Initiative for Excellent Young Researchers (LEADER),
MEXT, Japan. (Corresponding author: Kaoru Ota.) strategy can reduce more than 20% energy consumption
The authors are with the Department of Sciences and Informatics, compared to static scheduling strategy and more than 40%
Muroran Institute of Technology, Muroran 050-0071, Japan (e-mail: compared to the first-in, first-out (FIFO) strategy in most
[email protected]; [email protected]; mx.dong@
csse.muroran-it.ac.jp). settings. We list three major contributions of this paper as
Digital Object Identifier 10.1109/TGCN.2021.3100622 follows.
2473-2400 
c 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
80 IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO. 1, MARCH 2022

•We first design an intelligent edge computing frame- detection which is an important issue in automated manufac-
work to improve the energy efficiency of AI-driven IIoT turing. Supervised deep learning [24], [25], semi-supervised
applications. To the best of our knowledge, the proposed deep learning [26], and unsupervised deep learning methodolo-
framework is original and creative. gies [27] have been introduced to solve the anomaly detection
• We then state the scheduling problem in the proposed problem in IIoT environments. Meanwhile, some researchers
framework and give two solutions to solve the problem apply new deep learning methods, such as federated learn-
in both static and online scenarios, respectively. ing [28] and explainable AI [29], to detect anomalies in
• A demonstration framework is developed in our manufacturing.
testbed, and the intelligent edge devices show excel- Radio-frequency identification (RFID) is an efficient tech-
lent performance-to-power ratio in processing AI tasks. nology in sensing objects in manufacturing, which is stabler
Meanwhile, results from extensive simulations show than the vision-based way in many scenarios [30], [31].
that two proposed scheduling methods outperform other Huang et al. [32] show a deep learning-based approach
solutions. to monitor and predict manufacturing progress in factories.
Section II introduces related work of AI-driven IIoT and The proposed approach design a transfer learning-based AI
intelligent edge computing. Section III introduces the scenario model to improve the prediction performance with insufficient
of AI-driven IIoT applications and the design of the proposed labeled data. However, due to the gap between AI technolo-
framework. Section IV formulates the scheduling problem and gies and industrial informatics, the AI-driven IIoT area still
gives two solutions. Section V shows the demonstration frame- needs a stable and scalable solution for more comprehensive
work in the testbed and performance evaluation, followed by applications.
the conclusions and future work drawn in Section VI.

B. Intelligent Edge Computing


II. R ELATED W ORK
Intelligent edge computing is the opportunity for build-
In this section, we introduce existing research related to
ing a stable and scalable infrastructure for future AI-driven
AI-driven IIoT and the intelligent edge.
IIoT applications. After Edge computing meets AI technolo-
gies, intelligent edge computing brings an evolution to the
A. AI-Driven IIoT traditional edge and IoT areas.
Bisio et al. [20] first investigate the relationship between The first stage of the intelligent edge is offloading AI tasks
industry 4.0 and IoT concept and then analyze several context- from the cloud to the edge. However, the spec of earlier edge
aware schemes of AI-driven IIoT applications. From their hardware is too low to execute traditional compute-bound AI
work, AI-driven IIoT systems can understand the surround- tasks. There are several ways to reduce the computing require-
ing environment and react to different conditions according to ment of AI tasks for deployment in edge devices. The simplest
the sensed data and pre-trained models. way is pruning the AI model and trimming weights to reduce
In IIoT environments, two major areas apply AI technolo- the model complexity. Tiny-YOLO [33] is the most successful
gies, data analysis, and automated manufacturing. Various object recognition model for low-spec devices, which main-
data analysis-based predictions, such as demand forecast- tains good enough performance after drastically reducing the
ing and predictive maintenance, are significant to industrial network complexity. However, since it is hard to fit different
manufacturing. Therefore, AI technologies, including tradi- hardware with fixed pruning, Bhattacharya and Lane [34] pro-
tional machine learning and deep learning methods, have been pose an efficient pruning method named weight factorization
applied in analyzing data collected from IIoT environments. to optimize the complex AI models adaptively.
Demand forecasting is usually applied for forecasting energy Segmentation of AI models is another solution that divides
demand in the IIoT environment. Han et al. [21] propose existing AI models into different parts and pushes parts of
a deep learning framework that adopts a Long short-term models to edge devices. The benefit of AI model segmen-
memory (LSTM) model to predict the future energy con- tation reduces the computing load of edge devices without
sumption of IIoT devices. In their design, cloud computing model modification. Li et al. [35] propose a vertically par-
shows good performance to support data analysis and energy titioning to support convolutional neural networks (CNN) in
prediction in IIoT environments. Kanawaday and Sane [22] low-spec edge devices to support IoT applications. However,
apply a traditional machine learning method to analyze sens- the extensive data transferring in the network communications
ing data from a slitting machine and predict possible failures between the edge and the cloud still affects the processing
for further maintenance. Nguyen and Medjaher [23] design a performance.
deep learning-based framework to support dynamic predictive AI models with multiple outputs are also applied for adapt-
maintenance. A basic LSTM model is used in this framework, ing different edge devices. For low-spec devices, AI tasks are
which shows good precision in predicting the future system processed through the outputs with less computing load. In
failure time. high-spec devices, AI tasks will be processed through the
Unlike data analysis in IIoT, automated manufacturing entire model for best precision. Teerapittayanon et al. [36]
usually applies computer vision-based AI technologies to first propose the AI model with different outputs for the local
recognize the objects recorded by sensors for future decision- device, edge devices, and cloud servers. The entire network
making. AI technologies have been applied for anomaly will process input data for the cloud output and process

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
ZHU et al.: GREEN AI FOR IIoT: ENERGY EFFICIENT INTELLIGENT EDGE COMPUTING FOR IIoT 81

AI models have been stored in AI task processors to sup-


port related AI tasks. Meanwhile, the factory also builds an
access network to connect equipment for monitoring and man-
agement. The equipment has wireless connections with the
access network through access points. Access points are wired
connected to the access network. Cloud services are applied
to support other applications such as databases, management
tools, or information sharing.
The AI task processors usually have high energy consump-
tion due to the performance requirement. It is not easy to apply
traditional energy consumption optimization strategies in the
Fig. 1. A typical scenario for AI-driven IIoT applications. above scenario. Switching idle devices into a sleep state is a
key operation in most optimizations, which is not appropriate
to industrial scenarios because the state changing will affect
the input data with several neural layers for the local out- the equipment stability. Improving the utilization of comput-
put. Due to the rapid development of edge devices, most AI ing devices is another way to improve energy consumption.
models can be processed in new published hardware with- However, in the above scenario, AI task processors are iso-
out modification. NVIDIA has introduced the most powerful lated in different wireless access points, and it is slow and
A100 GPU to the edge, which can process AI tasks and unstable to transfer data flow from one equipment to another
processes in data centers [37]. Meanwhile, low-cost edge across the access network. Another problem is that dedicated
devices (e.g., NVIDIA Jetson Xavier NX [38]) can process computing devices are applied to process AI tasks from differ-
multiple AI tasks simultaneously. Therefore, in existing IIoT ent AI-enabled equipment, which means migrating tasks across
environments, the edge devices are powerful enough to pro- different equipment is not feasible. Therefore, it is necessary
cess AI tasks from IIoT devices. Due to the heterogeneous to apply a new structure to support AI-driven IIoT, which
architectures and distributed structure, efficient scheduling can apply the energy consumption optimization strategies and
strategies are among the most challenging intelligent edge satisfy the strict latency requirement.
computing issues. Q-learning-based scheduling is a promis- Furthermore, there are several opportunities in the AI-driven
ing methodology applied to optimizing resource scheduling in IIoT scenario: 1. The categories or types of tasks in the IIoT
edge environments [39]. However, it still needs an efficient scenario are much less than a standard IoT scenario. Usually,
resource scheduling strategy for intelligent edge computing the task categories in an IIoT scenario are determined by some
due to AI tasks’ particular computing model. This paper given requirements. After the IIoT system deployment, very
focuses on the dedicated scheduling strategy for intelligent few new task categories will be added to the scenario. 2. IIoT
edge computing in AI IIoT environments. devices in a given environment are knowable and manage-
able, and the factory or company also limits the scale of IIoT
III. M OTIVATION devices. 3. The network connections and power supplement in
an IIoT environment are very stable.
We present the research motivation in this section, including
the research scenario of AI-driven IIoT with intelligent edge
and the proposed framework’s design. B. Intelligent Edge Computing Framework
Intelligent edge computing is the opportunity to improve
A. AI-Driven IIoT Scenario the energy efficiency of AI-driven IIoT and meet the strict
We first introduce what is meant by an AI task in the IIoT requirement from the processing time. As shown in Fig. 2,
scenario. In a typical IIoT scenario, different machine learning we design the AI-driven IIoT framework based on intelligent
methods are deployed to process large amounts of data sens- edge computing.
ing from IoT devices. In recent years, various deep learning Edge nodes are deployed on the edge of the access network,
technologies that show advantages in processing complex data including the access points and leaf switches, which are very
types such as images, videos, and sound have been deployed near to industrial equipment. In a factory environment, it is
in IIoT environments. Deep learning is also the central part possible to organize all edge nodes with a centralized con-
of nowadays AI area. Therefore, AI tasks are considered tasks troller, which is stable and convenient to manage multiple edge
containing deep learning parts for processing sensing data in nodes. The intelligent edge controller consists of the edge node
IIoT environments. After that, we use a typical AI-driven IIoT scheduler, AI model converter, AI model storage, task man-
scenario shown in Fig. 1 to illustrate the issue in process- ager, AI processing interface, and management interface. We
ing AI tasks. In this scenario, a factory has applied AI-driven introduce these six parts and the working process between
IIoT applications in the industrial robot, heavy equipment, and them as follows. The edge node scheduler records all edge
transport vehicle for different purposes. To process AI tasks nodes’ status and information in the network, including the
from different applications in required latency, the factory workload, hardware specs, and software settings. Meanwhile,
has to place high-performance computers as AI task proces- the resource scheduling strategy is executed to find the best
sors physically connected to industrial equipment. Pre-trained edge node for processing the coming AI tasks.

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
82 IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO. 1, MARCH 2022

First, the centralized controller is deployed in a company’s


private cloud rather than in the public cloud. Replication of
the controller is also considered in the proposed framework
to provide consistent service even if one controller fails. All
edge nodes are connected to the controller with indepen-
dent network links, which reduces the influence of the access
network. To avoid the single point of failure, the controller
will assign multiple edge nodes to execute critical tasks.
The scalability of the entire intelligent edge computing
framework consists of two parts, the scalability of AI task
categories and the scalability of the edge system. It is easy to
add a new AI task category by adding the pre-trained AI model
into the AI Model Storage. For example, in our demonstration
system, converting an open neural network exchange (ONNX)
model to a TensorFlow model only cost several seconds, and
the size of most converted models is from 1 to 300 MB.
Usually, a lightweight server can store more than a million
of AI models for large amounts of AI task categories. For the
scalability of the edge system, there is no consistent operation
Fig. 2. The intelligent edge computing framework for AI-Drive IIoT. needed by adding or removing edge nodes in the central-
ized server. Moreover, since the proposed framework prepared
Edge nodes process AI tasks based on delegated pre-trained many enough AI models, the framework can deploy standard
AI models. However, most AI models are platform-related AI models to a new edge node without model conversion.
and not executed directly in edge environments without mod- Finally, based on the intelligent edge computing framework,
ification. The AI model converter automatically modifies the main motivations of our work are summarized as follows.
pre-trained AI models to the feasible models for given edge First, it is necessary to formulate the scheduling problem of
environments. Meanwhile, the AI model converter will gen- utilizing scarce computing resources in the edge for executing
erate different converted AI models for heterogeneous edge as many AI tasks as possible to minimize the energy con-
devices. Since model conversion is a time-consuming task, sumption of the entire AI-driven system. Second, we want to
it is challenging to convert AI-models online. We design an design an efficient strategy to solve the proposed problem in
AI model storage to store and manage all converted AI mod- an online scenario due to the dynamic workload of most IIoT
els for further deployment. The AI model storage will update environments.
deployed AI models in edge nodes after storing pre-trained AI
models with new versions. Therefore, a management table is
maintained by the AI model storage to record the deployed
IV. S CHEDULING P ROBLEM F ORMULATION AND
AI models in each edge node.
S OLUTION
The task manager module monitors tasks in all edge nodes,
including executing time, edge node identifications (IDs), In this section, we firstly formulate the problem of resource
task identifications (IDs), and resource consumption. The scheduling in intelligent edge computing and then describe the
edge node scheduler can schedule AI tasks according scheduling strategy’s details.
to the information from the task manager module. Meanwhile,
the task manager module controls the states of tasks and edge
nodes to switch a task or an edge node from one state to A. Problem Formulation
another. AI processing interface module provides a unified In the scheduling problem, the edge framework schedules
command set for the AI model storage. Since different edge an edge node to process the coming AI task. According to
nodes and AI platforms have different commands for process- state of the art in edge computing, the latency between edge
ing AI tasks, the AI processing interface module can transfer nodes is much larger than in the cloud environment [40], [41].
the device-related or platform-related interface to the unified Meanwhile, the energy consumption for transferring most AI
interface for upon module. Like the AI processing interface models and intermediate data between different edge devices is
module, the management interface module also provides the not affordable [42]. Meanwhile, due to most AI tasks focus on
unified interface for the upon module to control edge nodes. processing streaming data, the interruption of migration is not
After the task manager module sending the edge node ID, acceptable. Therefore, we assume there is no task migration
task ID, and operations to the management interface module, between different edge nodes in the scheduling problem. We
all operations will be transferred to the direct commands to use set E to denote all edge nodes and ei to denote an edge
the given edge node. node in set E. Let set D denote all equipment in the factory,
In the AI-driven IIoT framework, we add some risk control and dj denote one device in set D.
strategies to avoid potential failures since the industrial envi- All AI Tasks are denoted by set A, and one AI task is
ronment needs higher reliability than other IoT environments. denoted to ak . We use value Xjk to denote whether task ak

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
ZHU et al.: GREEN AI FOR IIoT: ENERGY EFFICIENT INTELLIGENT EDGE COMPUTING FOR IIoT 83

is processed for device dj , given by Therefore, in time series T, the total energy consumption
 denoted by E is given by
1, task ak is processed for device dj ,
Xjk = (1) |T |
0, elsewise. 
E = P · tk . (9)
Usually, the power consumption in processing a given AI t=1
task depends on the hardware of edge devices. Therefore, the
energy consumption of task ak in edge node ei is denoted by Therefore, the scheduling problem is summarized as
pik . We also use a value Yik to denote whether edge node ei
minimize: E
is scheduled to process task ak , given by
s.t.: lt ≤ Lt (10)
 |A|
1, task ak is processed in edge node ei , k =1 Yik · bk ≤ ci .
Yik = (2)
0, elsewise.
Resource scheduling problem for AI-driven IIoT in intel-
We use P to denote the power consumption of the schedul- ligent edge: given a set of intelligent edge nodes and a set
ing strategy is given by of industrial equipment, the edge node scheduler attempts
to assign edge nodes for processing AI tasks submitted by
|E | |A|
  equipment with minimal energy consumption.
P = Yik · pik . (3)
i=1 k =1
B. Scheduling Strategy
The processing latency of task ak on edge node ei is
p We first solve the scheduling problem in a static scenario
denoted by lik , and the network latency from equipment dj
in which AI tasks are submitted at the beginning and contin-
to edge node ei is denoted by lijn . Considering the communi-
d denote the time for uously processed until the end of the scheduling. The object
cation limitation of equipment ei , let lik of the scheduling problem is simplified to
transferring data of task ak to edge node ei , given by
|A| minimize: P
|E |
 lt ≤ Lt
lkd = Yik · k =1 Yik D subject: (11)
(4) |A|
k =1 Yik ≤ ci .
Bk
i=1

where D is the transferring data size and Bk is the ingress Theorem 1: The scheduling problem that assigning intel-
bandwidth of edge node ei . ligent edge nodes for processing AI tasks with the static
The total latency for processing task ak is denoted by lkt , assumption is NP-hard.
given by Proof: The NP-hardness of the scheduling problem can be
proved by reducing the bin-packing problem.
|E | |E | |D| Bin-packing Problem: given a set of items a1 , a2 , .., a|A| ,
 p
 
lkt = li.e., · Yik + lijn · Xjk · Yik + lkd . (5) each item ak with a size bk , and a number of |E | bins of
i=1 i=1 j =1 size C, the bin-packing problem accommodate all items into
We use Lt to denote the required processing latency in the |E | bins. We first suppose that there is a solution for the bin-
scheduling and for a given task ak , lkt is no more than Lt . packing problem. This solution accommodates |A| items to
Because edge node ei has a capacity for processing tasks, |E | bins, which is also a solution to the edge node scheduling
the resource scheduling strategy should satisfy problem. We suppose that the edge node scheduling problem
as a solution that each AI task ak is assigned to an edge
|A| node ei , which can solve the bin-packing problem. Since we

Yik · bk ≤ ci (6) can evaluate the objective function associated with a given
k =1 solution in a polynomial time, the edge node scheduling is
NP-hard.
where ci is the capacity of edge node ei , and bk is the overload
To solve the edge node scheduling problem, we design a
of task ak .
heuristic algorithm shown in Algorithm 1. We assume that
We consider that the scheduling strategy works on a time
edge nodes are less than equipment devices, i.e., |E | < |D|.
series T, which is divided into time slots. Let t denote a time
There are two loops in the static scheduling algorithm, a dou-
slot in time series T. We assume the length of each time slot
ble loop, and a normal loop. The complexity of the double
is the minimal scheduling unit. Therefore, the working time
loop is O(|A||D|) because the first layer of the loop has |A|
series of task ak , denoted by tk , is given by
iterations, and the second layer has no more than |D| itera-
 
tions. The normal loop has |E | iterations. Therefore, the time
tk = z1 , z2 , . . . , z|T | (7)
complexity of the proposed algorithm is O(|A||D| + |E |) =
and zt is a value to denote whether task tk is working, given by O(|A||D|) because |E | is smaller than |A||D|. Moreover, from
 the analysis in a previous paper [43], the number of edge nodes
1, task ak is working in time t employed by the static scheduling algorithm is no more than
zt = (8)
0, elsewise. two times that in an optimal solution.

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
84 IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO. 1, MARCH 2022

Algorithm 1 The Static Edge Node Scheduling Strategy for


AI-Driven Tasks
Sort A = {a1 , a2 , ..., a|K | } that b1 ≤ b2 ≤ ... ≤ b|A| ;
for ak ∈ A do
d ← 0;
for dj ∈ D do
if Xj k == 1 then
d ← dj ;
Break;
end if
end for
E  ← ∅;
Fig. 3. An example of input states, including the edge node scheduling and
for ei ∈ E do three tasks.
if lt < Lt andci ≥ bk then
E  ← E  ei
end if The Q ∗ (·) function can be estimated iteratively with the
end for Bellman equation as

Sort E  = {e1 , e2 , ..., e|E  | } that p1k ≤ p2k , ..., ≤
p|E  |k Qi+1 (Zn , On ) = EZn+1 rn + γ · max [Qi (Zn , On )]|Zn , On
On+1
Assign e1 for processing ak ; (15)
c1 ← c1 − bk ;
end for where i is the iterator. When i ← ∞, the optimal is equal Q∗
to Qi .
We use θ to denote the sequence of approximation weights,
and the Q ∗ (Zn , On ) can be estimated by
Since the static scheduling problem is a particular case of
the online scheduling problem, the online scenario’s schedul- Q ∗ (Zn , On ) ≈ Q(Zn , On ; θ). (16)
ing problem is NP-hard. Therefore, we define a reinforcement In our scheduling strategy, we adopt the deep reinforcement
learning-based strategy to find the solution of the online learning method and let Li (θi ) denote the loss function,
edge node scheduling problem for AI-driven IIoT. We set the given by
negative value of energy consumption as the reward of the

reinforcement learning and use rn to denote the single-step Li (θi ) = Eρ(Zn ,On ) EZn+1 [rn
reward at time n, given by
+ γ · Q(Zn+1 , On+1 ; θi−1 )|Zn , On ]
2
|E | |A|
  − Q(Zn , On ; θi ) (17)
rn = − Yik · pik · zt . (12)
i=1 k =1
where ρ(Z , O) denote the behavior distribution of Z and O.
We apply the stochastic gradient descent (SGD) to
Let Rn denote the total future discounted reward that is optimize (17) with the gradient given by
defined as
∇θi Li (θi ) = Eρ(Zn ,On ) [(rn + γ · Q(Zn+1 , On+1 ; θi−1 )

|T | − Q(Zn , On ; θi )) · ∇θi · Q(Zn , On ; θi ) (18)

Rn = γ t−n · rn (13)
We build the scheduling strategy states as distinct images
t=n
shown in Fig. 3, including the scheduled edge nodes and tasks.
where γ < 1 is the discounted factor. The image of the scheduled edge nodes is a table of edge
We use set O = {O1 , O2 , . . . , O|T | } and Z = capacities, in which each column is one unit of the process-
{Z1 , Z2 , . . . , Z|T | } to denote actions and states of the ing resource and each row means the time slot. Colored cells
scheduling strategy, respectively. We also use set O = represent allocated resources, and different color means dif-
{O1 , O2 , . . . , O|T | } and Z = {Z1 , Z2 , . . . , Z|T | } to denote ferent task. The image of a task is also a table in which
entire admissible action and state space. Let Q ∗ (·) denote the width is the needed number of resource units, and the
the action-reward function with maximum reward in state height represents the time series. The action space is given by
Zn and an optimal action Zn . From the Bellman optimality {e1a , e2a , . . . , e|A|
a } where e a means the edge node assigned to
k
equation [44], the value of Q ∗ (Zn , On ) can be formulated by task ak and Xeka k = 1. If the task ak ends or does not begin,
 the value of eka is zero. If the action for one task is valid, the
Q ∗ (Zn , On ) = EZn+1 rn + γ · max [Q ∗ (Zn , On )]|Zn , On value of eka is ∅. In each time slot, the scheduler will find a
On+1 valid action or an invalid action. After the scheduler finds an
(14) action, all tasks will put into the table for edge node schedul-
ing with different colors. If eka = ∅ exists, the scheduler will
where E is the expectation. shift up by a one-time slot and find another action space.

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
ZHU et al.: GREEN AI FOR IIoT: ENERGY EFFICIENT INTELLIGENT EDGE COMPUTING FOR IIoT 85

Algorithm 2 Training Algorithm for the Scheduler


Initial replay buffer
for m form 1 to M do
Z0 is set as an empty state;
for n from 1 to |T | do
p ← random(0, 1)
if p ≤ 1 −  then
On ← arg maxOn Q(Zn , On ; θn );
else
Fig. 4. Power consumption of four edge devices in the testbed.
Randomly select On from On ;
end if
Calculate rn from (12);
Assign edge nodes with On and get Zn+1 ;
Store (Zn , On , rn , Zn+1 ) in the replay buffer;
Sample batch B from the replay buffer;
for n  from 1 to |B | do
Update θn by performing SGD;
end for
θn−1 ← θn ;
end for
end for

We also define the training algorithm shown in Algorithm 2


to train the edge node scheduler. In the training phase, we use
an -greedy strategy to generate action On in each iteration.
Also, we build a replay buffer for the mini-batch sampling to
accelerate the SGD procedure.

Fig. 5. Time of loading AI models and processing tasks.


V. P ERFORMANCE E VALUATION
The performance evaluation consists of the demonstration
system in a small testbed and large-scale simulations. The results of the processing latency in the demonstration
system are given by Fig. 5. Before processing AI tasks in edge
devices, it is necessary to load the pre-trained and converted
A. Demonstration System models to the edge nodes. From the results shown in Fig. 5(a),
We first build a demonstration system to evaluate the because of its highest performance, Jeston AGX Xavier loads
latency and energy consumption of AI-driven applications. The the AI models in the shortest time. Meanwhile, loading AI
demonstration system is developed in a small testbed consists models for GPU computing is slower than models for CPU
of devices shown in Table I. All devices are connected to a computing. Loading converted models is slower than loading
gigabit Ethernet Network Switch with wired connections. models for general GPU computing. The Raspberry Pi 4B is
We test the power consumption and average processing time only able to process AI tasks with the CPU, but its model
of three deep learning models, YOLOv3, YOLOv3-tiny, and loading time is longer than Jeston AGX Xavier.
VGG16. In each test, the edge devices process selected 50 As shown in Fig. 5(b), the average processing time in edge
images from COCO 2017 image dataset, and we record the nodes is different from model loading time. As the quick-
model loading time and processing time. We also converted est device, Jeston AGX Xavier can finish the task in about
the models to TensorRT-enabled models with the AI model 50 ms with the converted YOLOv3-tiny model. The process-
converter. The intelligent edge framework is developed in a ing time with the converted YOLOv3 model in Jeston AGX
Macbook Pro 15 2018 computer with Python 3.9.2. The AI Xavier is about 280 ms. Without model converting, the pro-
model converter is based on the TensorRT 7.1.3 library, and cessing time with the original YOLOv3 in Jeston AGX Xavier
all GPU-based computing is based on CUDA 10.2. is more than 5800 ms which is more than 20 times longer than
As shown in Fig. 4, we test the power consumption in pro- the converted model. Although the Raspberry Pi 4 has the
cessing AI tasks and idle state of four edge devices. Since the lowest power consumption, its processing time with the orig-
Jetson AGX Xavier has the highest performance, its power inal YOLOv3 is more than 3640 times than Jetson Nano with
consumption is more than 30 watts. Jetson Xavier NX and converted YOLOv3 model. Even with YOLOv3-tiny, the pro-
Jeston Nano’s power consumption is similar, while the former cessing time in Raspberry Pi 4 is 160 times longer than Jetson
is more powerful than the latter. The weakest Raspberry Pi 4 Nano with converted YOLOv3-tiny. The average processing
also has the lowest power consumption. time of VGG16 is very similar with the time of YOLOv3 in

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
86 IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO. 1, MARCH 2022

TABLE I
H ARDWARE S ETTING OF THE T ESTBED

TABLE II
PARAMETER S ETTINGS IN A LL S IMULATIONS We compare the results between the deep reinforcement
learning-based online scheduling strategy (Online), the static
scheduling algorithm (Static), and the FIFO strategy in all
simulations. All simulations are executed 20 times, and the
average values are recorded with a confidence level of 95%.
All results of the performance evaluation are represented in
Fig. 6. As shown in Fig. 6(a), we change the number of tasks
from 200 to 500 and test the total energy consumption in one
hour. The online scheduling strategy shows the lowest energy
consumption with the different number of tasks. When the
number of tasks is less than 300, the FIFO method performs
better than the results with the static scheduling algorithm.
When the number of tasks increases to more than 300, the
all devices. The model converting can improve then energy static scheduling algorithm brings less energy consumption
efficiency in both high-spec and low-spec edge devices. GPU than the FIFO method.
accelerated AI processing is still essential to process AI tasks We also test the energy consumption with different numbers
with less energy consumption. of edge nodes. The number of tasks is set to 400, and the num-
ber of edge nodes increases from 100 to 400. From Fig. 6(b),
when the number of edge nodes is less than 200, three meth-
B. Experimental Result Analysis ods perform similarly, and the total energy consumption is
In a large environment, we apply simulations for the more than 5200 Wh. After more edge nodes begin to process
performance evaluation of the proposed scheduling strategy. AI tasks, the total energy consumption decreases. The online
We develop a simulator with Python 3.9.2. The training and scheduling method still has the lowest energy consumption in
evaluation is executed in a server that equips an Intel Core i9 the three methods, and the other two methods perform very
10900X CPU, 128GB memory, a 1TB SSD and two NVIDIA similarly.
GeForce TITAN RTX graphics cards. The operating system is The active ratio of the task in the time series also affects
Ubuntu 18.04 LTS. We adopt Keras 2.4.0 as the deep leaning the total energy consumption. In Fig. 6(c), we adjust the active
platform. In the experiments, we apply the RMSProp opti- ratio from 0.5 to 1.0 and test the total energy consumption. The
mizer for Keras library. We set the size of mini-batches to 64. number of edge nodes is set to 400, and the number of total
For the -greedy in the scheduler training, we set the value of tasks is set to 800. The results show that a higher active ratio
 to 0.1. will bring more energy consumption, and the online scheduling
As shown in Table II, we set the parameters of all sim- strategy performs better than the other two algorithms.
ulations according to the experimental results. The tasks are Finally, we test the energy consumption with different
randomly generated in the time series, and the power consump- required latency. We set the number of tasks to 400 with 400
tion for processing tasks is evenly distributed in [5, 25] from edge nodes, and the required latency increases from 100 ms to
Fig. 4. The required processing latency is evenly distributed 600 ms. As shown in Fig. 6(d), the total energy consumption
in [2, 600] ms from Fig. 5. According to the general industrial decreases with increased required latency. When the required
environment, [45], the network latency between devices and latency is less than 300 ms, the online scheduling strategy
edge nodes is evenly distributed in [1, 200] ms. The capacity performs much better than the other two methods. When the
of edge nodes is based on the memory size of each device and latency requirement becomes loose, three scheduling methods
evenly distributed in [1, 16]. According to memory usage, the perform similarly, and the FIFO has less energy consumption
required resource units of tasks are evenly distributed in [1, 2]. than the Static algorithm.
We use a small factory with 500 employees as the simulation
environment. Therefore, the number of edge nodes and equip-
ment devices is set from 100 to 400 and from 200 to 500, VI. C ONCLUSION AND F UTURE W ORK
respectively. Usually, since one device will send at most two From this article, intelligent edge computing shows its
tasks, the number of tasks is set from 200 to 800. We test the potential in the energy-efficient AI-driven IIoT scenario.
energy consumption in one hour with 3600 time-slots and 1 Depending on the development of AI frameworks, it is possi-
second per slot. The ratio that tasks are active in the total time ble to concentrate heterogeneous resources on edge to support
series is evenly distributed [0.8, 1]. different AI tasks. Meanwhile, novel resource scheduling is

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
ZHU et al.: GREEN AI FOR IIoT: ENERGY EFFICIENT INTELLIGENT EDGE COMPUTING FOR IIoT 87

Fig. 6. Energy consumption with different settings in intelligent edge computing.

also important to the energy efficiency in the IIoT environ- [6] Y. B. Zikria, M. K. Afzal, S. W. Kim, A. Marin, and M. Guizani, “Deep
ments. The demonstration system developed the small testbed learning for intelligent IoT: Opportunities, challenges and solutions,”
Comput. Commun., vol. 164, pp. 50–53, Dec. 2020.
shows that a well-designed edge system can outperform tra- [7] Y. Wang et al., “Benchmarking the performance and energy efficiency
ditional cloud-based solutions with short processing latency of AI accelerators for AI training,” in Proc. 20th IEEE/ACM Int. Symp.
and very low energy consumption. Finally, the proposed work Clust. Cloud Internet Comput. (CCGRID), 2020, pp. 744–751.
is a feasible solution to the AI-driven IIoT applications from [8] NVIDIA A100 GPUs Power the Modern Data Center. Accessed:
Mar. 10, 2021. [Online]. Available: https://www.nvidia.com/en-us/data-
the performance evaluation results. With the online schedul- center/a100/
ing strategy, energy consumption in different settings is less [9] H. Li, K. Ota, M. Dong, and M. Guo, “Learning human activities
than 20% and 30% of the number with the static schedul- through Wi-Fi channel state information with multiple access points,”
IEEE Commun. Mag., vol. 56, no. 5, pp. 124–129, May 2018.
ing strategy and FIFO algorithm. In the future, we will try [10] M. Mohammadi, A. Al-Fuqaha, S. Sorour, and M. Guizani, “Deep learn-
to enlarge the scale of our testbed and develop more deep ing for IoT big data and streaming analytics: A survey,” IEEE Commun.
learning applications such as NLP, time series forecasting, and Surveys Tuts., vol. 20, no. 4, pp. 2923–2960, 4th Quart., 2018.
audio recognition. We also plan to build intelligent edge ser- [11] M. Dong, K. Ota, and A. Liu, “RMER: Reliable and energy-efficient
data collection for large-scale wireless sensor networks,” IEEE Internet
vice in robots and UAVs to support autonomous driving and Things J., vol. 3, no. 4, pp. 511–519, Aug. 2016.
flying. [12] J. Xu, K. Ota, and M. Dong, “Saving energy on the edge: In-memory
caching for multi-tier heterogeneous networks,” IEEE Commun. Mag.,
vol. 56, no. 5, pp. 102–107, May 2018.
R EFERENCES [13] H. Li, K. Ota, and M. Dong, “Deep reinforcement scheduling for mobile
crowdsensing in fog computing,” ACM Trans. Internet Technol., vol. 19,
[1] S. G. Tzafestas, “Synergy of IoT and AI in modern society: The robotics
no. 2, pp. 1–18, 2019.
and automation case,” Robot. Autom. Eng. J., vol. 3, no. 5, pp. 1–15,
2018. [14] J. Xu, K. Ota, and M. Dong, “Big data on the fly: UAV-mounted mobile
[2] W. Sun, J. Liu, and Y. Yue, “AI-enhanced offloading in edge comput- edge computing for disaster management,” IEEE Trans. Netw. Sci. Eng.,
ing: When machine learning meets Industrial IoT,” IEEE Netw., vol. 33, vol. 7, no. 4, pp. 2620–2630, Oct.–Dec. 2020.
no. 5, pp. 68–74, Sep./Oct. 2019. [15] S. Pasteris, S. Wang, M. Herbster, and T. He, “Service placement with
[3] M. Haslgrübler, P. Fritz, B. Gollan, and A. Ferscha, “Getting through,” provable guarantees in heterogeneous edge computing systems,” in Proc.
in Proc. 7th Int. Conf. Internet Things, 2017, pp. 1–8. IEEE INFOCOM Conf. Comput. Commun., 2019, pp. 514–522.
[4] Y. Cong, D. Tian, Y. Feng, B. Fan, and H. Yu, “Speedup 3-D texture-less [16] R. Han, S. Li, X. Wang, C. H. Liu, G. Xin, and L. Y. Chen, “Accelerating
object recognition against self-occlusion for intelligent manufacturing,” gossip-based deep learning in heterogeneous edge computing platforms,”
IEEE Trans. Cybern., vol. 49, no. 11, pp. 3887–3897, Nov. 2019. IEEE Trans. Parallel Distrib. Syst., vol. 32, no. 7, pp. 1591–1602,
[5] J. Hiller, M. Henze, M. Serror, E. Wagner, J. N. Richter, and K. Wehrle, Jul. 2021.
“Secure low latency communication for constrained Industrial IoT sce- [17] H. Yan, Y. Zhang, Z. Pang, and L. D. Xu, “Superframe planning and
narios,” in Proc. IEEE 43rd Conf. Local Comput. Netw. (LCN), 2018, access latency of slotted MAC for industrial WSN in IoT environment,”
pp. 614–622. IEEE Trans. Ind. Informat., vol. 10, no. 2, pp. 1242–1251, May 2014.

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.
88 IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, VOL. 6, NO. 1, MARCH 2022

[18] H. Li, K. Ota, M. Dong, and H.-H. Chen, “Efficient energy trans- [41] N. Tziritas et al., “Data replication and virtual machine migrations to
port in 60 GHz for wireless industrial sensor networks,” IEEE Wireless mitigate network overhead in edge computing systems,” IEEE Trans.
Commun., vol. 24, no. 5, pp. 143–149, Oct. 2017. Sustain. Comput., vol. 2, no. 4, pp. 320–332, Oct.–Dec. 2017.
[19] K. Li, X. Tang, and K. Li, “Energy-efficient stochastic task scheduling [42] H. Liu, L. T. Yang, M. Lin, D. Yin, and Y. Guo, “A tensor-based holistic
on heterogeneous computing systems,” IEEE Trans. Parallel Distrib. edge computing optimization framework for Internet of Things,” IEEE
Syst., vol. 25, no. 11, pp. 2867–2876, Nov. 2014. Netw., vol. 32, no. 1, pp. 88–95, Jan./Feb. 2018.
[20] I. Bisio, C. Garibotto, A. Grattarola, F. Lavagetto, and A. Sciarrone, [43] H. Li, P. Li, S. Guo, and A. Nayak, “Byzantine-resilient secure software-
“Exploiting context-aware capabilities over the Internet of Things for defined networks with multiple controllers in cloud,” IEEE Trans. Cloud
industry 4.0 applications,” IEEE Netw., vol. 32, no. 3, pp. 101–107, Comput., vol. 2, no. 4, pp. 436–447, Oct.–Dec. 2014.
May/Jun. 2018. [44] R. Luus, “Dynamic programming,” in Iterative Dynamic Programming.
[21] T. Han, K. Muhammad, T. Hussain, J. Lloret, and S. W. Baik, “An Boca Raton, FL, USA: Chapman Hall/CRC, 2019, pp. 67–80.
efficient deep learning framework for intelligent energy management in [45] P. Ferrari, A. Flammini, E. Sisinni, S. Rinaldi, D. Brandão, and
IoT networks,” IEEE Internet Things J., vol. 8, no. 5, pp. 3170–3179, M. S. Rocha, “Delay estimation of Industrial IoT applications based
Mar. 2021. on messaging protocols,” IEEE Trans. Instrum. Meas., vol. 67, no. 9,
[22] A. Kanawaday and A. Sane, “Machine learning for predictive mainte- pp. 2188–2199, Sep. 2018.
nance of industrial machines using IoT sensor data,” in Proc. 8th IEEE
Int. Conf. Softw. Eng. Service Sci. (ICSESS), 2017, pp. 87–90.
[23] K. T. P. Nguyen and K. Medjaher, “A new dynamic predictive mainte-
nance framework using deep learning for failure prognostics,” Reliabil.
Eng. Syst. Safety, vol. 188, pp. 251–262, Aug. 2019.
[24] S. M. Erfani, S. Rajasegarar, S. Karunasekera, and C. Leckie, “High- Sha Zhu (Student Member, IEEE) received
dimensional and large-scale anomaly detection using a linear one-class the B.S. degree in electrical engineering from
SVM with deep learning,” Pattern Recognit., vol. 58, pp. 121–134, Honam University, South Korea, in 2011, and
Oct. 2016. the M.S. degree in electrical engineering from
[25] T. S. Buda, B. Caglayan, and H. Assem, “DeepAD: A generic Chonnam National University, South Korea, in 2013.
framework based on deep learning for time series anomaly detec- She is currently pursuing the Ph.D. degree in
tion,” in Advances in Knowledge Discovery and Data Mining. Cham, advanced information and electronic engineering
Switzerland: Springer, 2018, pp. 577–588. with Muroran Institute of Technology, Japan. Her
[26] D. F. Wulsin, J. R. Gupta, R. Mani, J. A. Blanco, and B. Litt, “Modeling research interests include smart grid, IoT, and edge
electroencephalography waveforms with semi-supervised deep belief computing.
nets: Fast classification and anomaly measurement,” J. Neural Eng.,
vol. 8, no. 3, 2011, Art. no. 036015.
[27] T. Schlegl, P. Seeböck, S. M. Waldstein, U. Schmidt-Erfurth, and
G. Langs, Unsupervised Anomaly Detection With Generative Adversarial
Networks to Guide Marker Discovery (Lecture Notes in Computer Kaoru Ota (Member, IEEE) was born in Aizu–
Science). Cham, Switzerland: Springer, 2017, pp. 146–157. Wakamatsu, Japan. She received the B.S. degree
[28] Y. Liu et al., “Deep anomaly detection for time-series data in Industrial in computer science and engineering from The
IoT: A communication-efficient on-device federated learning approach,” University of Aizu, Japan, in 2006, the M.S.
IEEE Internet Things J., vol. 8, no. 8, pp. 6348–6358, Apr. 2021. degree in computer science from Oklahoma State
[29] L. Antwarg, R. M. Miller, B. Shapira, and L. Rokach, “Explaining University, USA, in 2008, and the Ph.D. degree
anomalies detected by autoencoders using SHAP,” 2020. [Online]. in computer science and engineering from The
Available: arXiv:1903.02407. University of Aizu in 2012. She is currently an
[30] X. Xie, X. Liu, H. Qi, B. Xiao, K. Li, and J. Wu, “Geographical Associate Professor and Ministry of Education,
correlation-based data collection for sensor-augmented RFID systems,” Culture, Sports, Science and Technology Excellent
IEEE Trans. Mobile Comput., vol. 19, no. 10, pp. 2344–2357, Oct. 2020. Young Researcher with the Department of Sciences
[31] X. Xie et al., “Implementation of differential tag sampling for and Informatics, Muroran Institute of Technology, Japan. From March 2010
COTS RFID systems,” IEEE Trans. Mobile Comput., vol. 19, no. 8, to March 2011, she was a Visiting Scholar with the University of Waterloo,
pp. 1848–1861, Aug. 2020. Canada. She was also a Japan Society of the Promotion of Science Research
[32] S. Huang, Y. Guo, D. Liu, S. Zha, and W. Fang, “A two-stage trans- Fellow with Tohoku University, Japan, from April 2012 to April 2013. She is a
fer learning-based deep learning approach for production progress recipient of the IEEE TCSC Early Career Award 2017, the 13th IEEE ComSoc
prediction in IoT-enabled manufacturing,” IEEE Internet Things J., Asia–Pacific Young Researcher Award 2018, and the 2020 N2Women: Rising
vol. 6, no. 6, pp. 10627–10638, Dec. 2019. Stars in Computer Networking and Communications. She is a Clarivate
[33] C. Zhang, Q. Cao, H. Jiang, W. Zhang, J. Li, and J. Yao, “FFS-VA: A Analytics 2019 Highly Cited Researcher (Web of Science).
fast filtering system for large-scale video analytics,” in Proc. 47th Int.
Conf. Parallel Processing, 2018, pp. 1–10.
[34] S. Bhattacharya and N. D. Lane, “Sparsification and separation of deep
learning layers for constrained resource inference on wearables,” in
Proc. 14th ACM Conf. Embedded Netw. Sens. Syst. CD-ROM, 2016,
pp. 176–189. Mianxiong Dong (Member, IEEE) received the
[35] H. Li, K. Ota, and M. Dong, “Learning IoT in edge: Deep learning for B.S., M.S., and Ph.D. degrees in computer sci-
the Internet of Things with edge computing,” IEEE Netw., vol. 32, no. 1, ence and engineering from the University of Aizu,
pp. 96–101, Jan./Feb. 2018. Japan. He is the youngest ever Vice President and a
[36] S. Teerapittayanon, B. McDanel, and H. T. Kung, “Distributed deep Professor of Muroran Institute of Technology, Japan.
neural networks over the cloud, the edge and end devices,” in Proc. IEEE He was a JSPS Research Fellow with the School of
37th Int. Conf. Distrib. Comput. Syst. (ICDCS), 2017, pp. 328–339. Computer Science and Engineering, The University
[37] NVIDIA EGX Platform for AI Computing. Accessed: Mar. 10, of Aizu and was a Visiting Scholar with BBCR
2021. [Online]. Available: https://www.nvidia.com/en-us/data- Group, University of Waterloo, Canada, supported
center/products/egx/ by JSPS Excellent Young Researcher Overseas Visit
[38] NVIDIA Jetson Xavier NX for Embedded & Edge Systems. Accessed: Program from April 2010 to August 2011. He was
Mar. 10, 2021. [Online]. Available: https://www.nvidia.com/en- selected as a Foreigner Research Fellow (a total of three recipients all over
us/autonomous-machines/embedded-systems/jetson-xavier-nx/ Japan) by NEC C&C Foundation in 2011. He is a recipient of the IEEE TCSC
[39] H. van Hasselt, A. Guez, and D. Silver, “Deep reinforcement learning Early Career Award 2016, the IEEE SCSTC Outstanding Young Researcher
with double Q-learning,” in Proc. 30th AAAI Conf. Artif. Intell., 2016, Award 2017, the 12th IEEE ComSoc Asia–Pacific Young Researcher Award
pp. 2094–2100. 2017, the Funai Research Award 2018, and the NISTEP Researcher 2018
[40] H. Liu et al., “Thermal-aware and DVFS-enabled big data task schedul- (one of only 11 people in Japan) in recognition of significant contributions
ing for data centers,” IEEE Trans. Big Data, vol. 4, no. 2, pp. 177–190, in science and technology. He is a Clarivate Analytics 2019 Highly Cited
Jun. 2018. Researcher (Web of Science).

Authorized licensed use limited to: International Institute of Information Technology Bhubaneswar. Downloaded on January 30,2023 at 14:49:06 UTC from IEEE Xplore. Restrictions apply.

You might also like