2023 Wang
2023 Wang
Abstract—We propose an on-demand computing II. NETWORK MODEL AND PROBLEM STATEMENT
provisioning method in computing power network with mixed
CPU and GPU. Simulation results validate that more computing A. CPN with CPU and GPU
tasks can be accommodated by optimizing the usage of Due to the dynamic nature of computing tasks, the traffic
heterogeneous computing resources. has the feature of high burstiness and high throughput in CPN.
Optical networks have become the key infrastructure in large-
Keywords—computing power network, CPU and GPU,
scale computing networks, taking into account the tremendous
computing resource provisioning.
bandwidth in fibers. The elastic optical network (EON), with
I. INTRODUCTION its flexible bandwidth allocation, has been developed to
efficiently provision services with diverse bandwidth
Computing power network (CPN) is a novel information requirements[5]. Therefore, EON has become a promising
infrastructure that enables on-demand allocation and flexible architecture in computing power networks.
scheduling of computing resources, storage resources, and
network resources between the cloud, edge, and end [1]. Transmission node Unavailable computing resource CPU CPU Resource
Heterogeneous computing resources, such as CPUs and GPUs, Computing node Unavailable FS Available FS GPU GPU Resource
are coordinated and scheduled in CPNs. The CPN with CPU
and GPU can better satisfy the computing resource demands CPU GPU
Authorized licensed use limited to: UNIVERSIDADE FEDERAL DE CAMPINA GRANDE. Downloaded on October 21,2024 at 15:07:38 UTC from IEEE Xplore. Restrictions apply.
c is computing workload, and is latency of QoS. In this of computing resource (KSP-PCR) algorithm as benchmark
illustration, we assume that request as = { , , 2,24,10}, algorithm.
and the spectrum resource from node E to computing nodes A OPCR algorithm
and D meets the requirements for request transmission. Fig. Input: Give link spectrum resource and node types
2(b) present three computing resources provisioning options
G(V, E), request = ( , , , , ) ∈ {1,2, … }, as
for the general computing tasks at computing nodes A and
D. In Fig. 2(b), the requires 4 CUs when using only CPU well as CPU and GPU resource capacity 99A<, 9BA< .
resources in node A, but 6 CUs in node D due to a longer Output: the selected computing node C9 ,C ∈ D and
transmission time. Moreover, computing tasks requires 2
CUs when using GPU resources in node A. The CUs required provisioning of computing resource for r.
of computing tasks are calculated in (1). The resource 1: for each in
provisioning solutions for GPU-intensive computing tasks are 2: search for transmission paths set A{1,2, … E}
also similar. 3: if A is empty
Transmission node Unavailable CUs CPU CPU Resource CPU Unit
4: r=(s, d , ,c , ) is blocked; break
Computing node Unavailable FS GPU GPU Resource GPU Unit
5. else if A is not empty
Computing resource
calculate CUs required of in each C9 ;
CPU GPU
A CPU/GPU 6:
Node N CPU N GPU Capacity/a CPU(GPU)
A B A
E
3
4
3
0
5 Unit
5 Unit
7: then exclude C9 in D9 with insufficient CUs ;
bandwidth of each FS, which is 6.25 Gbps, D is distance from Fig. 3. NSFNET topology
source node to computing node, and is transmission speed As shown in Fig.3, NSFNET is used as the simulation
which is 5 / . The aforementioned three resource topology. We consider computing tasks in the form of data
provisioning options are all available for request deployment. flow, and their parameters value are set in one unit of time.
If we deploy computing tasks using a computing resource The request arrival and duration time follow Poisson and
allocation based on the K-shortest path (KSP) algorithm while exponential distributions, respectively. Specific parameters
disregarding the RIC of computing tasks, resources may be are shown in TABLE I. 95 6 and 9G 6 represent the capacity
wasted due to inefficient utilization. RIC is calculated using of CPU and GPU resource, where (10,15CUs) refers to 10
(2), where is transmission cost, is computing resource CPUs, with each CPU having 15 CUs in computing node.
cost, T is type of computing resource, is duration time,
and ) is cost of using inactive CPU or GPU resources. σ is 1 TABLE I. PARAMETER
when the CPU or GPU is active. Therefore, when deploying a Parameter Value
computing task, computing resources should be provided on- d [1,8] × 10H bit
demand to reduce resource waste and deployment costs. [2,5] FS
, 4 c [4,8] × 10I FLOPS
* =+ +2× 3+ (1) [10,20] ms
-×/×0 567
FS 80
89 = (2 × ; × + 9<= × × )) × (2)
99A< 3TFLOPs(10,15CUs)
III. ALGORITHM 99A< 6TFLOPs(10,15CUs)
Authorized licensed use limited to: UNIVERSIDADE FEDERAL DE CAMPINA GRANDE. Downloaded on October 21,2024 at 15:07:38 UTC from IEEE Xplore. Restrictions apply.
B. Results analysis The results of spectrum utilization ratio and blocking ratio
In this section, we evaluate the performance of KSP-PCR are shown in Fig. 5. In Fig. 5(a), we can see that OPCR
algorithm and our proposed OPCR algorithm in terms of CPU consumes 10.8% more spectrum resources than KSP-PCR
utilization, GPU utilization, spectrum utilization, blocking algorithm under the same traffic load. This is because KSP-
ratio and the sum of RIC in 350 Erlang. PCR has the shortest average path length, while the OPCR
0.8 algorithm may select a longer path to minimize the RIC. The
OPCR second factor influencing spectrum utilization is the blocking
KSP-PCR 67.3%
12.7% ratio. As shown in Fig. 5(b), KSP-PCR has a higher blocking
CPU Utilization
0.6
54.6% ratio, 3.6% higher than OPCR with 600 Erlangs.
7x104
0.4 OPCR
4x104
0.0 3x104 50.53%
(a)
0 200 400 600 800 2x104
Task ID 1x104
1.0 OPCR 0
KSP-PCR 93.2%
GPU Utilization
KSP-PCR 10.8%
0.6 61.8% and NSFC (61831003, 62021005, 62101063).
REFERENCES
0.4
[1] X. Tang et al., "Computing power network: The architecture of
convergence of computing and networking towards 6G requirement,"
0.2 China Communications, vol. 18, pp. 175-185, Feb. 2021
[2] D Du, Q Liu, X Jiang, Y Xia, B Zang, and H Chen. “Serverless
0.0 computing on heterogeneous computers”. In Proceedings of the 27th
( a) ACM International Conference on Architectural Support for
0 200 400 600 800 Programming Languages and Operating Systems (ASPLOS '22).
Task ID Association for Computing Machinery, New York, NY, USA, 797–
813. 2022.
0.08 OPCR
KSP-PCR [3] L. Liu, J. Feng, X. Mu, Q. Pei, D. Lan and M. Xiao, "Asynchronous
Deep Reinforcement Learning for Collaborative Task Computing and
Blocking Ratio
Authorized licensed use limited to: UNIVERSIDADE FEDERAL DE CAMPINA GRANDE. Downloaded on October 21,2024 at 15:07:38 UTC from IEEE Xplore. Restrictions apply.