Optimized Resource Allocation For Scalable Edge-Cloud Networks Using Hybrid Computing Models
Optimized Resource Allocation For Scalable Edge-Cloud Networks Using Hybrid Computing Models
Abstract: The increasing demand for real-time data Offloading, Context-Aware Systems, Predictive
processing in applications such as autonomous Analytics, Token-Based Resource Management,
systems, smart cities, and healthcare monitoring Latency Optimization, Throughput Enhancement,
has exposed the limitations of conventional edge- SLA Compliance, Intelligent Orchestration, Edge
only or cloud-only architectures. This paper Intelligence, Distributed Systems, Real-Time
presents a novel hybrid computing model that Processing, Scalable Networks.
enables intelligent, optimized resource allocation
between edge and cloud environments, targeting I. Introduction
scalability, reduced latency, and improved resource
utilization. The core innovation is the introduction The exponential growth of smart devices, Internet of
of a Context-Aware Predictive Resource Allocation Things (IoT) systems, and real-time applications such
Engine (CAPRAE), which employs a lightweight, as autonomous driving, remote healthcare, and
adaptive learning mechanism to forecast workload augmented reality has dramatically increased the
intensity and network fluctuations with an average demand for low-latency and high-throughput
prediction accuracy of 93.6%. This prediction computing. Traditional cloud computing
informs a Dynamic Workload Offloading infrastructures, though highly scalable and resource-
Framework (DWOF), which makes real-time rich, often suffer from communication delays and
decisions on task placement, based on latency bandwidth limitations when serving latency-sensitive
thresholds, energy profiles, and resource tasks. Conversely, edge computing, which brings
availability. A new Multi-Layer Token-Based computational resources closer to the data source, can
Resource Controller (MTBRC) is also proposed to reduce latency but lacks the scalability and
balance CPU, memory, and bandwidth allocation processing power required for complex workloads.
efficiently across nodes. Experimental evaluations However, efficient resource allocation in such hybrid
using a custom simulation environment, environments remains a significant challenge. Static
HyScaleSim, demonstrate that our approach or rule-based scheduling techniques are inadequate
achieves a 38% reduction in task execution latency, for handling dynamic workloads and fluctuating
42% improvement in throughput, and 31% network conditions. Moreover, unbalanced resource
enhancement in resource utilization efficiency, distribution can result in SLA (Service Level
compared to conventional cloud-first allocation Agreement) violations, increased latency, and
models. Additionally, the system maintains over underutilized infrastructure. To address these
92% SLA compliance, even under high-load limitations, we propose a context-aware and
scenarios. These results highlight the potential of prediction-driven hybrid computing model that
hybrid computing in enabling scalable, high- dynamically orchestrates task allocation between
performance edge-cloud networks for future digital edge and cloud layers. Our model introduces two
ecosystems. core innovations: a Context-Aware Predictive
Resource Allocation Engine (CAPRAE) and a
Keywords: Multi-Tier Token-Based Resource Controller
Hybrid Computing Model, Edge-Cloud (MTBRC). CAPRAE forecasts task demands using
Architecture, Resource Allocation, Workload lightweight machine learning algorithms, while
MTBRC ensures adaptive and fair distribution of dynamic task scheduling in edge-cloud platforms.
computing resources. We validate the proposed While accurate, the complexity and computational
model through an extensive simulation using a overhead of deep models make them less suitable for
custom-built tool, HyScaleSim, under various load real-time edge scenarios.
conditions. The results reveal substantial
improvements in latency, throughput, and SLA Unlike prior work, our proposed system integrates
compliance, demonstrating the effectiveness and lightweight machine learning prediction with
scalability of the proposed solution in next- context-aware orchestration, optimizing both
generation distributed systems. latency-sensitive and resource-intensive tasks.
Additionally, our Multi-Tier Token-Based
Resource Controller (MTBRC) introduces a novel
way to manage resource contention across
hierarchical computing layers, filling the critical gap
between scalability and real-time adaptability.
To ensure effective local processing, an Edge In cases of predicted failure or overload, tasks are
Resource Monitor (ERM) runs continuously on proactively reassigned to more capable nodes. This
each device. It records metrics like CPU usage, predictive and adaptive management ensures that
memory consumption, bandwidth availability, and both real-time responsiveness and computational
current task loads. This data is sent periodically to the throughput are optimized.
orchestration layer, enabling real-time awareness of
edge capabilities. Tasks that are identified as
C. Cloud Layer
lightweight or time-critical, such as motion detection
or anomaly filtering, are retained and processed
The Cloud Layer serves as the backbone for large-
locally.
scale computation, long-term storage, and backup
processing. Unlike edge nodes that are resource-
When resource constraints are detected or predicted,
constrained, cloud servers offer virtually unlimited
tasks are flagged for offloading. These decisions are
processing power and can host data-intensive
not static; instead, they are dynamically adjusted
applications such as machine learning training, big
using input from the orchestration layer, ensuring that
data analytics, and archival systems. The cloud layer
edge devices operate efficiently without being
also provides failover capabilities when edge or fog
overloaded or underutilized.
components become unavailable or saturated.
A. Context-Aware Predictive Resource Allocation The tiered structure of MTBRC respects the
Engine (CAPRAE) heterogeneous nature of edge-cloud architectures. For
example, edge nodes typically have fewer tokens due
CAPRAE functions as the brain of the system, to limited capacity, while cloud resources have a
gathering data from distributed edge devices, fog larger token pool but with higher latency. MTBRC’s
nodes, and cloud resources to predict workload allocation strategy balances these factors,
demand and resource availability proactively. Unlike maximizing overall system efficiency and SLA
traditional static allocation methods, CAPRAE uses adherence.
real-time context data such as CPU load, memory
consumption, network latency, energy levels, and C. Dynamic Workload Scheduler (DWS)
task priority. These inputs provide a comprehensive
view of system health and operational conditions. The DWS orchestrates the execution location of tasks
based on inputs from CAPRAE and MTBRC. It
To generate accurate forecasts without incurring considers multiple criteria including predicted
heavy computational overhead, CAPRAE employs a
resource demand, task latency sensitivity, current
node workloads, and network congestion metrics.
DWS also features a feedback mechanism, This section presents a comprehensive evaluation of
continuously monitoring task execution times and the proposed hybrid resource allocation framework,
success rates. This feedback refines future scheduling leveraging the HyScaleSim simulation platform. The
decisions, enabling the system to learn and improve system’s performance is analyzed using several
its responsiveness to environmental changes. critical metrics: task latency, resource utilization,
SLA compliance, and system throughput. The
D. Simulation Environment proposed approach, combining the Context-Aware
Predictive Resource Allocation Engine (CAPRAE),
To validate the proposed hybrid resource allocation Multi-Tier Token-Based Resource Controller
methodology, a comprehensive simulation (MTBRC), and Dynamic Workload Scheduler
environment named HyScaleSim was developed (DWS), is compared with two widely used baseline
using Python frameworks SimPy (for discrete-event models:
simulation) and TensorFlow Lite (for embedded
machine learning inference). A. Task Latency
HyScaleSim models realistic scenarios involving Task latency is a crucial factor for applications
heterogeneous edge devices (e.g., Raspberry Pi 4), demanding real-time or near-real-time
fog nodes, and cloud infrastructure with variable responsiveness, such as video surveillance,
network latencies and bandwidth constraints. The autonomous driving, and industrial IoT monitoring.
simulation incorporates stochastic task arrival The evaluation revealed that the proposed hybrid
patterns with varying SLA requirements, simulating model significantly reduces average task latency by
peak and off-peak loads. 32.7% compared to the cloud-only model and 18.4%
relative to the static allocation scheme. This
Metrics tracked include task completion latency, improvement is largely driven by the predictive
throughput, SLA compliance rates, and resource capabilities of CAPRAE, which forecasts resource
utilization. HyScaleSim enables fine-grained control needs and enables preemptive allocation of tasks to
over workload characteristics and network the most appropriate layer—edge, fog, or cloud.
parameters, allowing systematic evaluation of the
CAPRAE-MTBRC-DWS integrated system against By minimizing unnecessary cloud offloading,
baseline static and cloud-only models. especially for latency-sensitive tasks, the system
achieves faster processing close to the data source.
Additionally, the token-based mechanism prioritizes
critical tasks, preventing resource contention that can
cause delays. Latency distribution graphs indicate
that more than 85% of tasks meet their latency
deadlines even under peak workload scenarios, D. System Throughput
demonstrating the system’s robustness in handling
high traffic without significant performance System throughput measures the volume of tasks
degradation. successfully processed within a given timeframe,
reflecting the system’s capacity and efficiency. The
B. Resource Utilization proposed framework demonstrated a 27.5% increase
in throughput compared to baseline models. This
Optimizing resource usage is vital for operational enhancement results from reduced task queuing and
efficiency and cost savings. The proposed token- faster execution times enabled by the predictive
based controller enables balanced distribution of scheduling and token control.
CPU cycles, memory, and bandwidth across the edge,
fog, and cloud layers. Results show a 22% increase The dynamic workload scheduler optimizes task
in CPU and memory utilization at the edge layer placement, preventing bottlenecks and evenly
compared to baseline models, which means the distributing workloads across layers. It also adapts to
system effectively leverages nearby resources rather fluctuating demand by redirecting tasks in real-time,
than defaulting to cloud processing. which helps sustain high throughput even during
peak loads. These findings confirm the framework’s
Higher edge utilization reduces latency and network scalability and ability to handle variable workloads
traffic, leading to lower operational costs and energy typical in edge-cloud applications.
consumption. Simultaneously, cloud resource usage
decreases by 15%, reflecting that the system offloads E. Discussion and Insights
to the cloud only when necessary, avoiding excessive
dependency on centralized data centers. This The experimental results validate the effectiveness of
balanced approach not only extends the lifetime of integrating machine learning-based prediction with a
edge devices by avoiding overuse but also optimizes token-based resource allocation mechanism.
cloud costs through selective task offloading. CAPRAE’s lightweight decision-tree model strikes a
balance between prediction accuracy and
C. SLA Compliance computational efficiency, making it suitable for
deployment in resource-constrained edge
Maintaining Service Level Agreements (SLAs) is environments.
essential for ensuring Quality of Service (QoS) and
user satisfaction. The hybrid framework achieved an While the simulation environment models realistic
SLA compliance rate of 94.3%, outperforming static conditions including heterogeneous devices, variable
and cloud-only baselines by approximately 12% and network latency, and stochastic task arrivals, real-
18%, respectively. This improvement is attributed to world deployments may face additional challenges
the combined effects of predictive workload such as hardware failures, unpredictable network
forecasting and dynamic token-based resource outages, and security vulnerabilities. Future research
allocation. can focus on incorporating adaptive learning
techniques like reinforcement learning to enable the
By predicting resource bottlenecks in advance, system to better handle uncertainties and improve
CAPRAE allows proactive redistribution of tasks, decision-making over time. Additionally,
preventing SLA violations caused by resource experimental validation on physical testbeds could
shortages or delays. Moreover, MTBRC’s token further verify practical applicability.
mechanism guarantees that high-priority tasks receive
the necessary resources, reducing the likelihood of Overall, the proposed methodology demonstrates a
deadline misses. These results highlight the system’s robust and scalable approach to resource
capability to meet stringent QoS requirements in management in edge-cloud networks, effectively
dynamic and heterogeneous environments. improving latency, resource efficiency, SLA
adherence, and throughput, thus addressing key
limitations of existing static or cloud-centric models.
real-world testbeds will further assess its practical
applicability and scalability. The hybrid approach
outlined in this work paves the way for more
resilient, efficient, and scalable computing
infrastructures essential for the growing demands of
IoT, smart cities, and other latency-sensitive
applications.
VI. Conclusion