0% found this document useful (0 votes)
17 views46 pages

Module-1 Notes

The document discusses scalable computing over the Internet, highlighting the evolution of computing technologies from centralized systems to cloud computing and high-performance computing (HPC) and high-throughput computing (HTC) paradigms. It outlines the distinctions between various computing models, including centralized, parallel, distributed, and cloud computing, while emphasizing the importance of efficiency, dependability, and flexibility in future computing designs. Additionally, it explores the Internet of Things (IoT) and Cyber-Physical Systems (CPS), detailing their development, challenges, and potential applications in creating intelligent environments.

Uploaded by

Sneha Raju
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
17 views46 pages

Module-1 Notes

The document discusses scalable computing over the Internet, highlighting the evolution of computing technologies from centralized systems to cloud computing and high-performance computing (HPC) and high-throughput computing (HTC) paradigms. It outlines the distinctions between various computing models, including centralized, parallel, distributed, and cloud computing, while emphasizing the importance of efficiency, dependability, and flexibility in future computing designs. Additionally, it explores the Internet of Things (IoT) and Cyber-Physical Systems (CPS), detailing their development, challenges, and potential applications in creating intelligent environments.

Uploaded by

Sneha Raju
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 46

Department of Computer Science and Engineering

CLOUD COMPUTING (BCS601)

Module-01
Distributed System Models and Enabling Technologies
Scalable Computing Over the Internet
Scalable Computing over the Internet refers to the ability to dynamically allocate and
manage computing resources over the internet in a way that can handle growing demands.

.IN
This involves distributing computational tasks across multiple systems (often using cloud
computing or distributed computing platforms) to accommodate varying workloads.
C
Evolution of Computing Technology

 Over the last 60 years, computing has evolved through multiple platforms and
N
environments.
SY

 Shift from centralized computing to parallel and distributed systems.


 Modern computing relies on data-intensive and network-centric architectures.

The Age of Internet Computing


U

 High-performance computing (HPC) systems cater to large-scale computational needs.


VT

 High-throughput computing (HTC) focuses on handling a high number of simultaneous


tasks.
 The shift from Linpack Benchmark to HTC systems for measuring performance.

Platform Evolution

 First Generation (1950-1970): Mainframes like IBM 360 and CDC 6400.
 Second Generation (1960-1980): Minicomputers like DEC PDP 11 and VAX.
 Third Generation (1970-1990): Personal computers with VLSI microprocessors.
 Fourth Generation (1980-2000): Portable and wireless computing devices.
 Fifth Generation (1990-present): HPC and HTC systems in clusters, grids, and cloud
computing.

Dept. of CSE, SVIT Page 1


.IN
C
N
SY

FIGURE 1.1 Evolutionary trend toward parallel, distributed, and cloud computing with clusters,
MPPs, P2P networks, grids, clouds, web services, and the Internet of Things.

High-Performance Computing (HPC)


U

 Focused on raw speed, measured in floating-point operations per second (FLOPS).


VT

 Used mainly in scientific, engineering, and industrial applications.


 Limited to a small number of specialized users.

High-Throughput Computing (HTC)

 Shift from HPC to HTC for market-oriented applications.


 Used in Internet searches, web services, and enterprise computing.
 Emphasis on cost reduction, energy efficiency, security, and reliability.

Emerging Computing Paradigms

 Service-Oriented Architecture (SOA): Enables Web 2.0 services.


 Virtualization: Key technology for cloud computing.
 Internet of Things (IoT): Enabled by RFID, GPS, and sensor technologies.
 Cloud Computing: Evolution of computing as a utility.

Dept. of CSE, SVIT Page 2


Computing Paradigm Distinctions

 Centralized Computing: All resources in one system.


 Parallel Computing: Processors work simultaneously in a shared-memory or
distributed-memory setup.
 Distributed Computing: Multiple autonomous computers communicate over a
network.
 Cloud Computing: Uses both centralized and distributed computing over data
centers.

Distributed System Families

.IN
 Clusters: Homogeneous compute nodes working together.
 Grids: Wide-area distributed computing infrastructures.
 P2P Networks: Client machines globally distributed for file sharing and content
delivery.
C
 Cloud Computing: Utilizes clusters, grids, and P2P technologies.
N
Future Computing Needs and Design Objectives
SY

 Efficiency: Maximizing parallelism, job throughput, and power efficiency.


 Dependability: Ensuring reliability and Quality of Service (QoS).
 Adaptation: Scaling to billions of requests over vast data sets.
U

 Flexibility: Supporting both HPC (scientific/engineering) and HTC (business)


applications.
VT

Scalable Computing Trends and New Paradigms


Computing Trends and Parallelism

 Technological progress drives computing applications, as seen in Moore’s Law (processor


speed doubling every 18 months) and Gilder’s Law (network bandwidth doubling yearly).
 Commodity hardware advancements, driven by personal computing markets, have
influenced large-scale computing.

Degrees of Parallelism (DoP):

 Bit-level (BLP): Transition from 4-bit to 64-bit CPUs.


 Instruction-level (ILP): Techniques like pipelining, superscalar processing, and
multithreading.
 Data-level (DLP): SIMD and vector processing for efficient parallel execution.
 Task-level (TLP): Parallel tasks on multicore processors, though challenging to
program.

Dept. of CSE, SVIT Page 3


 Job-level (JLP): High-level parallelism in distributed systems, integrating fine- grain
parallelism.

Innovative Applications of Parallel and Distributed Systems

Transparency in data access, resource allocation, job execution, and failure recovery is
essential.

Application domains:

 Banking and finance: Distributed transaction processing and data consistency


challenges.
 Science, engineering, healthcare, and web services: Demand scalable and reliable

.IN
computing.
 Challenges include network saturation, security threats, and lack of software
support.
C
OR
N
Both HPC and HTC systems desire transparency in many application aspects. For example,
data access, resource allocation, process location, concurrency in execution, job replication,
SY

and failure recovery should be made transparent to both users and system management.
Table 1.1 highlights a few key applications that have driven the development of parallel and
distributed systems over the years. These applications spread across many important domains
U

in science, engineering, business, education, health care, traffic control, Internet and web
services, military, and government applications. Almost all applications demand computing
VT

economics, web-scale data collection, system reliability, and scalable performance. For
example, distributed transaction processing is often practiced in the banking and finance
industry. Transactions represent 90 percent of the existing market for reliable banking
systems. Users must deal with multiple database servers in distributed transactions.
Maintaining the consistency of replicated transaction records is crucial in real-time banking
services. Other complications include lack of software support, network saturation, and
security threats in these applications.

Dept. of CSE, SVIT Page 4


Utility Computing and Cloud Adoption
.IN
C
N
SY
U
VT

o Utility computing: Provides computing resources as a paid service (grid/cloud


platforms).
o Cloud computing extends utility computing, leveraging distributed resources and
virtualized environments.
o Challenges: Efficient processors, scalable memory/storage, distributed OS,
middleware, and new programming models.

Dept. of CSE, SVIT Page 5


Hype Cycle of Emerging Technologies

• New technologies go through five stages:


o Innovation trigger → Peak of inflated expectations → Disillusionment →
Enlightenment → Productivity plateau.
• 2010 Predictions:
 Cloud computing was expected to mature in 2-5 years.
 3D printing was 5-10 years from mainstream adoption.
 Mesh network sensors were more than 10 years from maturity.
 Broadband over power lines was expected to become obsolete.
• Promising technologies: Cloud computing, biometric authentication, interactive TV,

.IN
speech recognition, predictive analytics, and media tablets.

The Internet of Things and Cyber-Physical Systems


C
The Internet of Things (IoT)
N
 IoT extends the Internet to everyday objects, interconnecting devices, tools, and
computers via sensors, RFID, and GPS.
SY

 History: Introduced in 1999 at MIT, IoT enables communication between objects


and people.
 IPv6 Impact: With 2¹²⁸ IP addresses, IoT can assign unique addresses to all
U

objects, tracking up to 100 trillion static or moving objects.


VT

Communication Models:
 H2H (Human-to-Human)
 H2T (Human-to-Thing)
 T2T (Thing-to-Thing)

Development & Challenges:


 IoT is in its early stages, mainly advancing in Asia and Europe.
 Cloud computing is expected to enhance efficiency, intelligence, and
scalability in IoT interactions.

Smart Earth Vision: IoT aims to create intelligent cities, clean energy, better healthcare, and
sustainable environments.

Dept. of CSE, SVIT Page 6


.IN
C
N
SY

Cyber-Physical Systems (CPS)


U

CPS integrates computation, communication, and control (3C) into a closed intelligent
VT

feedback system between the physical and digital worlds.

Features:

 IoT vs. CPS: IoT focuses on networked objects, while CPS focuses on VR
applications in the real world.
 CPS enhances automation, intelligence, and interactivity in physical
environments.

Development:

 Actively researched in the United States.


 Expected to revolutionize real-world interactions just as the Internet transformed
virtual interactions.

Dept. of CSE, SVIT Page 7


Technologies for Network Based Systems
Introduction to Distributed Computing Technologies

Discusses hardware, software, and network technologies for distributed computing. Focuses
on designing distributed operating systems for handling massive parallelism.

Advances in CPU Processors

 Modern CPUs use multicore architecture (dual, quad, six, or more cores).
 Instruction-Level Parallelism (ILP) and Thread-Level Parallelism (TLP) improve

.IN
performance.

• Processor speed evolution:


C
 1 MIPS (VAX 780, 1978) → 1,800 MIPS (Intel Pentium 4, 2002) → 22,000 MIPS
N
(Sun Niagara 2,2008).
SY

• Moore’s Law holds for CPU growth, but clock rates are limited (~5 GHz max) due to heat and
power constraints.
U

• Modern CPU technologies include:


VT

 Superscalar architecture, dynamic branch prediction, speculative execution.


 Multithreaded CPUs (e.g., Intel i7, AMD Opteron, Sun Niagara, IBM Power 6).

Dept. of CSE, SVIT Page 8


.IN
C
N
SY
U
VT

Dept. of CSE, SVIT Page 9


The hierarchy of caches plays a crucial role in optimizing the performance of multicore CPUs
and many-core GPUs by reducing memory access latency.

Cache Hierarchy Overview:

Caches are small, fast storage layers closer to the processor than the main memory. A typical
hierarchy includes:

 L1 Cache: Closest to the processor core; fastest but smallest in size.


 L2 Cache: Larger than L1 but slightly slower; shared by one or more cores.
 L3 Cache: Even larger and slower; usually shared among all cores in a processor.

.IN
How it Works:

 Data Locality: Frequently accessed data is stored in the caches, exploiting spatial and
C
temporal locality to speed up access.
 Multilevel Approach: The cache hierarchy ensures faster access to data by maintaining
N
different levels of speed and size. L1 handles immediate needs, while L2 and L3 back it
up.
SY

 Coherence and Sharing: In multicore systems, maintaining cache coherence is vital.


Protocols like MESI (Modified, Exclusive, Shared, Invalid) ensure data consistency
across cores.
U

In GPUs:
VT

 GPUs also employ cache hierarchies, but their design is optimized for parallel workloads.
o Shared Memory: A fast, programmable memory shared among threads within a
core.
o L1 and L2 Caches: Designed to handle specific workloads like texture and global
memory accesses efficiently.

In both CPUs and GPUs, cache hierarchies reduce dependence on the slower main memory,
improving overall performance.

Multicore CPU and Many-Core GPU Architectures

 CPUs may scale to hundreds of cores but face memory wall limitations.
 GPUs (Graphics Processing Units) are designed for massive parallelism and data-
level parallelism (DLP).
 x86-based processors dominate HPC and HTC systems.
 Trend towards heterogeneous processors combining CPU and GPU cores on a
single chip.
Dept. of CSE, SVIT Page 10
Multithreading Technology

.IN
C
N
SY
U
VT

 Each processor category employs different scheduling patterns and techniques to exploit
Instruction-Level Parallelism (ILP) and Thread-Level Parallelism (TLP).
 Fine-grain and coarse-grain multithreading focus on reducing idle cycles, while
superscalar, CMP, and SMT enhance parallel execution in various ways.
 Blank slots in the functional units represent inefficiency, with SMT generally showing
the least idle time.

1. Four-Issue Superscalar Processor:


o Single-threaded processor with four functional units.
o Executes only instructions from a single thread at any given time.
o Limited by the availability of instructions from the same thread.
2. Fine-Grain Multithreaded Processor:
o Switches between different threads at every clock cycle.

Dept. of CSE, SVIT Page 11


o Ensures maximum utilization of functional units by avoiding stalls due to single-
thread dependencies.
o Suitable for workloads with many independent threads.
3. Coarse-Grain Multithreaded Processor:
o Executes multiple instructions from the same thread over several cycles before
switching to another thread.
o Reduces thread-switching overhead but may leave functional units idle during
thread stalls.
4. Dual-Core (2-Processor Chip Multiprocessing - CMP):

.IN
o Contains two independent processing cores, each functioning as a two-way
superscalar processor.
C
o Executes instructions from different threads on separate cores, fully exploiting
thread-level parallelism (TLP).
N
5. Simultaneous Multithreaded (SMT) Processor:
SY

o Allows multiple threads to execute simultaneously on a single core by sharing


functional units.
o Maximizes throughput by interleaving instructions from different threads in the
U

same cycle.
VT

GPU Computing and Exascale Systems

• GPUs were initially graphics accelerators, now widely used for HPC and AI.

• First GPU: NVIDIA GeForce 256 (1999).

• Modern GPUs have hundreds of cores, e.g., NVIDIA CUDA Tesla.

• GPGPU (General-Purpose GPU Computing) enables parallel processing beyond graphics.

How GPUs Work

• Early GPUs functioned as CPU coprocessors.

• Modern GPUs have 128+ cores, each handling multiple threads.

• GPUs optimize throughput, CPUs optimize latency.

• Used in supercomputers, AI, deep learning, gaming, and mobile devices.

Dept. of CSE, SVIT Page 12


GPU Programming Model

.IN
This section highlights the interaction between a CPU and a GPU for parallel execution of
floating-point operations:
C
1. CPU as the Controller:
N
o The CPU, with its multicore architecture, has limited parallelism.
o It offloads floating-point kernel computations to the GPU for massive data
SY

processing.
2. GPU as the Workhorse:
o The GPU features a many-core architecture with hundreds of simple processing
U

cores grouped into multiprocessors.


o Each core supports multiple threads, enabling extensive parallel execution.
VT

3. Key Process:
o The CPU instructs the GPU to perform large-scale computations.
o Efficient communication requires matching bandwidths between the on-board
main memory and the GPU's on-chip memory.
4. NVIDIA CUDA Framework:
o This process is commonly implemented using CUDA programming.
o Examples of GPUs for such tasks include NVIDIA's GeForce 8800, Tesla, and
Fermi GPUs.

The setup ensures that the computational workload is distributed effectively, leveraging the
strengths of both CPU and GPU architectures.

Example : the NVIDIA Fermi GPU Chip with 512 CUDA Cores

Dept. of CSE, SVIT Page 13


.IN
C
N
SY
U

The NVIDIA Fermi GPU architecture is designed for high performance, efficiency, and
VT

programmability. Here’s a quick summary:

 Structure:
o Built with 16 streaming multiprocessors (SMs), each containing 32 CUDA cores,
for a total of 512 CUDA cores.
o Each SM supports parallel execution, enabling massive computational power.
 Key Features:
o Optimized for floating-point operations and parallel processing.
o Allows for high memory bandwidth between on-chip GPU memory and external
device memory.
o Enhanced programmability using NVIDIA’s CUDA platform.

Dept. of CSE, SVIT Page 14


Power Efficiency of the GPU

Power Efficiency:

 GPUs consume significantly less power per instruction compared to CPUs.


 This efficiency makes GPUs more suitable for large-scale, power-constrained systems
like exascale computing.

Performance and Parallelism:

 GPUs are optimized for high throughput and parallel processing, with explicit on-chip

.IN
memory management.
 CPUs are designed for latency optimization in caches and memory, limiting their
C
parallelism.
N
Performance/Power Ratio:
SY

 In 2010, GPUs achieved a performance of 5 Gflops/watt per core, far exceeding CPUs at
less than 1 Gflop/watt.
U

 Exascale computing requires an estimated 60 Gflops/watt per core, highlighting the


need for further GPU advancements.
VT

Challenges and Future Directions:

 Power constraints and data movement dominate system limitations.


 Solutions include optimizing storage hierarchy, designing application-specific memory,
developing self-aware OS and runtime support, and building locality-aware compilers for
GPU-based massively parallel processors (MPPs).

Dept. of CSE, SVIT Page 15



.IN
C
Memory, Storage, and Wide-Area Networking
N
Memory Technology

• DRAM capacity growth: 16 KB (1976) → 64 GB (2011), increasing 4× every 3 years.


SY

• Memory wall problem: CPU speed increases faster than memory access speed, creating a
performance gap.
U

• Hard drive capacity growth: 260 MB (1981) → 250 GB (2004) → 3 TB (2011), increasing 10×
every 8 years.
VT

• Challenge: Faster CPUs and larger memory lead to CPU-memory bottlenecks.

Disks and Storage Technology

 Disk arrays exceeded 3 TB in capacity beyond 2011.


 Flash memory & SSDs are revolutionizing HPC (High-Performance Computing) and HTC
(High-Throughput Computing).
 SSD lifespan: 300,000–1 million write cycles per block, making them durable for years.

Storage trends:

 Tape units → obsolete


 Disks → function as tape units
 Flash storage → replacing traditional disks
 Memory → functions as cache

Challenges: Power consumption, cooling, and cost of large storage systems.

Dept. of CSE, SVIT Page 16


.IN
C
System-Area Interconnects
N
These three network types are frequently found in large clusters built with commercial
SY

components, showcasing flexibility in cluster design.

LAN (Local Area Network):


U

 Typically connects client hosts to large servers.


 Provides communication within a limited area, such as a cluster.
VT

SAN (Storage Area Network):

 Links servers to network storage like disk arrays.


 Allows efficient and centralized data storage management.

NAS (Network Attached Storage):

 Directly connects client hosts to storage devices (disk arrays).


 Offers shared storage access to multiple clients.

Small Clusters:

 Often built using a multiport Gigabit Ethernet switch and copper cables.
 Suitable when large distributed storage is not needed.

Dept. of CSE, SVIT Page 17


.IN
C
N
Wide-Area Networking
SY

Ethernet speed evolution:

 10 Mbps (1979) → 1 Gbps (1999) → 40-100 Gbps (2011) → projected 1 Tbps


U

(2013).
 Network performance grows 2× per year, surpassing Moore’s Law for CPUs.
VT

 High-bandwidth networking enables large-scale distributed computing.


 IDC 2010 report: InfiniBand & Ethernet will dominate HPC interconnects.
 Most data centers use Gigabit Ethernet for server clusters.

Virtual Machines and Virtualization Middleware


A conventional computer has a single OS image. This offers a rigid architecture that tightly couples
application software to a specific hardware platform.

Conventional Computers:

 Operate with a single OS image.


 Software is tightly linked to specific hardware, limiting flexibility.

Virtual Machines (VMs):

 Solve issues like resource underutilization and software inflexibility.


 Provide virtualized resources: processors, memory, and I/O.
Dept. of CSE, SVIT Page 18
Virtualization in Large Systems:

 Essential for clusters, grids, and clouds.


 Enables efficient, dynamic access to computing and storage.

Purpose of Virtualization:

 Aggregates resources for a unified system view.


 Improves manageability, flexibility, and efficiency.

Examples of Virtualized Resources:

 Virtual Machines (VMs).

.IN
 Virtual storage and networking.
C
N
SY
U
VT

(a) Physical Machine:

 The traditional setup where a single OS directly manages the hardware.


 No virtualization; applications and the OS depend entirely on the physical machine.

(b) Native VM Architecture:

 A virtual machine monitor (VMM or hypervisor) operates directly on the hardware.


 Efficient and high-performing, often used for managing resources in large-scale systems.

(c) Hosted VM Architecture:

 A VMM runs on top of a host operating system, treating VMs as applications.


Dept. of CSE, SVIT Page 19
 Simpler to set up but less efficient due to the extra layer introduced by the host OS.

(d) Dual-Mode VM Architecture:

 Combines features of both native and hosted architectures.


 Some tasks are handled directly by the VMM on hardware, while others pass through a
host OS.

Virtual Machines

Host Machine and Hardware:

 The physical hardware serves as the base for virtualization.

.IN
 Example: An x86 desktop running Windows OS as the host.

Virtual Machine (VM):


C
 Built with virtual resources managed by a guest OS to run specific applications.
 Provides hardware independence, allowing VMs to be ported across platforms.
N
Middleware Layer - Virtual Machine Monitor (VMM):
SY

 Positioned between the host and VM.


 Controls virtualization and resource management.
U

VM Architectures:
VT

 (b) Native (Bare-Metal) VM: Hypervisor (e.g., XEN) operates directly on hardware in
privileged mode. Guest OS could differ, like Linux on Windows hardware.
 (c) Hosted VM: VMM runs in nonprivileged mode on a host OS. No need to modify the
host.
 (d) Dual-Mode VM: Splits VMM functions between user level and supervisor level. May
require minor modifications to the host OS.

Key Benefits:

 Supports multiple VMs on the same hardware.


 Facilitates portability and flexibility with virtual appliances bundled with their dedicated
OS and applications.

VM Primitive Operations

VM Operations :

 Multiplexing: VMs are distributed across multiple hardware machines for efficient
resource use.
Dept. of CSE, SVIT Page 20
 Suspension: A VM can be paused and stored in stable storage for later use.
 Provision (Resumption): A suspended VM can be resumed on a new hardware
platform.
 Live Migration: A VM can be moved from one hardware platform to another without
disruption.

Benefits:

 Hardware Independence: VMs can run on any available hardware platform, regardless
of the underlying OS or architecture.
 Flexibility: Enables seamless porting of distributed applications.
 Enhanced Efficiency: Consolidates multiple server functions on a single hardware

.IN
platform, reducing server sprawl.
 Improved Utilization: VMware estimates server utilization increases from 5–15% to 60–
80% with this approach.
C
N
SY
U
VT

Virtual Infrastructures

Virtual Infrastructure:

Dept. of CSE, SVIT Page 21


 Connects physical resources (compute, storage, networking) to distributed applications
using VMs.
 Separates hardware and software for dynamic resource mapping.

Benefits:

 Reduces costs, increases efficiency, and improves responsiveness.


 Example: Server consolidation and containment using virtualization.

Future Support:

 Virtualization is pivotal for clusters, clouds, and grids, as discussed in later chapters.

.IN
Data Center Trends

 Shows customer spending, server installation growth, and cost breakdown over the
years.
C
 Highlights the "virtualization management gap," emphasizing the growing importance of
virtualization technologies.
N
SY
U
VT

Data Centre Virtualization for Cloud Computing

• Cloud Architecture: Uses commodity hardware (x86 processors, low-cost storage, Gigabit
Ethernet).

• Design Priorities: Cost-efficiency over raw performance, focusing on storage and energy
savings.
Dept. of CSE, SVIT Page 22
Data Centre Growth & Cost Breakdown

Large data centres contain thousands of servers.

Cost Distribution (2009 IDC Report):

o 30% for IT equipment (servers, storage).

o 60% for maintenance and management (cooling, power, etc.).

o Electricity & cooling costs increased from 5% to 14% in 15 years.

Low-Cost Design Philosophy

.IN
Uses commodity x86 servers and Ethernet networks instead of expensive hardware.

Software manages network traffic, fault tolerance, and scalability.


C
Convergence of Technologies Enabling Cloud Computing
N
1. Hardware Virtualization & Multi-Core Chips: Allows dynamic configurations.

2. Utility & Grid Computing: Forms the foundation of cloud computing.


SY

3. SOA, Web 2.0, and Mashups: Advances in web technologies drive cloud adoption.

4. Autonomic Computing & Data Center Automation: Enhances efficiency.


U

Impact of Cloud Computing on Data Science & E-Research


VT

• Data Deluge: Massive data from sensors, web, simulations, requiring advanced data
management.

• E-Science Applications: Used in biology, chemistry, physics, and social sciences.

• MapReduce & Iterative MapReduce: Enable parallel processing of big data.

• Multicore & GPU Clusters: Boost computational power for scientific research.

• Cloud Computing & Data Science Convergence: Revolutionizes computing architecture and
programming models.

System Models for Distributed and Cloud Computing

Distributed and Cloud Computing Systems

 Built over a large number of autonomous computer nodes.


 Nodes are interconnected using SANs, LANs, or WANs in a hierarchical manner.

Dept. of CSE, SVIT Page 23


 Clusters of clusters can be created using WANs for large-scale systems.
 These systems are highly scalable, supporting web-scale connectivity.

.IN
C
N
SY

Node Interconnection:
U

 Built over large numbers of autonomous computer nodes.


Nodes are interconnected hierarchically via SANs, LANs, or WANs.
VT


 LAN switches connect hundreds of machines; WANs link local clusters into massive
systems.

Massive Systems:

 Highly scalable, supporting web-scale connectivity.


 Involve hundreds, thousands, or millions of computers, working collectively or
collaboratively.

Classification (Table 1.2):

 Clusters: Homogeneous nodes for high-performance tasks like supercomputing (e.g.,


Google Search Engine, IBM Road Runner).
 Peer-to-Peer (P2P) Networks: Flexible client networks for file sharing, social networking
(e.g., BitTorrent, Skype).
 Grids: Distributed resource pooling for global problem-solving (e.g., TeraGrid,
ChinaGrid).

Dept. of CSE, SVIT Page 24


 Cloud Platforms: Virtualized servers for utility computing and dynamic resource
provisioning (e.g., AWS, Microsoft Azure).

Applications and Trends:

 Clusters dominate supercomputing (e.g., 417 of Top 500 supercomputers in 2009).


 P2P faces resistance due to copyright issues.
 Clouds offer cost-effective and simple solutions, gaining popularity.

Clusters of Cooperative Computers


A computing cluster consists of interconnected stand-alone computers which work cooperatively as a
single integrated computing resource. In the past, clustered computer systems have demonstrated

.IN
impressive results in handling heavy workloads with large data sets.

Cluster Architecture
C
Cluster Architecture:
N
 Built around a low-latency, high-bandwidth interconnection network, such as SAN (e.g.,
Myrinet) or LAN (e.g., Ethernet).
SY

 Hierarchical construction allows for scalable clusters using Gigabit Ethernet, Myrinet, or
InfiniBand switches.

Interconnection and Internet Access:


U

 The cluster is connected to the Internet via a virtual private network (VPN) gateway.
VT

 The gateway's IP address is used to locate the cluster.

Node Management:

 Most clusters have loosely coupled node computers, each managed by its own operating
system (OS).
 This results in multiple system images, as nodes operate autonomously under different
OS controls.

Shared Resources:

 Shared I/O devices and disk arrays support the cluster.


 Despite being composed of multiple nodes, the cluster acts as a single computational
unit connected to the Internet.

Dept. of CSE, SVIT Page 25


Single-System Image
.IN
C
Greg Pfister highlights that an ideal cluster should integrate multiple system images into a single-
system image (SSI). This SSI creates an illusion, via software or hardware, where a collection of
N
resources appears as one unified, powerful resource. SSI enables clusters to function as a single
machine for users, supporting shared CPUs, memory, and I/O across cluster nodes. In contrast, a
SY

cluster with multiple system images consists of independent computers.

Hardware, Software, and Middleware Support


U

• Cluster nodes include PCs, workstations, servers, or SMP.

• MPI and PVM used for message passing.


VT

• Most clusters run on Linux OS.

• Middleware is essential for SSI, high availability (HA), and distributed shared memory (DSM).

• Virtualization allows dynamic creation of virtual clusters.

Major Cluster Design Issues

• No unified cluster-wide OS for resource sharing.

• Middleware is required for cooperative computing and high performance.

• Benefits of clusters: Scalability, efficient message passing, fault tolerance, and job management

Dept. of CSE, SVIT Page 26


Grid Computing Infrastructures

.IN
C
N
SY

Over the past 30 years, there has been significant progression in computing services:

 Internet services like Telnet allow local computers to connect to remote ones.
U

 Web services like HTTP enable remote access to web pages.


 Grid computing facilitates simultaneous interaction among applications running on
VT

distant computers.

Computational Grids

 Grid Computing: Functions like a power grid, combining computers,


software/middleware, instruments, people, and sensors into an integrated
infrastructure across LAN, WAN, or the Internet at various scales.
 Grids include workstations, servers, clusters, and supercomputers, while users can
access them via PCs, laptops, or PDAs.
 Provides shared resources across resource sites owned by different organizations,
enabling collaborative computational tasks.
 Built over IP broadband networks (LAN/WAN) and presented as a unified resource pool.
 Uses instruments like radio telescopes (e.g., SETI@Home for alien life searches) and
supports scientific research (e.g., pulsar studies).
 Enterprises and consumers shape usage trends, offering computing, communication,
content, and transaction services.
 Global Examples: National and international grids include TeraGrid (US), EGEE
(Europe), and ChinaGrid, supporting scientific applications.

Dept. of CSE, SVIT Page 27


Grid Families

 Grid Technology Evolution: Requires new distributed computing models,


software/middleware, network protocols, and hardware.
 Industry Involvement: National grid projects have led to platforms developed by
companies like IBM, Microsoft, Sun, HP, Dell, and others.
 Emerging Grid Services: Grid service providers (GSPs) and applications have grown
quickly, paralleling the rise of Internet and web services.
 Grid System Categories:
 Computational/Data Grids: Operate primarily at a national level.
 P2P Grids: Focus on peer-to-peer resource sharing.

.IN
C
N
SY
U
VT

Dept. of CSE, SVIT Page 28


Peer-to-Peer (P2P) Network Families

 Client-server architecture connects clients (PCs, workstations) to a central server


for various applications.
 P2P (peer-to-peer) architecture is a distributed, client-oriented network model.
 P2P systems lack a central server and rely on peers to act as both clients and
servers.
 P2P systems are described at the physical level and overlay networks at the logical
level

P2P Systems

.IN
In a P2P system, each node acts as both a client and a server, sharing system
resources.
C
 Peer machines are simply client computers connected via the Internet, with the
autonomy to join or leave freely.

N
No master-slave relationship, central coordination, or global database exists.
 The system is self-organizing with distributed control.
 P2P networks form ad hoc physical networks across Internet domains, using
SY

protocols like TCP/IP and NAI.


 These networks dynamically change in size and topology as peers join or leave
voluntarily.
U

Overlay Networks
VT

 Data and files are distributed among participating peers.


 Peer IDs create an overlay network, defining logical connections between peers.
 The overlay network uses virtual links mapped to physical machines via IDs.
 Peer IDs are added or removed from the overlay network when peers join or leave.

Two types of overlay networks:

1. Unstructured: Random graph, uses flooding for queries, causing high traffic and
unpredictable searches.
2. Structured: Follows specific topology and rules, with routing mechanisms for better
efficiency.

Dept. of CSE, SVIT Page 29


P2P Application Families
.IN
C
 File Sharing: Distributed file sharing of digital content (e.g., Gnutella, Napster,
N
BitTorrent).
 Collaboration: Tools for chatting, instant messaging, and collaborative design (e.g.,
SY

MSN, Skype).
 Distributed Computing: Specific applications for computing power, such as
SETI@home with 25 Tflops over 3 million Internet hosts.
 Application Platforms: Systems like JXTA, .NET, and FightingAID@home offer support
for naming, discovery, security, communication, and resource aggregation.
U

P2P Computing Challenges


VT

 P2P systems face heterogeneity issues in hardware, software, and network requirements.
 Data locality, network proximity, and interoperability are key design objectives in P2P
applications.
 Routing efficiency, fault tolerance, failure management, and load balancing significantly
impact performance.
 Trust issues exist due to security, privacy, and copyright concerns among peers.
 P2P networks improve robustness by avoiding a single point of failure and replicating
data across nodes.
 Lack of centralization makes management complex.
 Security risks arise as any client can potentially cause damage or abuse.
 P2P networks are suitable for low-security applications without sensitive data.

Dept. of CSE, SVIT Page 30


Cloud Computing over the Internet
.IN
C
 Computational science is shifting to a data-intensive approach.
 Supercomputers need balanced systems, combining CPUs with petascale I/O and
N
networking arrays.
 Future data handling will involve sending computations to data rather than moving data
to workstations.
SY

 IT trends are moving computing and data from desktops to large data centers.
 Cloud computing emerged to handle the data explosion, offering on-demand software,
hardware, and data as a service.
U

 IBM defines a cloud as a pool of virtualized resources hosting diverse workloads,


including backend jobs and interactive applications.
 Clouds allow rapid provisioning of virtual or physical machines for scaling workloads.
VT

 They support redundant, self-recovering, and highly scalable models to handle


hardware/software failures.
 Real-time resource monitoring in clouds enables rebalancing of allocations when
necessary.

Internet Clouds

 Cloud computing uses virtualized platforms with elastic, on-demand resources.


 It dynamically provisions hardware, software, and data sets.
 The goal is to shift desktop computing to a service-oriented platform using server
clusters and large databases.
 Cloud computing is cost-effective and simple, benefiting users and providers.
 Machine virtualization enables the cost-efficiency of cloud computing.
 The cloud is designed to support many applications simultaneously.
 A secure, trustworthy, and dependable ecosystem is essential for cloud computing.

Dept. of CSE, SVIT Page 31


 Some users view the cloud as a centralized resource pool; others see it as a distributed
server cluster.

The Cloud Landscape

Traditional Distributed Computing Systems:

 Owned by autonomous domains (e.g., labs or companies) for on-premises needs.


 Performance bottlenecks: system maintenance, poor utilization, and high upgrade costs.

Cloud Computing:

 An on-demand computing paradigm addressing traditional system challenges.

Cloud Service Models:

.IN
1. Infrastructure as a Service (IaaS): Provides servers, storage, networks, and data center
C
resources. Users deploy and run VMs but don’t manage the cloud infrastructure.
2. Platform as a Service (PaaS): Enables deployment of user-built applications on a cloud
N
platform. Includes middleware, databases, tools, and APIs.
3. Software as a Service (SaaS): Browser-based software for business processes (CRM,
SY

ERP, HR, etc.), eliminating upfront costs for users and reducing hosting costs for
providers.

Cloud Deployment Modes:


U

 Private, public, managed, and hybrid, each with differing security implications.
VT

Cloud Advantages:

 Protected locations and energy efficiency.


 Improved peak-load capacity utilization.
 Separation of infrastructure maintenance and application development.
 Reduced costs compared to traditional systems.
 Support for application development.
 Efficient data discovery and service distribution.
 Addressing privacy, security, and reliability issues.
 Flexible service agreements and pricing models.

Dept. of CSE, SVIT Page 32


.IN
C
N
SOFTWARE ENVIRONMENTS FOR DISTRIBUTED SYSTEMS AND CLOUDS
SY

Service-Oriented Architecture (SOA)

 In grids/web services, Java, and CORBA:


 An entity is a service, a Java object, or a CORBA distributed object, respectively.
U

 These architectures build on the traditional seven Open Systems Interconnection


(OSI) layers for networking abstractions.
VT

Base Software Environment:

 .NET or Apache Axis for web services.


 Java Virtual Machine (JVM) for Java.
 Broker networks for CORBA.
 On top of the base software environment, a higher-level environment is built, tailored
to distributed computing features.
 This includes entity interfaces and inter-entity communication that reconstruct the
top four OSI layers, but at the entity level rather than the bit level.

Layered Architecture for Web Services and Grids

 Entity interfaces include WSDL, Java methods, and CORBA IDL for distributed systems.
 High-level communication systems use SOAP, RMI, and IIOP for RPC, fault recovery,
and routing.
 Middleware like WebSphere MQ or JMS supports routing and virtualization.
Dept. of CSE, SVIT Page 33
 Fault tolerance is achieved via WSRM, adapting OSI layer features for entity
abstractions.
 Security is implemented using OSI concepts like IPsec and secure sockets.
 Higher-level services provide registries, metadata, and entity management.
 Discovery tools like JNDI, UDDI, LDAP, and ebXML enable service discovery.
 Management services include CORBA Life Cycle, Enterprise JavaBeans models, and
Jini’s lifetime model.
 Shared memory models simplify information exchange.
 Distributed systems improve performance using multiple CPUs and support software
reuse.
 Older approaches like CORBA are being replaced by SOAP, XML, and REST in modern

.IN
systems.
C
N
SY
U
VT

Web Services and Tools

 Loose coupling and heterogeneous implementations make services more attractive than
distributed objects.
 Service architectures include web services and REST systems, each with distinct
approaches to interoperability.

Dept. of CSE, SVIT Page 34


Web Services:

 Fully specify the service and environment using SOAP messages.


 SOAP acts as a universal distributed operating system, but efficiency and agreement on
protocols remain challenging.

REST Systems:

 Prioritize simplicity and delegate complex problems to application-specific software.


 Minimal information is included in the header; the message body carries required data.
 REST is suitable for rapid technology environments and uses "XML over HTTP" instead
of SOAP.

.IN
CORBA and Java are linked via RPCs.

 In Java, method calls are replaced by Remote Method Invocation (RMI).


C
 CORBA uses C++-style syntax for linking object interfaces.
N
Grids and Clouds:

 Grids refer to either single services or collections of services.


SY

 Sensors represent entities producing data, while grids and clouds handle multiple
message-based inputs and outputs.

The Evolution of SOA


U

 Service-Oriented Architecture (SOA) has evolved to support grids, clouds, interclouds,


VT

and systems of systems.


 Sensors (e.g., ZigBee, Bluetooth, WiFi devices) collect raw data as sensor services (SS).
 Sensor services interact with grids, databases, and various clouds, such as compute,
storage, filter, and discovery clouds.
 Filter services (fs) process raw data, removing unwanted information to address
specific requests.
 A collection of filter services forms a filter cloud.
 SOA processes raw data into information, knowledge, and intelligence for decision-
making.
 Distributed systems use a web interface/portal to transform sensor data through
compute, storage, filter, and discovery clouds.
 Example portals include OGFCE and HUBzero, which employ web service and Web 2.0
technologies.

Dept. of CSE, SVIT Page 35


.IN
C
N
SY
U

Grids versus Clouds


VT

 Service-Oriented Architecture (SOA) has evolved to support grids, clouds, grids of


clouds, clouds of grids, and interclouds (clouds of clouds).
 Sensor services (SS) collect raw data; sensors include ZigBee devices, Bluetooth
devices, WiFi access points, personal computers, and wireless phones.
 SS devices interact with grids, databases, and various clouds, such as compute, storage,
filter, and discovery clouds.
 Filter services (fs) process raw data by eliminating unwanted information, forming
filter clouds to respond to specific requests.
 SOA transforms massive raw data into useful information, knowledge, and wisdom for
intelligent decision-making.
 Distributed systems often require web interfaces or portals to process and present
transformed data.
Dept. of CSE, SVIT Page 36
 Data streams from sensors pass through sequences of compute, storage, filter, and
discovery clouds before converging at portals accessed by users.
 Example portals include OGFCE and HUBzero, which use web service (portlet) and Web
2.0 (gadget) technologies.

Grids versus Clouds

 The boundary between grids and clouds is increasingly blurred.


 Workflow technologies in web services coordinate and orchestrate services, supporting

.IN
business models like two-phase transactions.
 Workflow standards and approaches include BPEL, Pegasus, Taverna, Kepler, Trident,
and Swift.
C
 Grids use static resources, while clouds focus on elastic resources.
N
 Key difference: Dynamic resource allocation via virtualization and autonomic
computing.
SY

 Grids can be built from multiple clouds, offering better support for negotiated resource
allocation.
U

 These combinations result in systems like clouds of clouds, grids of clouds, clouds of
grids, or inter-clouds in SOA architectures.
VT

Trends toward Distributed Operating Systems

 Distributed systems are loosely coupled, with multiple system images due to
independent operating systems on node machines.
 A distributed OS promotes resource sharing and efficient communication among nodes.
 Such a system is typically a closed system, relying on message passing and Remote
Procedure Calls (RPCs) for internode communication.
 A distributed OS is essential for improving the performance, efficiency, and flexibility
of distributed applications.

Dept. of CSE, SVIT Page 37


Distributed Operating Systems

 Network OS: Built over heterogeneous OS platforms, offers the lowest transparency,
and relies on file sharing for communication.
 Middleware: Provides limited resource sharing, similar to MOSIX/OS for clustered
systems.
 Truly Distributed OS: Achieves high resource usage and system transparency.

.IN
C
N
SY
U
VT

Amoeba versus DCE

 DCE is a middleware-based system for distributed computing environments.


 Amoeba was developed academically at Free University in the Netherlands.
 OSF promoted DCE for distributed computing.
 Amoeba, DCE, and MOSIX2 remain research prototypes, mainly used in academia.
 No successful commercial OS products emerged from these research systems.
 New web-based operating systems are needed to support resource virtualization in
distributed environments.
 These distributed OSs should distribute functionalities across available servers.

Dept. of CSE, SVIT Page 38


 Conventional OSs run on centralized platforms, while distributed OSs require a
lightweight microkernel approach (e.g., Amoeba).
 Alternatively, distributed OSs could extend existing systems like DCE by modifying UNIX.
 The trend is to relieve users of most resource management responsibilities.

MOSIX2 for Linux Clusters

 MOSIX2 is a distributed OS that operates with a virtualization layer in Linux.


 It provides a partial single-system image to user applications.
 Supports sequential and parallel applications, and migrates software processes among
Linux nodes.

.IN
 Manages Linux clusters or grids of multiple clusters with resource discovery and
process migration.
 Allows flexible grid management, enabling resource sharing among cluster owners.
 MOSIX-enabled grids can scale indefinitely, provided there is trust among cluster
C
owners.
 It is being applied to manage resources in Linux clusters, GPU clusters, grids, and
N
clouds (when VMs are used).
SY

Transparency in Programming Environments

 Transparent computing infrastructure separates user data, applications, OS, and


U

hardware into four distinct levels.


 User data is owned independently of applications.
VT

 The OS provides clear interfaces and standard system calls to application


programmers.
 Future cloud infrastructure will decouple hardware from the OS through standard
interfaces.
 Users can choose different OSes to run on their preferred hardware devices.
 Cloud applications as SaaS enable users to switch between services without binding
data to specific applications.

Parallel and Distributed Programming Models

 Four programming models for distributed computing aim for scalable performance and
application flexibility.

o MPI (Message Passing Interface):


 A library for C or FORTRAN to create parallel programs on distributed systems.
 Supports synchronous/asynchronous communication and I/O operations in user
programs.

Dept. of CSE, SVIT Page 39


MapReduce:

 A web programming model for scalable data processing on large clusters.


 Includes Map and Reduce functions for processing large data sets.

Hadoop:

 A software library for running large applications on vast data sets.


 Scalable, efficient, and reliable, often used in business and cloud computing.
 Service clouds rely on extending tools like Hadoop, EC2, and S3 for distributed
computing over storage systems.

.IN
C
N
SY
U
VT

Dept. of CSE, SVIT Page 40


Message-Passing Interface (MPI)
.IN
C
 MPI (Message Passing Interface): A primary standard for developing parallel and
concurrent programs on distributed systems.
N
 Provides a library of subprograms callable from C or FORTRAN for writing parallel
programs.
SY

 Designed to support clusters, grid systems, and P2P systems with enhanced web
services and utility computing applications.
 PVM (Parallel Virtual Machine): Another low-level primitive for distributed
programming.
U

 Both MPI and PVM are discussed in detail by Hwang and Xu [28].
VT

Hadoop Library

 Hadoop, originally developed by a Yahoo! group, is a software platform for writing and
running applications over vast distributed data.
 It can scale easily to store and process petabytes of web data.
 Hadoop is economical, offering an open-source version of MapReduce that reduces
overhead.
 It is efficient, achieving high parallelism across many commodity nodes.
 It is reliable, automatically maintaining multiple data copies for redeployment after
unexpected system failures.

Dept. of CSE, SVIT Page 41


Open Grid Services Architecture (OGSA)
.IN
C
 Grid infrastructure development is motivated by large-scale distributed computing
applications requiring extensive resource and data sharing.
N
 OGSA (Open Grid Services Architecture) is a common standard for public grid
services.
SY

 Genesis II is an implementation of OGSA.


 Key features of Genesis II include:
o A distributed execution environment.
o Public Key Infrastructure (PKI) services using a local certificate authority (CA).
U

o Trust management and security policies in grid computing.


VT

Globus Toolkits and Extensions

 Globus is a middleware library developed by the U.S. Argonne National Laboratory and
USC Information Science Institute.
 Implements OGSA standards for resource discovery, allocation, and security in grid
environments.
 Supports multisite mutual authentication using PKI certificates.
 The current version, GT 4, has been in use since 2008.
 IBM has extended Globus for business applications.

PERFORMANCE, SECURITY, AND ENERGY EFFICIENCY


Performance Metrics and Scalability Analysis

 Performance metrics are essential for assessing distributed systems.


 This section focuses on various dimensions of scalability and performance laws.

Dept. of CSE, SVIT Page 42


 System scalability will be evaluated in relation to OS images and the limiting factors
affecting performance.

Performance Metrics

 Processor and network performance are estimated using metrics like CPU speed in
MIPS and network bandwidth in Mbps.
 System throughput is measured in MIPS, Tflops, or TPS.
 Additional performance measures include job response time and network latency.
 A preferred interconnection network has low latency and high bandwidth.
 System overhead factors include OS boot time, compile time, I/O data rate, and
runtime support system.
 Other performance metrics involve QoS for Internet/web services, system availability

.IN
and dependability, and security resilience against network attacks.

Dimensions of Scalability
C
 Scalable performance requires backward compatibility with existing hardware and
software.
N
 Overdesign can be cost-ineffective, with scaling influenced by practical factors.
 Size scalability involves adding processors, cache, memory, storage, or I/O channels to
SY

boost performance or functionality.


 Software scalability includes upgrading OS, compilers, libraries, and application
software while ensuring compatibility with larger systems.
U

 Application scalability aligns problem size with machine size for efficiency and cost-
effectiveness, often by enlarging problem size.
VT

 Technology scalability adapts to new component and networking technologies,


considering time (generation updates), space (packaging and energy), and heterogeneity
(compatibility across diverse hardware or software).

Scalability versus OS Image Count

 Scalable performance improves speed by adding processors, servers, memory, disk


capacity, or I/O channels.
 OS image refers to independent OS instances in clusters, grids, P2P networks, or clouds.
 SMP (Symmetric Multiprocessor) has a single system image, limited to a few hundred
processors by 2010, with scalability constrained by packaging and interconnects.
 NUMA (Nonuniform Memory Access) uses SMP nodes with shared memory, scaling to
thousands of processors, such as a 2,048-processor NUMA machine with 32 OS images.
 Clusters consist of loosely coupled SMP servers or high-end machines, offering higher
scalability than NUMA. Total processors or cores exceed the number of OS images by
orders of magnitude.

Dept. of CSE, SVIT Page 43


 Clouds, being virtualized clusters, scaled up to a few thousand VMs by 2010.
 Grids include clusters, mainframes, supercomputers, or MPP systems, with fewer OS
images than processors.
 P2P networks can scale to millions of peer nodes, with performance relying on public
network QoS.
 Comparisons of scalability should consider similar networking levels across P2P, clouds,
and clusters.

.IN
C
N
SY

Amdahl’s Law
U

Amdahl's Law explains how much faster a program can run when split across multiple
VT

processors.

 Some part of the program (called the sequential bottleneck) must always run on a
single processor, and it can't be parallelized. This is represented by α\alphaα.
 The rest of the program can run in parallel across nnn processors. This is

(1−α)(1 - \alpha)(1−α).

The total time for the program on nnn processors is:

Tparallel=αT+(1−α)TnT_{\text{parallel}} = \alpha T + \frac{(1 - \alpha) T}{n}Tparallel


=αT+n(1−α)T

Where TTT is the original time on a single processor.

Speedup (how much faster the program runs) is:

Dept. of CSE, SVIT Page 44


S=TαT+(1−α)Tn=1α+(1−α)nS = \frac{T}{\alpha T + \frac{(1 - \alpha) T}{n}} = \frac{1}{\alpha +
\frac{(1 - \alpha)}{n}}S=αT+n(1−α)TT=α+n(1−α)1

As you add more processors, the speedup increases, but there's a limit. The maximum
speedup happens when α\alphaα (the part that can't be parallelized) is zero.

For example, if 25% of the code can't be parallelized (α=0.25\alpha = 0.25α=0.25), the best
speedup you'll ever get is 4, even with hundreds of processors.

Problem with Fixed Workload

In Amdahl's Law, the assumption is made that the workload remains the same for both

.IN
sequential and parallel execution, with a fixed problem size or data set. This is known as
fixed-workload speedup. C
The system efficiency EEE is defined as the speedup SSS divided by the number of
processors nnn, as follows:
N
E=Sn=1αn+(1−α)E = \frac{S}{n} = \frac{1}{\alpha n + (1 - \alpha)}E=nS=αn+(1−α)1
SY

In many cases, especially with large clusters, the system efficiency becomes quite low. For
instance, when executing the program on 256 processors, the efficiency is:
U

E=1(0.25×256)+0.75≈1.5%E = \frac{1}{(0.25 \times 256) + 0.75} \approx


1.5\%E=(0.25×256)+0.751≈1.5%
VT

This low efficiency occurs because only a few processors (e.g., 4) are actively used, while the
majority of the processors remain idle. As the cluster size increases, the parallel processing
becomes less efficient, mainly due to the sequential bottleneck (the part of the code that can't
be parallelized).

Gustafson’s Law

To improve efficiency when using large clusters, scaled-workload speedup is used, which
scales the problem size to match the cluster's capabilities. This approach, proposed by John
Gustafson (1988), adjusts the workload based on the number of processors.

Scaled-Workload Speedup:

 Let WWW be the original workload. When using nnn processors, the workload is scaled to
W′=αW+(1−α)nWW' = \alpha W + (1 - \alpha) n WW′=αW+(1−α)nW, where only the parallelizable
portion of the workload is scaled by nnn.

Dept. of CSE, SVIT Page 45


 The total execution time for the scaled workload W′W'W′ on nnn processors gives the scaled-
workload speedup S′S'S′:

S′=W′W=αW+(1−α)nWW=α+(1−α)nS' = \frac{W'}{W} = \frac{\alpha W + (1 - \alpha) n W}{W} = \alpha +


(1 - \alpha) nS′=WW′=WαW+(1−α)nW=α+(1−α)n

This is known as Gustafson's Law.

Efficiency with Scaled Workload:

The efficiency E′E'E′ for the scaled workload on nnn processors is:

.IN
E′=S′n=αn+(1−α)E' = \frac{S'}{n} = \frac{\alpha}{n} + (1 - \alpha)E′=nS′=nα+(1−α)

For example, if the workload is scaled for a 256-node cluster with α=0.25\alpha = 0.25α=0.25,
the efficiency is:
C
E′=0.25256+0.75≈0.751E' = \frac{0.25}{256} + 0.75 \approx 0.751E′=2560.25+0.75≈0.751
N
When to Apply Amdahl's vs. Gustafson's Law:
SY

 Amdahl’s Law should be used for fixed workloads, where the problem size doesn't change.
 Gustafson’s Law should be used when the workload is scaled according to the number of
processors, as it leads to better efficiency for large clusters.
U
VT

Dept. of CSE, SVIT Page 46

You might also like