Module 3 Notes
Module 3 Notes
Module-3
Cloud Platform Architecture over Virtualized Datacenters: Cloud Computing and Service Models,
Data Center Design and Interconnection Networks, Architectural Design of Compute and Storage
Clouds, Public Cloud Platforms: GAE, AWS and Azure, Inter-Cloud Resource Management.
Textbook 1: Chapter 4: 4.1 to 4.5
o
Examples of Real-World Private Clouds:
• NASA’s Climate Modeling Cloud:
o Enables researchers to run remote models without investing in HPC clusters.
o Cost-effective and centralized.
• CERN’s Scientific Cloud:
o Distributes applications, data, and computing resources to thousands of global
researchers.
Modern data center networking is designed for massive scalability, Internet-based access, and
dynamic workload handling. While supercomputers are built for high-performance tasks using
customized networks, cloud data centers prioritize flexibility, security, and multi-user
environments with efficient networking structures. These architectural distinctions make cloud
platforms ideal for evolving computational demands.
•
o Cloud Consumers: Businesses, developers, end-users leveraging cloud
services, demand flexible platform.
o Cloud Management: Intermediaries that manage service usage and
performance. Provides virtualized resources over an IaaS platform.
o VI Management: the manager allocates VMs’ over multiple clusters.
o VM managers: VM manager handle VMs installed individual host machines.
• Key Enabling Technologies:
o Virtualization: Allows multiple virtual machines (VMs) to run on a single
physical server, improving efficiency and flexibility.
o Containerization: Docker and Kubernetes enable efficient application
deployment across cloud environments.
o Microservices Architecture: Supports modular application development,
enhancing scalability and manageability.
o Artificial Intelligence (AI) & Machine Learning (ML): Enables cloud-based
predictive analytics, automation, and intelligent decision-making.
o Blockchain Technology: Used for secure transactions, data integrity, and
decentralized cloud services.
• Definition:
o Provides essential computing resources (compute, storage, networking) as
virtualized services over the internet.
• Key Characteristics:
o Pay-as-you-go pricing model.
o Highly scalable infrastructure.
o Automated resource provisioning.
• Examples:
•
o AWS EC2: Provides virtual machines with flexible configurations.
o
o Google Compute Engine (GCE): Scalable VM instances for cloud workloads.
o Microsoft Azure Virtual Machines: Supports both Windows and Linux
environments.
• Advantages:
o Eliminates the need for physical hardware investment.
o Provides high availability and disaster recovery solutions.
o Offers global reach and reduced latency through distributed data centers.
• Challenges:
o Requires advanced cloud expertise for configuration and management.
o Potential security risks if not properly configured.
4.1.4 Platform-as-a-Service (PaaS) and Software-as-a-Service (SaaS)
• Platform-as-a-Service (PaaS):
o Provides a cloud-based environment for application development and
deployment.
o Includes integrated tools such as databases, runtime environments, and
development frameworks.
o Examples:
o
▪ Google App Engine: Scalable application hosting.
▪
o Benefits:
▪ Reduces the complexity of software development.
▪ Allows developers to focus on writing code instead of managing
infrastructure.
▪ Scalable and flexible resource allocation.
o Challenges:
▪ Limited control over the underlying infrastructure.
▪ Vendor lock-in concerns.
• Software-as-a-Service (SaaS):
o Delivers software applications over the internet without requiring installation
on local machines.
o Users access software via web browsers on a subscription basis.
o Examples:
▪ Google Workspace (Docs, Sheets, Gmail): Cloud-based productivity
tools.
▪ Microsoft 365: Office applications and collaboration tools.
▪ Salesforce: Customer relationship management (CRM) software.
o Benefits:
▪ Easy access from anywhere with an internet connection.
▪ Automatic updates and maintenance handled by the provider.
▪ Lower upfront costs compared to traditional software licensing.
o Challenges:
o
o Security measures: Includes biometric access control, surveillance, and fire
suppression systems.
•
4.2.2 Data-Center Interconnection Networks
• Purpose: Facilitates communication between servers, storage, and network devices.
• Common Network Topologies:
o Fat-tree architecture: Provides high bandwidth and low latency, used in cloud
data centers.
o Clos network: Ensures efficient routing and redundancy.
o Software-Defined Networking (SDN): Enables dynamic network
configuration and automation.
• Load Balancing: Distributes network traffic evenly to prevent bottlenecks.
• Fault Tolerance: Ensures network resilience through redundant connections.
•
4.2.3 Modular Data Centers in Shipping Containers
• Definition: Prefabricated data centers housed in shipping containers.
• Advantages:
o Portability: Can be transported to different locations as needed.
o Scalability: Easy to add more modules for additional capacity.
o Energy Efficiency: Designed with advanced cooling systems to reduce power
consumption.
• Use Cases: Disaster recovery, military applications, rapid deployment in remote areas.
• Example: Microsoft’s Azure Modular Data Centers, which support cloud workloads in
various locations.
4.2.4 Interconnection of Modular Data Centers
• Techniques Used:
o Fiber-optic networking: Ensures high-speed data transfer between modules.
o SDN integration: Manages dynamic network reconfigurations.
o Data replication strategies: Ensures data consistency across distributed
centers.
• Challenges:
o Latency issues: Distance between modular centers can impact performance.
o Security risks: Data transmission over long distances requires encryption
o
o .
4.2.5 Data-Center Management Issues
Here are basic requirements for managing the resources of a data center. These suggestions
have resulted from the design and operational experiences of many data centers in the IT and
service industries.
• Making common users happy The data center should be designed to provide quality
service to the majority of users for at least 30 years.
• Controlled information flow Information flow should be streamlined. Sustained
services and high availability (HA) are the primary goals.
• Multiuser manageability The system must be managed to support all functions of a data
center, including traffic flow, database updating, and server maintenance.
• Scalability to prepare for database growth The system should allow growth as workload
increases. The storage, processing, I/O, power, and cooling subsystems should be
scalable.
• Reliability in virtualized infrastructure Failover, fault tolerance, and VM live migration
should be integrated to enable recovery of critical applications from failures or
disasters.
• Low cost to both users and providers The cost to users and providers of the cloud system
built over the data centers should be reduced, including all operational costs.
• Security enforcement and data protection Data privacy and security defense
mechanisms must be deployed to protect the data center against network attacks and
system interrupts and to maintain data integrity from user abuses or network attacks.
• Green information technology Saving power consumption and upgrading energy
efficiency are in high demand when designing and operating current and future data
centers.
• Key Challenges:
o Resource allocation: Ensuring optimal usage of compute, storage, and network
resources.
The servers may be physical machines or virtual machines (VMs) depending on the
deployment model.
• Users interact with the cloud via user interfaces, requesting services or resources.
o A provisioning tool automatically allocates and configures: Servers, Software
environments, Database resources
• This process allows for flexibility and scalability in resource management.
• Beyond computing power, the cloud requires distributed storage to handle large
volumes of data.
• A distributed file system is used to manage and access this data efficiently.
o Accompanying services include:Storage Area Networks (SANs), Database
systems, Security devices (e.g., firewalls, intrusion detection)
• These resources are part of the cloud platform and managed collectively.
IaaS enables cloud providers to offer virtual machines and infrastructure components on
demand.
Platform Layer (PaaS - Platform as a Service): Built on top of the infrastructure layer.
Offers a general-purpose software environment that supports:
o Application development
o Workflow testing
o Execution monitoring
Acts as a system middleware between the infrastructure and the application layers.
Provides reusable software libraries, development frameworks, and APIs to simplify
application building. PaaS is ideal for developers who want to build and deploy applications
without managing the underlying hardware.
Application Layer (SaaS - Software as a Service): The topmost layer in the cloud
architecture. Consists of ready-to-use software applications provided as services over the
Internet. Applications in this layer are used for:
o Daily office tasks (document processing, calendars)
o Information retrieval
o Enterprise operations (CRM, sales, supply chain management, finance)
SaaS allows users to access applications without worrying about the underlying platform or
infrastructure.
Layer Interdependence
• The layers are interdependent, and real-world cloud services may utilize multiple layers
simultaneously.
• Applications may not be confined to one layer but span across infrastructure, platform,
and application levels.
• Hybrid clouds are becoming increasingly common, blending private and public cloud
services for flexibility and scalability.
Cloud architecture is modular and layered, supporting different types of services through
IaaS, PaaS, and SaaS. Virtualization and standardization enable dynamic provisioning,
resource pooling, and scalability. Users and providers interact differently with each layer,
and choosing the right model depends on use case, control, and cost requirements. Hybrid
cloud models are favored for applications that need both private control and public scalability.
The provider’s responsibilities and level of management increase from IaaS to SaaS.
Cloud providers allocate resources through Service Level Agreements (SLAs), ensuring CPU,
memory, and bandwidth availability for a preset period. Balancing underprovisioning (risking
SLA violations) and overprovisioning (causing resource waste) is a key challenge. Efficient
provisioning involves VM installation, live migration, and failure recovery. Examples
include Amazon EC2, IBM’s Blue Cloud, and Microsoft Azure, all of which rely on
virtualization technologies.
2. Resource Provisioning Methods
Three primary resource provisioning methods exist:
• Demand-Driven Provisioning: Adjusts resources based on utilization thresholds (e.g.,
Amazon EC2 auto-scaling). It is simple but ineffective for abrupt workload changes.
• Event-Driven Provisioning: Allocates resources based on predicted workload spikes
during specific events (e.g., seasonal sales). This method minimizes Quality of Service
(QoS) loss if predictions are accurate.
• Popularity-Driven Provisioning: Allocates resources based on Internet search
trends. It anticipates traffic surges but may waste resources if popularity forecasts are
incorrect.
Limitations:
• Not effective for abrupt workload changes (e.g., a sudden surge in users).
• Reacts after usage crosses thresholds — not proactive.
Performance (Figure 4.25a):
• Starts with high CPU load fluctuations.
• Gradually stabilizes to maximum 20 VMs, showing effective scaling.
• Suitable for applications with gradual changes in demand.
Event-Driven Resource Provisioning
Concept: Resources are provisioned based on predefined events or time triggers.
Use Cases:
• Seasonal or predictable events (e.g., Christmas shopping, Lunar New Year).
• Preemptive scaling before a known traffic spike.
Advantages:
• Minimal Quality of Service (QoS) loss if the event is well-predicted.
• Ideal for predictable user demand patterns.
Limitations:
• Inaccurate predictions lead to resource wastage or degraded QoS.
• Not useful for unpredictable traffic patterns.
Performance (Figure 4.25b):
• Peaks at 17 VMs, then drops quickly post-event.
• More conservative than demand-driven provisioning.
Popularity-Driven Resource Provisioning
Concept: Provisions resources based on application popularity trends (e.g., search traffic, social
media mentions).
• Uses Internet popularity metrics to estimate likely demand.
• Allocates resources in anticipation of increased usage.
Advantages:
• Can pre-allocate resources for trending apps.
• Helps maintain QoS if popularity trends are accurately predicted.
Limitations:
• Incorrect predictions result in wasted resources.
• Assumes linear correlation between popularity and usage, which may not hold.
Performance (Figure 4.25c):
• Shows fluctuating CPU utilization and VM provisioning.
• Peaks mid-way, aligning with predicted popularity.