CC Module-4
CC Module-4
Introduction
The Internet has grown into a huge source of information, with data centers at its center.
These centers provide services (like Google, Facebook) and resources (like Amazon EC2,
Microsoft Azure). With the rise of cloud computing, data centers are now essential for
businesses. They help manage data and IT infrastructure but are complex and expensive to
maintain.
Data centers are classified into four tiers based on how reliable they are. Tier IV is the most
reliable, designed for critical systems, while Tier I is for smaller businesses. Because of
high costs, small companies often rely on cloud service providers like Google and Amazon.
As cloud services grow, data centers face challenges in cutting costs, saving energy, and
improving performance. This chapter looks at these challenges, especially in data center
networks.
Overview of Data Center Environment
Initially, organizations relied on server rooms to store servers and networking equipment
for LANs. Some had a main server room and a backup in another building for fault
tolerance.
With the rise of client-server computing and the Internet, the concept of data centers
emerged. These centers became vital during the dot-com era, offering high reliability to
support the rapid growth of Internet companies. Over time, data centers became the
backbone of IT, essential for running businesses or serving as the business itself.
Data centers are known for their "mission-critical" role, ensuring maximum uptime. They
are built for reliability at any cost. However, in the past 5–10 years, advancements have
transformed and accelerated innovation in data center technology.
1. Web server
2. Application
3. Database
Another design, the server cluster model, originated in scientific and academic fields and
is now used in industries like finance, manufacturing, and entertainment. It supports high-
performance, parallel, and high-throughput computing, often customized to meet specific
business needs.
These tiers can run on the same machine using interprocess communication or across
separate machines via network connections. Separating tiers enhances resiliency,
allowing servers to be replaced without disrupting operations, and improves security by
isolating compromised web servers from application or database tiers. Typically, web and
application servers may share physical hardware, while databases remain separate.
CEDCs
New technological and economic pressures are forcing organizations to maximize their IT
infrastructure, as the current environment is strained, making it harder to maintain
efficiency and service quality, regardless of the number of users or applications. The
solution lies in cloud computing, which reduces costs and increases flexibility for
applications and IT infrastructure. The Cloud-Enabled Data Center (CEDC) takes
virtualization to the next level, transforming traditional virtualized environments into more
agile systems. It incorporates intelligent, integrated Infrastructure-as-a-Service (IaaS) and
Platform-as-a-Service (PaaS) to manage dynamic workloads, providing resources based on
business policies, while offering automation and coordination of resources across various
data centers. This progression from virtualization management to CEDC addresses
common business goals, improving overall efficiency.
Physical Organization
Data centers can occupy an entire building, a few floors, or even just a single room. They
typically consist of many servers mounted in rack cabinets arranged in rows, creating
aisles for easy access to the front and rear of each cabinet.
Some equipment, like storage devices, can be as large as the racks and are often placed
beside them. In large data centers, which can store thousands of servers, shipping
containers holding 1,000 or more servers may be used. When a failure occurs or upgrades
are needed, entire containers are replaced instead of individual servers.
Multiple applications run inside a data center, typically with each application hosted on its
own set of (potentially virtual) server machines. Each application is associated with one or
more publicly visible and routable IP addresses, which clients on the Internet use to send
requests and receive replies. Inside the data center, requests are spread among a pool of
frontend servers by a specialized load balancer.
A two-tier network topology is commonly used in data center networks today. Access
switches for server connectivity are collapsed into high-density aggregation switches,
which handle the switching and routing for access switching interconnections and various
LAN servers. It offers several benefits:
However, the two-tier design has limited scalability. When the ports on an aggregation
switch pair are fully utilized, adding another pair of aggregation switches introduces
complexity. The connection between aggregation switch pairs must be fully meshed with
high bandwidth to avoid bottlenecks. Since the aggregation switch pair also runs routing
protocols, more switch pairs mean more routing protocol peering and more routing
interfaces, adding complexity to the full mesh design.
Cooling Infrastructure
Since data centers typically span large areas and house thousands of servers, a
sophisticated cooling system is required, often including building-level air-conditioning
units, fans, and air recirculation systems. The server racks are placed on raised platforms
with alternating back-facing and front-facing aisles. Cold air is forced into the front-facing
aisles, and server fans draw the cold air through the servers, pushing the hot air to the
back. This hot air rises and is directed toward the chiller plant for cooling and recirculation.
While this setup is cost-effective, it can create hot spots due to uneven cooling or the
mixing of hot and cold air.
1. Mice traffic: Small query traffic, such as Google searches or Facebook updates. It
makes up most of the traffic in a DCN but involves low data transmission volumes.
2. Cat traffic: Control and coordination messages, such as small- to medium-sized
file downloads.
3. Elephant traffic: Large data transfers, like antivirus updates or movie downloads.
These types of traffic—bursty query traffic, delay-sensitive cat traffic, and throughput-
sensitive elephant traffic—often coexist in DCNs, each with its own application and
performance requirements.
Availability
A major challenge for cloud providers is ensuring maximum uptime. Even short downtime
can harm the organization’s reputation, breach SLAs, and impact revenue. Downtime not
only disrupts user services but also erodes trust, making reliability critical for business
continuity. High availability is typically achieved through strategies like data replication,
load balancing, and regular backups to minimize service interruptions.
Data Center Networks (DCNs) require high burst tolerance, low latency, and high
throughput to handle diverse traffic types: mice, cat, and elephant traffic, each with unique
performance needs. However, traditional TCP/IP, designed for internet scenarios,
struggles to deliver optimal performance in DCNs.
Security
Ensuring data security during transmission and at rest remains a key concern for cloud
providers. Accidental data loss due to power outages or backup failures can result in
significant damage. Additionally, providers must safeguard the physical security of data
center buildings and networking infrastructure to prevent attacks from malicious insiders.
Despite decades of evolution, TCP faces challenges due to the diversity of modern
networks and application demands. These issues expose limitations in TCP’s fundamental
design, particularly in its congestion control mechanisms. This section highlights the key
challenges encountered by state-of-the-art TCP in DCNs.
TCP Incast
TCP incast refers to the issue where TCP fails to fully utilize link capacity in many-to-one
communication patterns, such as those used in web search, MapReduce, and social
networks. This problem is common in large-scale applications and data centers, including
distributed storage and data-intensive computing systems.
As the number of concurrent worker nodes increases, packet losses rise, reducing
throughput. Lost packets are only retransmitted after a delay, affecting time-sensitive
applications like mice traffic. Fast retransmit mechanisms often fail for mice traffic due to
the small data volume and lack of sufficient duplicate acknowledgments.
To address TCP incast, various solutions have been proposed, from modifying TCP itself to
exploring fixes at the application, transport, or link layer. Ren et al. provide an in-depth
analysis of these solutions.
TCP Outcast
TCP outcast occurs when a large set of flows and a small set of flows compete for the
same bottleneck output port in data center switches, causing the small flows to lose their
throughput share. This happens due to drop-tail queues in switches, which lead to
consecutive packet drops and frequent TCP timeouts, resulting in high latency and poor
response times. This issue, known as port blackout, significantly affects the performance
of small flows.
Additionally, TCP’s behavior causes RTT bias, where flows with low RTT get a larger share
of bandwidth than those with high RTT. However, in TCP outcast scenarios, the low RTT
flows are outcasted by the high RTT flows, causing inverse RTT bias.
The two main causes of TCP outcast are the use of drop-tail queues and many-to-one
communication patterns, which are common in data centers.
Mitigating TCP Outcast: To mitigate TCP outcast, alternatives to drop-tail queues can be
used, such as Random Early Detection (RED) or Stochastic Fair Queueing (SFQ). Another
approach is to design efficient TCP congestion control mechanisms at end hosts to reduce
buffer occupancy at switches.
Queue Buildup
In Data Center Networks (DCNs), mice traffic, cat traffic, and elephant traffic coexist.
Elephant traffic, being long-lasting and greedy, often leads to extreme congestion,
overflowing bottleneck buffers. When both mice and elephant traffic share the same route,
the performance of mice traffic is severely impacted.
There are two main ways elephant traffic degrades mice traffic performance:
1. Elephant traffic fills up most of the buffer, increasing the likelihood of packet drops
for mice traffic, similar to the TCP incast issue. Frequent packet losses lead to
timeouts and reduced performance.
2. Even when no packets are lost, mice traffic experiences increased queuing delays
as they wait behind elephant traffic, leading to queue buildup.
Mitigating Queue Buildup: To address queue buildup, the key is to minimize queue
occupancy in DCN switches. Most existing TCP variants use a reactive approach, adjusting
the sending rate only after packet loss occurs. A proactive approach is needed to reduce
queue occupancy and prevent queue buildup.
Buffer Pressure
Buffer pressure occurs due to the greedy and long-lasting nature of elephant traffic. When
both mice and elephant traffic share the same route, elephant traffic occupies most of the
buffer space, leaving little room for the bursty mice traffic. This results in a high number of
packet losses for mice traffic, leading to poor performance. Additionally, since most traffic
in DCNs is bursty, mice traffic packets are frequently dropped because elephant traffic
keeps the buffer space full.
Pseudocongestion Effect
Virtualization, a key technology in cloud computing, allows data centers to run virtual
machines (VMs) to provide on-demand services. While virtualization offers benefits like
efficient server utilization and lower maintenance costs, it can negatively impact
traditional protocols such as TCP and UDP. A study on Amazon EC2 showed that
virtualization reduces TCP and UDP performance, causing instability in throughput and
increased end-to-end delays.
When multiple VMs run on the same physical machine, hypervisor scheduling latency
increases, leading to longer waiting times for each VM to access the processor. This
latency, ranging from microseconds to hundreds of milliseconds, causes unpredictable
network delays, affecting throughput and increasing delays. If the latency is high enough, it
can trigger a Retransmission Timeout (RTO), causing the VM sender to mistakenly assume
network congestion and lower its sending rate. This phenomenon is known as
pseudocongestion.
Introduction
Emerging types include Security as a Service (SeaaS), Knowledge as a Service, and Data
Analytics as a Service.
Major companies like Amazon, Microsoft, Google, and IBM offer a range of cloud tools and
services to benefit both users and businesses.
EMC
EMC is a leading global enterprise that needs dynamic scalability and agile infrastructure
to adapt to changing applications and business demands. Cloud computing was adopted
as the ideal solution to simplify operations and optimize infrastructure. By offering
Information Technology as a Service (ITaaS), EMC also reduces energy consumption
through resource sharing.
EMC IT
Provides secure frameworks for applications and information, offering services like Oracle
Database, SQL Server, and Greenplum as a Service, along with platforms for application
development, content management, information cycle management, security, and
integration.
The Captiva Cloud Toolkit, developed by EMC, is a Software Development Kit (SDK) aimed
at helping developers integrate scanning and imaging capabilities into web-based
business applications. It is tailored for document capture vendors, commercial software
developers, and enterprises looking to create custom, scan-enabled web solutions that
complement their business offerings.
This toolkit enables developers to build scan-enabled applications in as little as one week,
reducing development, testing, and support costs. It also shortens time-to-market,
accelerates return on investment, and enhances competitiveness in the document
capture market.
1.Scan:
This module allows importing documents into Captiva via a scanner at the page level. It
supports various formats like PDF, TIFF, and JPG. Scanning is the entry point for
documents to be processed by Captiva.
Acts as another entry point for importing documents from folders or repositories
such as email attachments. Unlike the scan module, MDW does not interact with
scanners, making it ideal for businesses handling soft copy documents.
This module enhances image quality for smoother processing. Features like deskewing
and noise removal help repair unclear images. Configuration options are available to suit
specific business requirements.
4.Index:
A data capture module that extracts key information (e.g., account numbers) from
documents. It supports adding validation fields to ensure data accuracy during entry.
5.Export:
The exit point where processed images or data are sent to repositories or systems.
Exported data can be utilized for specific business needs, such as mapping captured
details to respective departments.
6.Multi:
This is the final module, used to delete batches that have successfully passed through all
other modules. It can be configured to retain batches if backups are needed.
By combining these modules, the Captiva Cloud Toolkit streamlines document processing,
reduces operational costs, and supports businesses in creating efficient and tailored
solutions for their needs.
Google
Google is a top cloud provider offering secure and reliable data storage solutions. Its
services include the Google Cloud Platform, App Engine, Cloud Print, Cloud Connect, and
more, all designed for scalability and security. Many of these services are free or available
at minimal cost, making them highly accessible and user-friendly.
Cloud Platform
Google Cloud Platform allows developers to build, test, and deploy applications on its
highly scalable and reliable infrastructure. Supported by one of the largest global
networks, it incorporates advanced software infrastructures like MapReduce, BigTable,
and Dremel for industrial innovation.
• Infrastructure: Virtual machines, block storage, NoSQL datastore, and big data
analytics.
• Storage: Offers versatile storage services for easy maintenance and quick data
access.
• Flexibility: Users can choose between a fully managed platform or customizable
virtual machines based on their needs.
• Integration: Simplifies integrating user applications within the cloud platform.
The platform supports automatic scaling to handle demanding workloads, scaling down
during low traffic. Designed to manage spikes in traffic, it operates like Google’s own
products. Managed services like App Engine and Cloud Datastore ensure seamless
autoscaling, and users pay only for what they use.
Cloud Storage
Google Cloud Storage is a RESTful web service for storing and accessing data on Google’s
infrastructure. It uses the REST architectural style, coordinating components, connectors,
and data elements in a distributed system. The service combines Google’s high
performance and scalability with robust security features, ensuring data is protected
through redundant storage across multiple physical locations. Google Cloud Storage is
designed to efficiently handle growing data needs while providing advanced security and
sharing capabilities, making it a reliable choice for users seeking secure and scalable
cloud storage solutions.
The following are the few tools for Google Cloud Storage:
• Google Developers Console is a web application where one can perform simple storage
management tasks on the Google Cloud Storage system.
• gsutil is a Python application that lets the user access Google Cloud Storage from the
command line.
When changes are made to a document, those updates are reflected in real-time for all
users viewing it. If multiple people edit the same section, Cloud Connect allows users to
choose which changes to keep.
When a document is uploaded to Google Cloud Connect, the service inserts metadata into
the file. This metadata helps track changes across all copies, ensuring updates are
properly synced. The backend infrastructure is similar to the Google File System and
leverages Google Docs to sync and distribute changes to all copies using metadata.
Most printers are not cloud-ready, so users often need a computer to act as a bridge.
Google Cloud Print is integrated into the Google Chrome Browser but must be explicitly
enabled. Once activated, a "connector" is used to send print commands from the
computer’s printer software to the printer.
For cloud-ready printers, direct Internet connection is possible without the need for a
computer. These printers must be registered with Google Cloud Print to be functional.
Since Google allows developers to integrate Cloud Print into their apps and websites, the
feature’s availability varies, and not all platforms support it. While Google incorporates
Cloud Print in its own products, its adoption in third-party apps and services may be
limited.
App Engine supports multiple programming languages, including Java, Python, PHP, and
Go. It provides environments tailored for each language, such as Java with the JVM and
servlets, Python with a fast interpreter, and go with native support for compiled code. App
Engine’s infrastructure ensures that applications run quickly, securely, and without
interference from others.
Users pay only for the resources they use, including storage and bandwidth, which are
measured in gigabytes and billed competitively. App Engine offers a free tier, providing up
to 1 GB of storage and enough resources to support about five million page views per
month. Once billing is enabled, users can scale beyond the free limits and only pay for
additional resources consumed.
Amazon Web Services (AWS)
Amazon Web Services (AWS) is a suite of remote computing services, collectively forming
a cloud computing platform provided over the Internet by Amazon.com. Key services
include Amazon Elastic Compute Cloud (EC2), Amazon Simple Queue Service (SQS), and
Amazon S3. Amazon EC2 is a computing service, while SQS and S3 are support services.
EC2 offers scalable computing capacity, allowing businesses to deploy virtual servers
much faster and cheaper than building physical server farms. AWS data centers are spread
globally across locations like Ashburn, Virginia, Dallas/Fort Worth, Los Angeles, Miami,
Newark, New Jersey, Palo Alto, California, Seattle, St. Louis, Amsterdam, Dublin,
Frankfurt, London, Hong Kong, Singapore, and Tokyo. AWS provides a highly reliable and
cost-efficient cloud infrastructure for various business needs.
EC2 instances are created by booting Amazon Machine Images (AMIs), which contain all
the necessary information to launch an instance. Users can interact with EC2 using the
AWS Management Console (GUI) or a web service API that supports protocols like SOAP
and Query Requests. The API provides libraries and resources for multiple programming
languages including Java, PHP, Python, Ruby, Windows, and .NET. The virtual
infrastructure is powered by the Xen hypervisor.
• Instance Types:
❖ Standard instances: Suitable for general applications.
❖ Micro instances: For low-throughput applications.
❖ High-memory instances: Ideal for high-throughput applications.
❖ High-CPU instances: For compute-intensive tasks.
❖ Cluster compute instances: Best for high-performance computing (HPC).
• On-Demand & Reserved Instances:
Instances are billed hourly, and users can reserve instances at discounted rates.
Spot Instances can be obtained by bidding on unused capacity.
• Regions and Availability Zones:
Instances can be placed across multiple locations defined by regions and
availability zones. This ensures fault tolerance and reliability.
• Monitoring and Management:
Instances can be monitored and managed through the AWS Management Console
or API. Amazon CloudWatch provides visibility into resource utilization, operational
performance, and demand patterns (e.g., CPU utilization, disk I/O, and network
traffic).
• Security:
EC2 instances are authenticated using a signature-based protocol with key pairs.
Amazon VPC allows integration with existing IT infrastructure via a virtual private
network (VPN), offering isolated computing resources with customizable security
settings like firewalls and intrusion detection.
• Elastic Load Balancing (ELB):
ELB distributes incoming application traffic across multiple instances based on
metrics such as request count and latency. It ensures fault tolerance and
automatic scaling of resources.
• Persistent Storage:
EC2 instances can be paired with storage solutions like Elastic Block Storage (EBS)
for secure, reliable, and scalable storage or Amazon S3 for highly durable storage of
mission-critical data. EBS volumes can be attached to instances as block devices,
while S3 allows data storage as objects.
Amazon EC2 is a robust and flexible platform for running applications on the cloud, with
capabilities that ensure high availability, scalability, and security.
S3 is highly scalable, allowing users to store large amounts of data and handle fluctuating
data loads with ease. Its reliability ensures that data is always available when needed.
S3 provides a cost-effective solution for storing easily reproducible content or data where
an original copy is stored elsewhere. RRS reduces storage costs by providing lower
redundancy for less critical data.
S3 is well-suited for storing data for analysis and computation. Data stored in S3 can be
transferred to Amazon EC2 for large-scale analytics or computations without incurring
data transfer costs between services.
S3 provides a secure, durable solution for backup and archiving of critical data. The AWS
Import/Export feature enables users to move large amounts of data into and out of AWS
using physical storage devices, ideal for disaster recovery or periodic backups.
S3 offers static website hosting for websites that serve static content, such as HTML files,
images, videos, and JavaScript. It’s a simple and cost-effective way to host content without
the need for a web server.
Amazon S3 is built with performance and security in mind. Data is stored securely and can
be accessed via a highly reliable web interface. It also supports encryption to ensure data
protection.
Amazon S3 is an essential tool for developers and businesses looking to store and manage
large volumes of data, providing flexibility, scalability, and security, along with an easy-to-
use interface.
Amazon Simple Queue Service (SQS)
Amazon SQS is a fast, reliable, scalable, and fully managed message queuing service
offered by AWS. It simplifies the decoupling of the components within a cloud application,
making it cost-effective to transmit data at any volume and throughput. The service
ensures that messages are not lost and does not require other services to always be
available, providing reliability for applications that depend on message-driven
architecture.
As a distributed queue system, Amazon SQS enables web service applications to quickly
and reliably queue messages. These messages can be generated by one component and
consumed by another, making it easy to handle communication between different parts of
an application. A queue serves as a temporary repository for messages awaiting
processing.
Amazon SQS is built on a redundant infrastructure that ensures high availability for sending
and receiving messages. It can scale easily to handle varying volumes of messages.
The service offers built-in access control facilities, ensuring secure interactions with the
queue. Users can define who can send and receive messages, adding an extra layer of
security.
SQS allows for variable-length messages, providing flexibility in the size of the data that
can be transmitted.
• Configurable Queue Settings:
Users can customize the settings of each queue to meet their specific needs, such as
adjusting message retention times and setting visibility timeouts.
SQS guarantees the reliable delivery of messages, even during high traffic volumes or if
there are failures in one part of the system.
Amazon SQS is an essential tool for building distributed, message-driven applications that
require high availability, scalability, and performance without the complexity of managing
message queues manually. It supports a wide range of use cases from simple message
delivery to complex event-driven architectures.
Microsoft
Microsoft IT (MSIT) has made cloud computing the preferred environment for new and
migrated applications. MSIT developed a methodology and best practices to assess and
migrate applications to the cloud, ensuring optimal performance and cost-efficiency.
These insights are shared with other organizations to help them transition smoothly to
cloud computing. By embracing the cloud, Microsoft empowers businesses to achieve
scalability, flexibility, and cost savings while maintaining performance.
Additionally, Microsoft provides tools like Azure to support businesses in their digital
transformation. Azure offers a comprehensive set of cloud services, including compute,
storage, and networking, helping companies build, deploy, and manage applications. With
a focus on security, compliance, and innovation, Microsoft ensures that organizations can
leverage cloud technologies effectively and safely. As the cloud landscape evolves,
Microsoft continues to support businesses with cutting-edge solutions for long-term
growth.
Windows Azure
Windows Azure Cloud Services (web and worker roles/PaaS) offer developers an easy way
to deploy and manage application services while Windows Azure handles the underlying
infrastructure and operating system.
The Migration Assessment Tool (MAT) helps determine the effort required to migrate an
application to Windows Azure by generating a report based on responses to binary
questions. The Windows Azure Pricing Calculator compares an application’s current
operational costs with potential costs on Windows Azure and SQL Azure, helping
businesses evaluate cost-effectiveness.
The Windows Azure Pack for Windows Server provides a collection of Windows Azure
technologies that allow Microsoft customers to create a self-service, multitenant cloud in
their data centers, running on Windows Server 2012 R2 and System Center 2012 R2. This
technology offers an experience similar to the public Windows Azure, enabling businesses
to enhance their infrastructure.
SharePoint
Microsoft SharePoint is a web application platform offering a wide range of collaboration
tools backed by a common technical infrastructure. Integrated with the Microsoft Office
suite, SharePoint provides a user-friendly interface that is accessible to nontechnical
users. It is designed to enable intranet portals, document management, collaboration,
enterprise search, and business intelligence. Additionally, SharePoint offers system
integration, process integration, and workflow automation capabilities. While not a free
tool like Google Cloud Connect, SharePoint’s advanced features set it apart from other
collaboration platforms.
IBM Cloud
IBM is a prominent player in cloud computing, offering a range of solutions through its IBM
SmartCloud. The platform includes IaaS, SaaS, and PaaS, available via public, private, and
hybrid cloud delivery models. IBM’s cloud offerings begin with physical hardware
supporting virtualization, followed by IBM Websphere for application infrastructure. The
management layer uses IBM Tivoli middleware, providing tools for automated provisioning,
monitoring, and usage tracking. Integrated workload tools support cloud-based
collaboration, development, analytics, and security, helping businesses optimize their
cloud operations effectively.
Cloud Models
IBM provides a range of cloud models to suit different business needs. These include fully
private clouds, where companies manage their infrastructure, as well as hybrid clouds,
which combine public and private elements. The flexibility in choosing between virtual
private cloud services and public cloud services allows businesses to scale operations
securely and efficiently. IBM also offers secure private cloud solutions for businesses
concerned about security risks, with tools like IBM Workload Deployer and Cloudburst to
facilitate the building and deployment of private clouds. These customizable options give
enterprises the ability to integrate and manage their workloads in a way that best fits their
requirements.
IBM SmartCloud
IBM SmartCloud is a comprehensive suite of cloud computing products and solutions,
encompassing IaaS, PaaS, SaaS, and BPaaS across public, private, and hybrid cloud
models. It is organized into three key areas: SmartCloud Foundation, SmartCloud
Services, and SmartCloud Solutions. SmartCloud Foundation provides the infrastructure,
hardware, provisioning, and management for private or hybrid clouds. Built on this
foundation, SmartCloud Services includes IaaS, PaaS, and backup services, while
SmartCloud Solutions offers a range of SaaS applications for collaboration, analytics, and
marketing. BPaaS (Business Process as a Service) extends the cloud offering by delivering
business processes through the cloud with elastic scaling and web interfaces.
SAP Labs
SAP Labs is a leading developer of enterprise software that helps businesses manage
operations and customer relations. Known for its dominance in the enterprise applications
market, SAP’s best-known products include SAP ERP (Enterprise Resource Planning), SAP
BW (Business Warehouse), SAP Business Objects, and SAP HANA, an in-memory
computing appliance. Additionally, SAP has expanded its portfolio with Sybase mobile
products. As one of the largest software companies globally, SAP continues to innovate in
the areas of business management and data processing.
This platform enables rapid development and deployment of business and consumer
applications, offering seamless integration with minimal coding requirements, and
supports open programming standards to enhance developer productivity.
A joint service from SAP and VMware supports the transition to a more open and flexible
private cloud platform, leveraging proven virtualization technology to deliver these
benefits.
Salesforce
Salesforce.com is a cloud computing and social enterprise SaaS provider based in San
Francisco. Of its cloud platforms and applications, the company is best known for its
Salesforce CRM product, which is composed of Sales Cloud, Service Cloud, Marketing
Cloud, Force.com, Chatter, and Work.com. In addition to its products and platforms,
Salesforce.com created AppExchange, a custom application building and sharing
platform. The company also has consulting, deployment, and training services.
Sales Cloud
Sales Cloud is a module in Salesforce.com designed to streamline the entire sales
process. It includes features such as Leads, Accounts, Contacts, Contracts,
Opportunities, Products, Pricebooks, Quotes, and Campaigns, with limits applying in
certain cases. It offers functionalities like web-to-lead for online lead capture and
autoresponse rules. Sales Cloud manages contact information and integrates social
media, real-time customer collaboration, and Chatter to enhance customer engagement.
This platform provides up-to-date account information and social insights, all available in
one place, anytime, anywhere. Sales Cloud ensures real-time updates, from contact
details to deal changes and discount approvals. Salesforce built it to be user-friendly,
similar to consumer websites like Amazon, and in the cloud to eliminate the risks and
costs of traditional software. With its open architecture and automatic updates, Sales
Cloud removes the hidden costs and lengthy implementations associated with traditional
CRM software, continuously innovating with mobile, collaboration, and social intelligence
to stay ahead of the competition.
Cloud Servers are managed through the Rackspace Cloud Control Panel (GUI) or via API
using a RESTful interface, allowing users to automate scaling and load balancing. Servers
are equipped with RAID10 disk storage for persistent data storage, and Cloud Servers can
scale automatically to balance loads. Rackspace provides support through a knowledge
base, forums, and live chat, while the API is open-sourced and supports high-level
languages like C++, Java, Python, and Ruby. The platform also offers the option of a load
balancing solution, currently in beta, and provides additional processing power at no extra
cost when available.
VMware
VMware is a leader in virtualization technology and provides enterprise cloud computing
solutions for both private and public clouds, as well as hybrid cloud environments.
VMware's suite of products, including VMware vCloud Director, VMware vCloud
Datacenter Services, VMware vSphere, and VMware vShield, help organizations optimize IT
infrastructure and leverage cloud services efficiently. VMware’s private cloud offerings
enhance operational efficiency by consolidating physical infrastructures, providing greater
standardization, rapid provisioning, and self-service. These solutions lead to cost savings
and improved resource management for enterprises.
VMware vSphere transforms IT infrastructures into virtual storage, compute, and network
resources, offering both infrastructure and application-level services. At the infrastructure
level, it facilitates the operation and management of resources, while at the application
level, it ensures secure, scalable, and available applications. VMware vCloud Director,
used with VMware vSphere, enables enterprises to build secure, multitenant private
clouds by pooling resources into virtual datacenters, which can be accessed through
automated, catalog-based services. VMware vShield provides security services like
perimeter protection, firewalls, and VPN to safeguard these environments.
VMware also offers public and hybrid cloud solutions through partnerships with certified
service providers, enabling users to extend their private cloud into a public cloud
environment. VMware vCloud Datacenter Services and VMware vCloud Express provide
scalable, reliable, and pay-as-you-go infrastructure, allowing users to augment their
internal resources with external cloud resources. VMware’s vCloud API enables
automated orchestration of operational tasks and workflows, further enhancing the
flexibility and efficiency of VMware’s cloud solutions.
Manjrasoft
Manjrasoft is a lesser-known provider of cloud services, but it offers a platform called
Aneka that simplifies the development of applications. Aneka provides a set of services
designed to accelerate the development, deployment, and management of applications,
helping businesses save time and money. The platform is market-oriented and aims to
enhance business productivity and profitability by providing efficient cloud computing
solutions tailored to meet business needs. Manjrasoft's cloud services focus on building
scalable and high-performance applications, ultimately driving business growth.
Aneka Platform
The Aneka platform simplifies enterprise cloud construction and application development
while maintaining flexibility, scalability, reliability, and extensibility. Key features of Aneka
include:
Aneka enables servers and desktop PCs to be linked together, forming a powerful
computing infrastructure. This approach allows businesses to be more energy-efficient
and cost-effective without the need for extensive hardware investments. Each Aneka node
includes configurable containers for information indexing, scheduling, execution, and
storage services, supporting various programming models, security protocols, persistence
options, and communication protocols.
Summary
This chapter covered various companies offering cloud computing tools and technologies
to help users adapt to the cloud environment. It highlighted the cloud features provided by
each company, with some services like Google Docs and Google Cloud Print being free,
while others, like AWS and Microsoft, are proprietary.A summary of tools and services
offered by each company is provided in Table 11.1, and Table 11.2 outlines pricing, service
models, and deployment models for selected providers. Additional details can be found in
the references section.