Building A Modern Data Center Principles and Strategies of Design (PDFDrive)
Building A Modern Data Center Principles and Strategies of Design (PDFDrive)
Written by
Scott D. Lowe, James Green, & David Davis
Partners, ActualTech Media
Building a Modern Data Center
Principles & Strategies of Design
Author:
Scott D. Lowe, ActualTech Media, James Green, ActualTech Media,
David Davis, ActualTech Media Editor:
Hilary Kirchner, Dream Write Creative Cover Design:
Atlantis Computing
Layout:
Braeden Black, Avalon Media Productions Project Manager: Geordie
Carswell, ActualTech Media Ebook Conversion: Mark D'Antoni, eBook
DesignWorks
Copyright © 2016 by Atlantis Computing All rights reserved. This book or any
portion there of may not be reproduced or used in any manner whatsoever
without the express written permission of the publisher except for the use of
brief quotations in a book review.
Printed in the United States of America First Printing, 2016
ISBN 978-1-943952-07-6
Foreword
Chapter 1 — Introduction
IT Is Changing . . . and It Must
Simplification Is the New Black
Focus on Cost Reduction
Focus on Customer Service and the Business
Adapt or Die
Flattening of the IT Organization
IT as an Operational Expense
What’s Next?
Chapter 10 — IT Transformation
The Challenge
The Solutions
Automation
Orchestration
DevOps
Industry Needs
Open Standards and Interoperability
Transparency on Pricing and Cost
Performance Benchmarking Standards
Chapter Recap
Leading 3rd party IT industry influencers Scott D. Lowe, David M. Davis and
special technical partners cover hot topics from the software defined data center
to hyperconvergence and virtualization.
Cutting through the hype, noise and claims around new data center technologies
isn’t easy, but ActualTech Media helps find the signal in the noise. Analysis,
authorship and events produced by ActualTech Media provide an essential piece
of the technology evaluation puzzle.
Visit www.AtlantisComputing.com
Foreword
For too long, IT has been thought of as a cost center, where the increased
demands and shrinking budgets put it at odds with being responsive to new
business requirements. The role of the CIO and IT as a whole has been in a state
of flux – outsourcing, cloud computing (often in the form of stealth IT), and a
heavy focus on reducing costs typically meant that infrastructure is aging and
internal resources spend most of the time fighting fires and optimizing
environments, at the cost of any new initiatives or enhancements. Now is the
time for IT to take its place as a critical driver to business success by creating a
modern data center. The authors of this book have real life experience as both IT
practitioners and years of educating IT communities on transformational
technologies.
Introduction
The modern data center is an exciting place, and it looks nothing like the data
center of only 10 years past. The IT industry and the world in general are
changing at an exponential pace. In fact, according to Moore’s Law (named after
the co-founder of Intel, Gordon Moore), computing power doubles every few
years. Due to the limitations of the laws of physics, this pace has slowed a bit
over the past decade, but processing power still doubles every 2.5 to 3 years.
This means that, in the modern data center, tasks that were accepted as
unattainable only a few years ago are commonplace today. One practical
example that will be discussed many times in this book is leveraging software
defined storage (SDS) to serve I/O to Tier 1 applications. This model would
have likely been cost-prohibitive and really slow just a few years back. But
today, SDS offers IT organizations some of the greatest storage flexibility and
ease of use they’ve ever known.
IT Is Changing . . . and It Must
With this exponential growth, there are more opportunities for entrepreneurs to
make a killing, for organizations to multiply their revenues tenfold by leveraging
emerging technology, and for top-tier graduate students to create projects that
once seemed like pure science fiction. While bleeding edge technology trickles
down into the mainstream IT organization, the changing nature of the IT
business itself rivals the pace of technological change.
Over the next couple of years, IT organizations and business leaders will have
the opportunity to focus on dramatic transformations in the way they conduct
themselves. Technological change will create completely new ways of doing
business. Those who learn and embrace these new developments will thrive —
those who don’t, will struggle and possibly fail. The exponential growth in
technology in the next decade will generate the rise of entirely new industries
and cause everything in our day-to-day experience to be different, from the way
we shop to the way we drive our cars.
Simplification Is the New Black
Thanks to the ever-growing number of devices interconnected on private and
public networks, and thanks to the Internet, the scope of IT’s responsibility
continues to expand.
For example, not so many years ago, a printing press would have been operated
and maintained by a number of highly specialized printing press operators; IT
would have had nothing to do with the print shop, except to provide the
computers on which the projects were designed. Today, however, it is not at all
unusual that most of the printing press staff are gone, and the press itself is
primarily controlled by a computer. The computer is responsible for operating
the press at maximum efficiency, and sometimes it does so well that the print
shop doubles its profits and the computer becomes critical to the business. All of
a sudden, IT is in the bulk-printing business.
This printing press example is just part of the boom of devices, such as sensors,
lights, actuators, and medical devices that are being networked and exchanging
data. This boom was dubbed the “Internet of Things” (IoT) way back in 1999,
believe it or not! Today, we’re still right at the beginning of this paradigm shift.
The estimations of IoT proliferation in the next few years are staggering. A 2014
Gartner report shows an estimated 25 trillion connected devices by 2020.
Because this trend means that IT departments become responsible for more
systems and endpoints, one of the primary goals for these departments in the
near future is to simplify administration of their systems.
Figure 1-1: The IT Budget Gap
Despite the fact that the number of integrations and the amount of data IT has to
manage because of these devices is increasing, in many industries budgets are
not. (See Figure 1-1) This means that, in order to succeed, CIOs will need to
find ways to boost efficiency in major ways. Five years from now, the same
team of administrators may be managing 10 times the number of resources that
they’re managing today, and expectations for stability, performance and
availability will continue to increase.
There are two main venues where the push for simplicity must come from: the
manufacturers and the IT executives. First of all, manufacturers must begin to
design and redesign their products with administrative simplicity in mind. To be
clear, this doesn’t mean the product must be simple. In fact, the product will
almost assuredly be even more complex. However, the products must have the
intelligence to self-configure, solve problems, and make decisions so that the
administrator doesn’t have to.
Secondly, IT Directors and CIOs must survey their entire organization and
ruthlessly eliminate complexity. In order to scale to meet the needs of the future,
all productivity-sapping complexity must be replaced with elegant simplicity
such that the IT staff can spend time on valuable work, rather than on putting out
fires and troubleshooting mysterious failures that take hours to resolve. One of
the primary ways this might be done is to eliminate the organizational siloes that
prevent communication and collaboration.
Focus on Cost Reduction
As shown in Figure 1-2, a recent Gartner report shows that IT budgets are
predicted to remain nearly flat as far out as 2020. This is despite 56% of
respondents reporting that overall revenues are expected to increase. The reality
is that many of those IT organizations will be required to do more with less, or at
the very least, do more without any increase in budget. As this trend isn’t likely
to change, the IT department of the future will be focused on reducing expenses
where possible and increasing control over the absolutely necessary expenses.
24.6% of survey respondents have four or less dedicated IT staff supporting their
business. Without budget increases, this same small handful of people will be
expected to continue supporting the organization’s increasing needs.
Many leading manufacturers are leveraging increased connectivity and big data
analytics to preemptively solve their customers’ problems before they have
them. After all, the best support call an administrator could ask for is the one that
never has to be made. Plus, in the coming years, CIOs and IT Directors will look
to shift more of the burden of maintaining infrastructure and software resources
to their partners and vendors, while they look to their own IT staff to assume a
role more focused on innovation and providing value to the business.
This means that shadow IT poses significant security, compliance, and control
risk to the entire business, and the only way to really stop it is to serve the
internal customers so well that they don’t need to look elsewhere for their
technical needs. Shadow IT is a term used to describe business units (or
individuals) other than IT who deploy technical solutions — not sanctioned or
controlled by IT — to solve their business problems. A simple example of this
phenomenon is a few individuals in the finance department using personal
Dropbox folders to share files while the business’ chosen direction is to share
company files in SharePoint.
Adapt or Die
The world of IT operates in the same way as the rest of the world: things change
over time, and those companies, technologies, or individuals who choose not to
keep up with the current state of the industry get left behind. It’s unfortunate, but
it is reality. The movie rental giant Blockbuster was dominating the market until
Netflix and Redbox innovated and Blockbuster failed to adapt. Blockbuster
eventually went bankrupt, and is now an afterthought in the movie consumption
industry while Netflix’s fortunes are at an all-time high.
This lifecycle is exactly the same in IT; there are household names in the
industry that are more or less irrelevant (or quickly becoming so) at this point
because of their failure to adapt to the changing market. Unfortunate as it may
be, this is also happening at the individual level. As IT administrators and
architects adapt or do not adapt over the course of time, they either advance in
their organization and career or they become irrelevant.
Flattening of the IT Organization
A decade ago, IT administrators prided themselves on their silos. The
stereotypical gray-ponytailed storage administrators were the pride and joy of a
thriving IT department. However, that organizational mentality has come and
gone, and the IT department of the future is becoming much more tightly
integrated and highly generalized.
With the ongoing need to do more with less, the most lucrative jobs will be the
roles whose responsibilities span a wide variety of technical specializations. The
industry will place great value on highly proficient generalists as compared to
specialists.
One of the primary disruptions to the industry that’s driving this dissolution of
silos was the mainstream adoption of x86 virtualization. As organizations across
the globe have consolidated rooms full of physical servers down into a few
racks, the supporting technologies have necessarily come closer together as well.
Since cloud resources can be billed just like a monthly phone bill, shifting IT
resources to the cloud also shifts the way the budget for those resources is
allocated. While buying a pile of servers and network equipment to complete
projects over the next year is budgeted as a capital expenditure, the
organization’s “cloud bill” will be slotted as an operational expenditure. This is
because, with a capital purchase, once the equipment is purchased, it is owned
and depreciating in value from the moment it hits the loading dock. This
removes an element of control from IT executives as compared to an operational
expense.
If a SaaS application is billed per user on a monthly basis, there’s no need to pay
for licenses now to accommodate growth in headcount six months down the
road. It can also be in use this month and cancelled next month. This is in
contrast to the IT director who can’t just “cancel” the stack of servers purchased
six months ago because the project got cancelled. Due to these advantages from
a budgeting and control standpoint, products and services offering a model that
will require little or no capital expense and allow budgeting as an operational
expense will be preferred.
The survey didn’t explore this, but it’s likely that a good number of those looking to
adopt cloud-based storage services are interested in doing so because of the
operational expense associated with the model. They’re tired of the storage hardware
and maintenance refresh cycle.
What this means for manufacturers is that transparency, granular control, and
offering a OpEx-based model like renting or leasing, billing based on monthly
usage, and expanding in small, predictable chunks based on need, will position
them for adoption by the IT department of the future. Also, offering insight and
helping the customer to increase efficiency and accuracy over the course of the
billing cycle will create lasting loyalty.
This shifting focus from CapEx to OpEx is also giving rise to new kinds of data
center architectures that allow organizations to keep data centers on premises
and private, but that enable some economic aspects similar to cloud. Rather than
having to overbuy storage, for example, companies can begin to adopt software
defined storage (SDS) or hyperconverged infrastructure (HCI) solutions that
enable pay-as-you-grow adoption methodologies.
What’s Next?
The future always holds a lot of promise, but it also creates trepidation as people
struggle to understand emerging offerings and how they personally fit in to the
new order. In this book, you will learn about the trends that will shape modern
data centers over the next decade and how you can seize the opportunities that
come with this transformation.
2
This happens in all kinds of areas as technology matures. The cyclical nature
doesn’t mean that progress isn’t being made, however. Each time the cycle starts
a new iteration, it’s substantially improved from the last iteration. Before we get
into the state of IT today, and ultimately where it is headed, let’s go on a bit of a
journey to the past.
A History of the Modern Data Center
The first few decades in the life of the room that eventually became known as
the “data center” were characterized by electromechanical computers made from
electrical switches and mechanical relays, and later by all electronic computers
that used vacuum tubes as switches.
The innovation responsible for the data center as we know it today was the
transistorized, integrated circuit based microprocessor. Maturity in this
technology eventually led to Intel’s 8086 chip, and all of its successors. The x86
instruction set lives on today, and is the foundation of many components of the
modern data center. Although none of today’s modern processors has an “86” in
its name at all, the name “x86” comes from the 8086 and its successors, such as
the 80186, 80286, and so on.
As computing technology developed, so did the ability to store the data that was
being manipulated and recorded. Tape-based data storage technology began to
be displaced when IBM released the first disk-based storage unit in 1956 ( ). It
was capable of storing a whopping 3.75 megabytes — paltry by today’s terabyte
standards. It weighed over a ton, was moved by forklift, and was delivered by
cargo plane. Magnetic, spinning disks continue to increase in capacity to this
day, although the form factor and rotation speed have been fairly static in recent
years. The last time a new rotational speed was introduced was in 2000 when
Seagate introduced the 15,000 RPM Cheetah drive. CPU clock speed and
density has increased many times over since then.
Figure 2-1: The spinning disk system for the IBM RAMAC 305
The Rise of the Monolithic Storage Array
The inefficiency at scale actually had two components. The first is that servers
very commonly only used a fraction of the computing power they had available.
It would have been totally normal at this time to see a server that regularly ran at
10% CPU utilization, thus wasting massive amounts of resources. (The solution
to this problem will be discussed in the next section.) The second problem was
that data storage utilization had the same utilization issue. With the many, many
islands of storage created by placing direct attached storage with every server,
there came a great inefficiency caused by the need to allow room for growth.
As an example, imagine that an enterprise had 800 servers in their data center. If
each of those servers had 60 GB of unused storage capacity to allow for growth.
That would mean there was 48 TB of unused capacity across the organization.
Using the lens of today’s data center to look at this problem, paying for 48 TB of
capacity to just sit on the shelf seems absurd, but until this problem could be
solved, that was the accepted design (Figure 2-2).
Block level access, on the other hand, sends SCSI commands directly from the initiator (client side) to the
target (storage array side).
The determining factor in which method is used is the storage protocol. Examples of storage protocols are:
Fibre Channel (block), iSCSI (block), NFS (file), and SMB (file). Each type of storage protocol and the
resulting access granularity has a use case. File-based protocols may be used where an end user or
application will be accessing the files directly — a network share, for example. Block-based protocols are
more likely to be used when an operating system or hypervisor is accessing the storage, as direct access to
the disk is preferred.
Block-based storage can be formatted by the client with a filesystem like NTFS of VMFS, whereas a file-
based volume is already formatted by the storage platform.
Data Services
Most storage platforms come with a variety of different data services that allow the administrator to
manipulate and protect the stored data. These are a few of the most common.
Snapshots
A storage snapshot is a storage feature that allows an administrator to capture the state and contents of a
volume or object at a certain point in time. A snapshot can be used later to revert to the previous state.
Snapshots are also sometimes copied offsite to help with recovery from site-level disasters.
Replication
Replication is a storage feature that allows an administrator to copy a duplicate of a data set to another
system. Replication is most commonly a data protection method; copies of data a replicated offsite and
available for restore in the event of a disaster. Replication can also have other uses, however, like
replicating production data to a testing environment.
Data Reduction
Especially in enterprise environments, there is generally a large amount of duplicate data. Virtualization
compounds this issue by allowing administrators to very simply deploy tens to thousands of identical
operating systems. Many storage platforms are capable of compression and deduplication, which both
involve removing duplicate bits of data. The difference between the two is scope. Compression happens to a
single file or object, whereas deduplication happens across an entire data set. By removing duplicate data,
often only a fraction of the initial data must be stored.
This incorrect usage of the term SAN continues to this day, and is so common
that it’s accepted nomenclature by all but the most academic storage
administrators.
As the industry matured and more organizations adopted a shared storage model,
the value of the architecture continued to increase. Manufacturers added features
to the management platforms of the storage arrays to allow operations like
storage snapshots, replication, and data reduction. Again, rather than 800 places
to manage file system snapshots, administrators could make use of volume-level
snapshots from just a few (or even one) management console. This created new
possibilities for backup and recovery solutions to complete backups faster and
more efficiently. Storage systems also contained mechanisms for replicating data
from one storage array to another. This meant that a second copy of the data
could be kept up-to-date in a safe location, as opposed to backing up and
restoring data all the time.
Perhaps one of the greatest efficiencies achieved by adopting the shared storage
model was potential for global deduplication of data across the enterprise. Even
if deduplication was available in the Direct Attached Storage (DAS) model,
deduplicating 800 silos of data individually would not result in high
consolidation ratios. However, deduplicating data across all 800 systems that
were likely similar would result in much higher consolidation.
By the mid-2000s, average data centers had the efficiency of using shared
storage across servers and applications, combined with the added efficiency of
being able to globally deduplicate that data. Performance of the shared storage
systems grew as manufacturers continued to improve the networking protocols,
the physical disk media, and the file systems that governed the storage array.
Due to its size and scope in many organizations, managing the storage network
and the storage arrays became a job for entire teams of people, each with highly
specialized skill sets.
Using shared storage allowed more agility and flexibility with servers than was
known with direct-attached storage. During this time, many organizations chose
to provision the operating system disk for a server on the storage array and use a
“boot from SAN” model. The benefit of deploying operating systems this way
was this: if one physical server failed, a new server could replace it, be mapped
to the same boot volume, and the same operating system instance and
applications could be back up and running in no time.
Blade server form factors became popular in this era as well. Blade servers have
a smaller footprint because of their small number of drives (if any); this allows
higher density per rack unit.
As effective as all of this consolidation was at driving down costs in the data
center, there was still the problem of compute resources. CPU and memory
resources were still generally configured far above the actual utilization of the
application the server was built for. Eliminating this problem was the second
frontier in solving inefficiency in the modern data center.
The Virtualization of Compute — Software Defined Servers
Virtualization as a concept is not a new development. Virtualization has been
around since the 1960s when the technology was developed to allow multiple
jobs to run simultaneously on a mainframe. This was in contrast to the prior
capability of running a single batch process at a given time. Virtualization
allowed for multiple workloads to run in tandem on shared hardware, yet be
isolated from one another. As mainframes gave way to microcomputers and PCs,
virtualization as a technology became less important, at least for a time.
In the late 1980s, as different companies struggled to control the PC market, end
users found themselves in a bit of a pickle. Certain applications would be
designed only for one platform. If a user owned a Unix-based computer and
wanted to run a Microsoft DOS program, they were out of luck. That is, until a
company released a technology that allowed the virtualization of the application
developed for one operating system to run on an operating system it was not
developed for. For approximately 10 years, this technology matured on desktop
computers.
That was great for the desktop, but the true power of modern virtualization came
in 2001, when VMware released ESX, a bare-metal hypervisor capable of
virtualizing server workloads in the data center. The hypervisor, a term used to
describe the software that abstracts physical resources like CPU and memory
from the virtual machines, fulfilled the same purpose as the virtualization
technology developed for mainframes: running multiple workloads
simultaneously and effectively isolated from one another.
Many people were skeptical of this idea at first, but data centers had run into
some significant issues as they grew. There’s the capacity issue which has been
discussed. Further, servers often used on a tiny fraction of the CPU and memory
resources allocated. But there were also environmental issues; the electricity and
cooling bills were growing, and floor space was becoming an increasingly scarce
commodity. CIOs could see that the problem was only getting worse, and server
virtualization had the potential to solve this issue.
As the market watched and waited and allowed x86 virtualization to mature, the
footprint of physical servers in the data center continued to grow to outrageous
numbers and consumed equally outrageous quantities of electricity. An EPA
study estimated that the data center industry in 2006 consumed 61 billion
kilowatt hours of electricity, which was 1.5% of the total U.S. electricity
consumption that year. The total cost for this energy would have been around 4.5
billion dollars. [1]
The electric bill for the data center was also reduced. If physical servers could be
consolidated at the rate of 4 virtual machines to 1 physical machine (a 4:1
consolidation ratio), then the data center could power off 3 out of 4 physical
servers — a huge reduction in the overall power consumption.
Also, if only 1 out of 4 physical servers was running, the cooling and battery
backup systems didn’t need to be nearly as robust. The physical footprint of the
data center was reduced. That reduction was especially economically important
in co-location scenarios where square footage is a premium.
Another example is the amount of copies of duplicate data that can be represented by only one copy of the
data.
In both cases, the ratio will be expressed as [consolidated amount]:1. For example, 4:1 vCPUs to physical
cores would indicate in a P2V project that for every 1 physical CPU core available, 4 vCPUs are allocated
and performing to expectations.
Once there were fewer physical servers to manage, the administration of said
servers became easier as well. Administrators needed to apply a fraction of the
firmware updates, perform a fraction of the hardware upgrades during a refresh
cycle, and repair a fraction of the motherboard failures.
The impact of virtualization changed networking as well, all the way down to the
physical connections. Where there may have once been two data cables for every
server in the environment, in a post-virtualization data center there are perhaps
two data cables per hypervisor with many virtual machines utilizing those
physical links. This creates a favorable level oversubscription of the network
links as waste from the legacy model. See Figure 2-4 for an example of this
consolidation.
Figure 2-4: Pre-vs. Post-virtualization network consolidation
Hypervisor and virtual machine performance increased, and with it the demands
on related infrastructure components. Driving virtual machine density
necessitated higher bandwidth networking to allow for the high quantity of
traffic sharing a single pipe. It also required higher disk performance and lower
latency due to the virtual machines sharing the same storage path.
By now, administrators have had time to get used to server virtualization. Roughly
95% of survey respondents indicated that they have either “some” or “expert-level”
knowledge of server virtualization. Only 5% said they have “little or no” server
virtualization knowledge.
This need for better performance led to absurd practices like “short stroking”
(the practice of only utilizing the outer edge of a spinning disk and wasting the
rest of the capacity) and buying disks solely for the performance impact that
extra disks have on a storage array even though capacity was not needed.
Clearly, the x86 virtualization movement called for serious innovation in the
data storage market.
The No-Spin Zone: The Move from Disk to Flash
Magnetic storage media has been the dominant choice for data storage for the
majority of data center history. Spinning disks have served as primary storage,
and tape-based storage systems have served higher capacity longer term storage
needs. However, the performance of spinning disk eventually leveled off due to
physics-induced limitations. The speed by which data on a spinning disk can be
accessed is based upon a few factors, but the one that is the biggest problem is
the rotation speed of the disk platter. Eventually, the platter can’t be spun any
faster without damaging it.
Based on what the storage industry has produced in the past few years, it would
appear that 15,000 rotations per minute (15K RPM) is the fastest speed that
manufacturers have been able to maintain while keeping the disk economically
viable to the customer. A 15K SAS drive is a high-performing disk, to be sure.
However, the number of I/O operations that any spinning disk can perform in a
second doesn’t seem to be changing all that much. The fastest, most efficient
spinning disks can deliver less than 200 random I/O per second (IOPS). While
this is beyond adequate for a use case like a PC, it has left something to be
desired when serving I/O to a dense, mixed workload virtual server or virtual
desktop environment. The numbers get even trickier when RAID write penalties
are factored in; depending on the RAID configuration a number of disks may be
needed to achieve 200 IOPS rather than just one.
There’s also the issue of latency. Due to the mechanical nature of a spinning disk
drive, latency (the time it takes to retrieve or write the data in question) can’t be
pushed below a certain threshold. Tiny bits of latency added together across
many drives becomes an issue at scale.
The solution to both the IOPS problem and the latency problem is found in flash
storage. In short, flash storage media makes use of non-volatile memory to store
data as opposed to magnetic platters.
Although the use of flash storage was initially troublesome due to durability
issues, the performance has always been quite attractive and often worth the risk.
Because flash storage is not mechanical in nature, it doesn’t suffer from the same
limitations as spinning disks. Flash storage is capable of latency on the order of
microseconds as opposed to spinning disk’s multiple milliseconds. It’s also
capable of far more I/O operations per second than a handful of spinning disks.
The issue of durability has been solved over time as manufacturers improve the
physical memory, storage controllers use intelligence like wear leveling, and
different types of flash cells are developed, like single level cell (SLC), multi-
level cell and enterprise-grade multi-level cell (MLC/eMLC), and triple level
cell (TLC). A typical eMLC drive on the market in 2015 is warrantied for 10 full
writes per day over a period of 5 years. Alternatively, some manufacturers
specify simply a total amount of data written. The same eMLC drive would
probably be warrantied for something like 3.5 PB of data written.
Lastly, because of the non-mechanical (or “solid state”) nature of flash storage, it
requires much less power to operate when compared to spinning disk. As data
center power bills have always run high, any way to reduce power consumption
is attractive to the data center manager — and the CFO! In some countries,
governments offer substantial incentives for making environmentally friendly
changes like reducing power consumption. In some cases, purchasing boatloads
of flash storage to reduce power consumption may be cheaper than the cost of
the fine for failure to comply.
According to survey results, 38.5% of respondents have 10% or more of their data
center storage provided by solid state disks. Another 40% serve between 1% and 10%
of their storage needs with SSD. This means that only 21.6% of respondents aren’t
making use of at least some flash storage.
Flash storage becoming widely available has been a huge win for the data center
industry. It allows much higher performance with substantially less power, and
in the future flash storage will be available at a cost per gigabyte comparable to
that of spinning disk. Exactly how long this will take is anyone’s guess, but
industry experts predict it could take until at least 2020. This maturing of flash
storage has led data center architects to reconsider the way storage is accessed in
the data center yet again. Just as the utilization and management issues of direct-
attached storage gave birth to the monolithic storage array, performance issues
and power/environmental concerns have birthed a new storage design. The data
center of the future will likely see less of the monolithic storage array in favor of
a return to direct attached storage . . . but with a twist.
The Fall of the Monolithic Storage Array
Monolithic storage arrays solved many of the data center’s problems and
allowed IT to achieve greater efficiencies and scale. Unfortunately, the things
that made this architecture so attractive also eventually became its downfall. The
virtualization of compute led to densities and performance requirements that
storage arrays have struggled to keep up with ever since.
One idea is the concept of server-side caching. This design is less radical than
others, because it continues to make use of existing shared storage. One of the
first really well done implementations of this technology in the enterprise was a
solution that used a virtual machine to consume local DRAM and use it as a
high-speed cache. Another early option was a very expensive but high-
performing PCIe SSD that accelerated remote storage with local caching. These
designs solved common problems like boot storms in VDI environments,
because the virtual machines on each host were able to retrieve hot blocks from
local DRAM before ever traversing the network. This technology was mimicked
and improved on, and today a number of options exist for caching with local
SSDs and DRAM in front of shared storage arrays.
The resilience could be thought of like a network RAID, although it’s more
complex than that. The performance and scale implications of this model are
massive: because each node added to the cluster with local storage contributes to
the pool, this means that the storage pool can grow to virtually limitless heights.
Each server that is added has its own storage controller, meaning that throughput
never becomes an issue. Increasing capacity of the pool is as easy as adding
disks to existing servers or adding more servers overall. The control of all of this
is done by either virtual machines (VSAs) or by kernel-level software, and the
administrator typically manages it from the hypervisor’s existing management
interface (like vCenter or SCVMM).
SDS is changing the data center in tangible ways, and as more organizations
begin to adopt this architecture, vendors of monolithic storage arrays will have
to innovate or pivot in order to stay relevant and survive.
A deep dive on SDS is coming in a later chapter, but the stage isn’t fully set yet.
There’s a lot of moving pieces in the data center and it can get a bit
overwhelming at times. Plus, it moves so fast. Wouldn’t it be nice if the design
and building of it was left to someone else, and what showed up on the loading
dock was ready to use rather than ready to begin a 6-month project?
The Emergence of Convergence
As the challenges for IT have grown in equal proportions with the ever-
increasing scope of their responsibilities, IT decision makers have often looked
to outsource parts of their operation. A notable trend for data center
“outsourcing” of sorts is now referred to as convergence. Put simply,
convergence is multiple pieces of the infrastructure assembled prior to delivery
to the customer. Convergence saves time and frustration during the deployment
phase and provides decreased time-to-value after procurement.
The value in convergence comes not only from the fact that the solution comes
pre-assembled, but also from the fact that it includes all the pieces necessary.
Half the challenge in traditional piecemeal solution-building is getting all the
right parts and ensuring interoperability. Convergence guarantees that with the
purchase of a certain SKU, all the components contained within it will be
compatible with one another, and all the necessary parts will be included.
Now, what’s a technology designed for cloud data centers like Facebook or
Amazon doing in a small corporate data center? It turns out that the cloud isn’t
just for the top 1% anymore. Cloud technology is being leveraged all over the
world by even small companies.
The Role of Cloud
The term cloud has always been a bit confusing and hard to nail down.
Unfortunately, many misconceptions exist about exactly what “the cloud” is, but
in the most general sense, the cloud is pretty easy to grasp.
The Cloud
Keep in mind that the cloud is simply a method of offering and provisioning on-
demand services. With this definition in mind, it’s easy to see that a private cloud
deployment is simply an on-premises deployment of a tool like OpenStack that allows
for rapid, on-demand provisioning of resources that can easily be created and
destroyed.
But why does anyone do this anyway? What is the value in cloud-based solutions as opposed to the way
virtual infrastructure has been deployed for the previous decade? The Cloud Drivers section answers these
questions.
Cloud Types
Different cloud deployment models fit different organizations. There are certain
cases where an application has been developed from the ground up to run in a
cloud. In this case, it may make sense to use a public cloud model, where all
resources are provisioned in a third party data center provided by the likes of
AWS, Microsoft, VMware, Google, or your friendly neighborhood cloud
provider. Especially for some small businesses, being entirely public-cloud-
based allows for an extremely light IT footprint in the office or storefront,
resulting in less overhead.
Public cloud can be very affordable. It also offloads risk and overhead in terms
of compliance, patching, equipment failure, hardware refreshes, and so on.
When purchasing cloud resources, you’re purchasing a service and do not care
what happens on the back end. If it works for the organization, an exclusively
public cloud model is great — but it doesn’t work for everyone.
An example of how hybrid cloud might work is that of a retailer. If Black Friday
is coming up, the retailer may be able to spin up an extra 20 instances of their
website and shopping cart application in the public cloud. The back end
databases still exist in the on-premises data center and need not be migrated.
This is commonly referred to as “bursting” to the cloud.
Another example where a hybrid cloud model could work out well is in an
organization that has a heavy in-house development workload. If developers are
constantly creating and destroying test environments, it can require lots of
horsepower to keep things running fast enough that developers are happy, and
project scopes can change with a moment’s notice. A much easier way to handle
this situation would be to run production workloads in the on-premises data
center, but have development and testing workloads provision to the public
cloud. This can also save on cost as opposed to running the resources locally.
The third option for cloud deployment models is a private cloud. This phrase can
be rather confusing if one thinks of “cloud” as a third party selling services on
the Internet, or worse yet, if one thinks the Internet itself is a cloud.
Cloud Drivers
Although virtualization revolutionized the data center, expectations for
performance, availability, and cost never cease to change the way things must be
done. Eventually, the increased speed and reduced cost of virtualization wasn’t
enough anymore. There are a few things driving the adoption of cloud models,
currently; as with most decisions, it ultimately comes back to the business and
the bottom line.
Therefore, the first driver for cloud models (public, hybrid, or private alike) is
agility. By nature, any sort of cloud model will dramatically increase the speed
of the software development lifecycle.
A third driver is scale. By leveraging a public cloud provider, the scale to which
an organization’s systems can grow is practically limitless. Where physical
space and limitations on power, cooling, and other factors make scale a
challenge when hosting the entire infrastructure on-premises, the public cloud
makes scaling a breeze.
Roughly 10% of all survey respondents said that they’re managing over a petabyte of
storage at their primary site. Storage at this scale alone is enough reason for some
organizations to look for cloud-based alternatives.
The capacity and rotational speed of spinning disk grew quickly for
many years. Today, capacity is still increasing, but rotational speed
hasn’t increased in the mainstream market since the year 2000.
The monolithic shared storage array was the foundation of the data
center architecture of choice for the past decade.
Agility, cost, and scale are common reasons businesses today are
adopting cloud-focused architectures.
Now that you have a good feel for the past and the present, and you’ve begun to
see what the future holds, Chapter 3 will take a deeper look at what’s on the
horizon for the modern data center. The software defined data center (SDDC)
vision is all the rage today, and for good reason. You will see how the software
defined model is helping enable hyperconvergence and transforming the data
center as we know it.
[1]At the time of writing, this EPA report can be downloaded at:
www.energystar.gov/ia/partners/prod_development/downloads/EPA_Datacenter_Report_Congress_Final1.pdf
3
The software defined approach took hold with compute, but is now starting to
encompass all areas of the data center, which has led to the term software
defined data center (SDDC). The SDDC isn’t any one thing specifically, but
rather a way of describing a data center where as many pieces as possible are
abstracted into software. The SDDC is characterized by automation,
orchestration, and abstraction of resources into software and code. By nature,
code is more reliable than humans, which means that compared to a legacy data
center, the SDDC is more secure, more agile, and moves more rapidly. The
fallout of abstracting physical resources across the data center is that all of a
sudden, the hardware is substantially less important to the big picture.
Commoditization of Hardware
Historically, computing has been enhanced by the creation of specialized
hardware that is created to serve a specific purpose. Application-specific
integrated circuits (ASICs) are developed, as the name suggests, to serve one
specific purpose. In other words, they have one primary application. While this
model of computing can lead to increased performance, lower latency, or any
number of desirable metrics as compared to commodity hardware, it also comes
with substantial costs that must be weighed.
How does this relate to the software defined data center? Well, because the
SDDC’s goal is to abstract as much physical function into software as possible,
the physical equipment becomes less important. This means that platforms
which would previously have required special hardware can now be emulated or
replaced with software and run on commodity hardware. We’ll discuss some
more specific examples of this later.
All of this is not to say that there isn’t a case for specialized hardware. There are
times when it makes sense based on the application to use custom silicon created
just for that task. Hardware being a commodity also doesn’t mean certain
vendors can’t set themselves apart. One vendor may beat another to integrating a
new technology like NVMe, and that might make them the best choice for a
project. But again, something like NVMe is a standard; it is meant to replace
proprietary manufacturer-specific ways of doing things.
When hardware resources are abstracted into software and allowing commodity
hardware to run the workload, IT is afforded more flexibility, choice, longevity
of hardware, and likely a lower cost as well.
Shift to Software Defined Compute
Compute is data center jargon meaning “CPU and memory resources.” The term
is necessary since the mainstream adoption of server virtualization, which
abstracted the resources that formerly defined a server — those resources being
CPU, memory, networking, and storage. In a post-virtualization data center,
CPU and memory are grouped together as “compute,” and networking and
storage are handled separately.
The advantages businesses have seen from deploying software defined compute
(which is just a different way of saying “server virtualization”) are broad and
numerous. But some of them include massive consolidation of physical
resources, increased IT agility, and increased performance and utilization.
Features of modern hypervisors allow for compute workloads to be migrated
between physical servers without any downtime, and software intelligence
places workloads on the most appropriate physical servers based on utilization.
The challenge for other SDDC resources like storage and networking is to get
adoption and trust of the technologies to become mainstream. Software-defined
compute has already reached that point, so the challenge for it today is to add
more control and intelligence to the platform. Software-defined is all about the
ability to manipulate the system from outside the system (typically via
Application Programming Interfaces, or APIs), so expanding on the features and
robustness of the compute platform’s APIs will be key moving forward.
Shift to Software Defined Storage
One the heels of server virtualization is the idea of abstracting and controlling
storage resources in the same way that had already been done with compute.
Software-defined storage (SDS) could be thought of as storage virtualization.
The purpose is to take disparate groups of physical storage medium, pool them,
and abstract them so that they can be consumed programmatically via the SDS
platform instead of accessing each resource independently.
Software-defined is about abstraction, but it’s also about control. So, SDS will
also include data services that allow administrators to optimize and protect the
data stored. Some possible data services to include are:
Thin provisioning
Deduplication
Compression
Cloning
Replication
Snapshotting
One of the primary benefits of SDS is the fact that the abstraction of underlying
storage allows the use of heterogeneous storage platforms in a way that looks
and feels homogenous to the administrator and to the applications. In fact, SDS
commonly moves the granularity of storage management from the storage
aggregate and volume level to the virtual machine or virtual disk level. This
allows far greater control and flexibility in the environment.
By decoupling underlying storage hardware from the SDS platform, one of the
biggest IT pains of the last decade — a storage refresh — is more or less
eliminated. Where mixed-generation hardware configurations would not be
allowed on a legacy platform, SDS typically makes the transition smooth by
allowing various types of underlying hardware to coexist.
Similar to how SDS decouples the physical storage from storage management,
software defined networking (SDN) decouples the control plane from the data
plane and allows programmatic interaction with the control plane. This change
allows the network to be redesigned on the fly or scaled to meet demand.
Traditional networking has relied heavily on link oversubscription, and as
virtualization has changed the game, SDN allows network architects to create
dynamic networks that can be adapted to meet changing requirements.
SDN, as with storage and compute, also removes the dependence on specific
manufacturers and vendors as the physical hardware is abstracted. This means
that IT is no longer bound by the product life cycle and development iterations
of a given vendor. Rather, they can move hardware in and out at will underneath
their abstraction layer. This is made possible by projects like OpenFlow, which
has become almost synonymous with SDN.
The answer to this security problem in the SDDC could be called software
defined security.
As with the other software defined systems that have been discussed, software
defined security is characterized by the abstraction of security management and
policy from the devices and platforms that are providing security, and being
secured. Software-defined security allows the automation of security policies
and changes to said policies. Plus, automating changes allows for higher
precision which, in turn, leads to fewer security incidents due to human error.
It’s really many different layers of the SDDC that make hyperconvergence
possible. To really understand hyperconvergence, a deeper level of
understanding of software defined storage is especially critical. Without SDS,
the flexibility that makes hyperconvergence what it is would be impossible.
The Details of SDS
Software-defined storage is a really tricky subject to nail down and understand.
This is because, similar to Cloud and DevOps, software defined is a philosophy
and not any one specific thing. Thus, the categorization of what is and is not
SDS can be a bit challenging.
There are a few broad features that characterize SDS which should apply to any
solution or technology purporting to be SDS:
The requirement to provide abstraction does not mean that SDS can’t be a way
of consuming storage from a more traditional, monolithic storage array. SDS is
commonly associated with hyperconvergence; however, that’s only one of many
ways that SDS can be leveraged. An SDS layer can provide the method for
managing, automating, and scaling an already specialized storage solution.
That abstraction is typically found in one of two implementation types. The first
type is a virtual appliance deployed in the infrastructure. This virtual appliance
contains the software that provides and manages the SDS platform and abstracts
the storage behind it from the workloads in front of it. The other method is
kernel-level storage virtualization. Rather than running in a virtual machine,
software runs on the hypervisor itself to provide the storage features of the SDS
platform.
Policy-Based
The application of policy rather than specific settings reduces administrative
burden, eliminates opportunity for administrator error, and introduces a method
of ensuring consistency over time in the environment. In an SDS environment,
policy may dictate any number of settings related to the storage devices
themselves or the how the workloads are placed, protected, or served.
Imagine applying these specific settings to one virtual machine a single time.
The task is not incredibly daunting, given the right software. However, imagine
applying these same settings to 1,000 virtual machines in an environment where
six new virtual machines are provisioned each week. It’s only a matter of time
before mistakes are made, and with each new virtual machine an administrator
will burn time setting it up. With policy-driven SDS, simply by having applied
the policy (created once), the virtual machines will be treated exactly as desired
with accuracy and consistency over time.
Programmability
Management automation is the hallmark of the SDDC. For helpful and
approachable automation to take place, the functions of a system must be
accessible to third parties via the use of APIs. An API is a developer friendly
way of exposing resources in such a way that another program can query them
(get data about a resource) or manipulate them (initiate actions or change
properties). Some examples of API implementations are SOAP, which is
becoming less common, or REST, which is becoming more common.
APIs are necessarily present in SDS because the SDDC as a whole uses some
sort of orchestration engine to make all the pieces work together. That
orchestration engine needs a way to interface with each of the individual
components, and APIs provide that integration point. The Programmable Data
Center is a subset of the overall software defined data center vision, which aims
to allow anything and everything to be accessible via API.
Scalability
Finally, SDS is highly scalable in nature. This characteristic works in
conjunction with the abstraction; in part, it is the abstraction that provides the
scalability. By seamlessly allowing different physical hardware to be added and
removed underneath the abstraction layer, changes to the scale of the system can
be completed without the workloads every being aware. This gives organizations
leveraging SDS a distinct advantage over the prior method of scaling storage.
Historically, storage was scaled by buying a bigger unit and painfully,
methodically migrating data to it. SDS allows the addition of more storage, or a
shift to a new platform to take place, that is totally transparent to the workload.
What Is Hyperconverged Infrastructure?
Hyperconvergence is an evolution in the data center that’s only just beginning to
take hold. The past couple of years have seen hyperconverged solutions
developing at an incredibly rapid pace and taking hold in data centers of all
sizes. Hyperconvergence is a data center architecture, not any one specific
product. At its core, hyperconvergence is a quest for simplicity and efficiency.
Every vendor with a hyperconverged platform approaches this slightly
differently, but the end goal is always the same: combine resources and
platforms that are currently disparate, wrap a management layer around the
resulting system, and make it simple. Simplicity is, perhaps, the most sought
after factor in systems going into data centers today.
57% of survey respondents say they are looking at both appliance-based and software-
only options for hyperconvergence. Of those who have a preference one way or the
other, 25% prefer the appliance-based approach, and 18% are more interested in
software-only solutions.
The opposite view is that leveraging a VSA and no custom hardware opens up
the solution to a wide variety of hardware possibilities. While flexible, the
downside of this approach is that it consumes resources from the hypervisor
which would have served virtual machine workloads in a traditional design. This
can add up to a considerable amount of overhead. Which direction ends up being
the best choice is dependent on myriad variables and is unique to each
environment.
The Relationship Between SDS and HCI
It’s important to realize how much software defined storage (SDS) technology
makes the concept of hyperconvergence infrastructure (HCI) possible. If SDS
didn’t exist to abstract the physical storage resource from the storage consumer,
the options left would be the architectures that have already been shown to be
broken. Namely, those architectures are silos of direct attached storage and
shared storage in a monolithic storage array. Pooled local storage has advantages
over both of those designs, but would not be possible without the help of SDS
which performs the abstraction and pooling.
Most hyperconverged platforms offer the ability to apply data protection and
performance policies at a virtual machine granularity. This capability is also a
function of the SDS component of the hyperconverged system. Policy from the
management engine interacts with the SDS interface to apply specific changes to
only the correct data. This granularity, again, would not be possible without
software defined storage.
The Role of Flash in Hyperconvergence
There are many things that go into making a hyperconverged model successful,
but one component that hyperconvergence absolutely could not be successful
without is flash storage.
The performance capabilities of modern flash storage are the only reason it’s
possible to attain acceptable performance from a hyperconverged platform.
In a legacy monolithic storage array, there was one way of achieving additional
performance for quite some time: add more disks. Each disk in a storage array
can serve a certain amount of data at a time. This disk performance is measured
in I/O Operations per Second (IOPS). In other words, how many individual I/O
requests (reads or writes) can the disk complete in one second.
In a massive monolithic array, this was no problem. Add another shelf of disk,
and you’re on your way. In the land of hyperconvergence, this becomes a serious
problem however. You can’t just go on adding disks in perpetuity. A disk-
focused 2U server using 2.5 inch disks can usually only fit 24 of them. So what
happens if the workload requires more IOPS per node than 24 spinning disks are
capable of providing?
Flash storage is orders of magnitude faster than magnetic disk due to its solid
state (non-mechanical) nature. A single solid-state drive (SSD) could easily
deliver the IOPS performance of all 24 spinning disks. Because of this dramatic
performance benefit, flash storage is critical to hyperconvergence. Physical
limitations would not allow for the creation of a high performing
hyperconverged system without the performance boost that flash can provide.
Raw performance aside, SSDs can also provide high performing cache which
can front-end a large amount of capacity. Using SSDs as cache allows
hyperconverged platforms to get high performance and great capacity numbers
at the same time.
Using flash to provide caching for a group of higher capacity, slower spinning
disks is commonly referred to as hybrid.
There are a number of different disk configurations that you might see used in
hyperconvergence (Figure 3-1):
The last few years have seen flash storage and software defined infrastructure
growing in maturity, and both are finally getting to the point of being ready for
the mainstream IT organization. The cost of flash storage has been a challenge
historically, but within the next year or two the cost of flash storage will drop to
the point where it’s affordable for most situations. Hyperconvergence will
become more mature from a development standpoint but even more simple from
an administration standpoint.
Is there anything holding the industry back? Well, hyperconvergence and the
SDDC are entirely based on the prospect of a virtualized infrastructure.
Therefore, customers who are only lightly virtualized (or haven’t virtualized at
all) have a long way to go before hyperconvergence can be a reality in their data
center. There are also decisions for them to make regarding how and where the
workloads will move to as they phase out physical servers. Virtualization and
hyperconvergence in an on-premises data center is one option, but they could
just as easily choose to move those workloads to a public cloud provider and
remove all physical equipment from the facility. In one way or another, it’s safe
to say that the forward progress of the SDDC ideal is totally dependent on
virtualization.
Truth: Modern enterprise flash drives are commonly capable of withstanding 10 full drive writes per day
over the course of five years. That durability is covered under warranty and the drive
can be easily replaced if it doesn’t live up to that promise.
Truth: Depending on workload requirements, flash storage likely has a lower cost per
IOPS than spinning magnetic disks. As the cost of flash capacity comes down, SSDs
will likely be cheaper in terms of both capacity and performance.
Use the following expression to determine whether flash is an economically superior choice for a
given workload:
IOPS required GB required < cost per GB (SSD) cost per GB (HDD)
If the expression is true for the given workload, flash storage is a good choice.
If the expression evaluates to false (meaning the left side of the expression is NOT less than the right side)
perhaps a hybrid or spinning disk approach would be preferred.
This consideration doesn’t just apply to raw capacity, either. When applying data reduction techniques like
compression and deduplication, the cost per gigabyte of flash in “effective capacity” is even lower than
when evaluating the raw capacity.
As flash storage prices do fall even more in the near future and the software
defined data center model becomes easier to implement, it’s quite likely that
there will be an exponential increase in the number of IT organizations who
decide they’re ready to make the leap into a new kind of data center architecture.
The consideration for them will be whether they’re going to build a new kind of
on-premises data center with software defined compute, networking, and storage
or whether they’re going to shift toward a public or hybrid cloud consumption
model. Rightscale’s 2015 State of the Cloud report showed that 82 percent of IT
organizations surveyed have a hybrid cloud strategy, so clearly this will be a
major part of the coming years in the data center. The SDDC will be able to
abstract the cloud provider so the SDDC management platform can provision
workloads in an on-premises data center or a public cloud data center based on
policy.
Also, 15% are currently using hyperconverged infrastructure, and another 35% plan to
add some in the next 12 to 36 months. Hyperconvergence will help virtualization gain
traction inside the organization due to its simplicity and efficiency.
Chapter Recap
In this chapter, you saw how the SDDC approach is revolutionizing the way the
modern data center operates. You also briefly got a look at the characteristics of
software defined storage and hyperconverged infrastructure. Recall that
hyperconverged infrastructure is an architecture that is enabled by software
defined storage (among other things). Following are a few key terms and
concepts to take away from this chapter.
The SDDC isn’t any one thing specifically, but rather a way of
describing a data center where as many pieces as possible are
abstracted into software. The SDDC is characterized by Automation,
Orchestration, and Abstraction of resources into software and code.
Data center functions being abstracted into software allows for the
commoditization of hardware. This drives cost down and decreases
potential for vendor lock-in.
In the next chapter, you will learn about defining modern business requirements
so that IT can be successful. Namely, this means accurately addressing concerns
like Risk, Cost, and Agility. Defining these requirements is critical to IT’s
success as their responsibility to the business grows. Look at this alarming
statistic: according to research done by Oracle, “By 2020, IT departments will
need to manage... 10-times the servers, 50-times the data, 75-times the files —
All with only 1.5-times the people.” Defined, accurate business requirements are
the only way this will be possible.
4
We’ve covered the software defined data center (SDDC), how it relates to
software defined storage (SDS), and the adoption growth of the SDDC model in
data centers around the world. It’s the flexibility and agility that the SDDC
model brings which can help enterprises become vastly more efficient and agile.
Only through this model can enterprises hope to meet the needs of the business,
today and tomorrow.
The Business Requirement Challenge
Infrastructure is required to run applications that meet the business’ needs.
Businesses, though, have been notoriously bad at defining requirements for a
projects and sticking to them. This has caused many headaches and wasted IT
investments, in both time and dollars.
In many cases, business requirements change because the business didn’t really
know what they were getting into to begin with (they had never implemented
such a project or they didn’t know to ask the right questions in order to outline
the requirements accurately). At other times, those requirements change because
the business has to react quickly to new information, for example finding out
that their competition is about to implement something on a larger scale, or
unexpectedly getting an offer to acquire the competition in a certain area.
In the realm of IT, we don’t always take the time to understand the business or
connect with the people who run the business (who are the customers /
consumers of the IT organization). In many cases, IT practitioners get frustrated
when the requirements for a project involving IT are defined, then unexpectedly
change. In other cases, there are IT pros who simply don’t like change — at all.
Those IT pros are frustrated with the change in requirements because they know
that those changes will require that they invest unexpected and unplanned time
into getting the infrastructure ready to meet the business requirements.
Additional time investment, for full-time employees might require reprioritizing
tasks or extra work hours during nights or weekends. With full-time employees
those additional hours invested are soft-dollar costs, meaning they don’t cost the
company any extra real money. However, if the person working on the project is
a contractor who is billing the company on an hourly basis, then those additional
hours required to meet the new business requirements are hard-dollar costs,
meaning they affect the company’s bottom line.
Even worse, in many cases, a change in business requirements requires not only
unexpected change in time investment but also unexpected changes in
infrastructure requirements, and additional infrastructure (or even reconfiguring
new infrastructure). These are all hard-dollar costs, and they are the costs that
really get the attention of executives.
Thus, the best option for IT infrastructure groups is to ensure that the
infrastructure design and solutions are as efficient, scalable, and agile as
possible. With such an infrastructure design, IT infrastructure groups can more
easily adapt to eminent changes in infrastructure requirements due to changes in
the business.
When analyzing IT infrastructure designs and solutions for the modern data
center, there are 5 overarching challenges that such designs and solutions must
be able to solve.
They are:
Risk
Cost
Complexity
Just like the old business saying, “everyone is in sales,” a similar saying applies to IT:
“everyone is in security.”
The question for you is, can you ensure that your data center is designed in such
a way that you can meet the ever-changing business requirements without
unnecessary risk?
For example, modern SDCC solutions utilize policy-based systems such that
policies can be placed on a virtual machine or application data to ensure that it
remains encrypted and highly available no matter where it moves across the
enterprise data center or into the cloud.
Other types of business risks that must be taken into account are as follows.
Assurance, Business Continuity, and Disaster Recovery
One of the requirements that the business needs from IT is assurance that
company data is secure and that applications will be available even if the
unexpected should happen. IT can only provide this type of assurance if they
have first worked with the business to understand the business priorities,
processes, and operations to create resiliency in the infrastructure and plan for
business continuity and disaster recovery.
The question for you now is, how can your future infrastructure design and
selection help you better provide the necessary business assurances and more
easily support BC and DR processes if the unexpected does occur? After all,
everyone in the business expected zero downtime of the infrastructure or
applications until they are presented the cost required to make that possible. By
leveraging more modern SDDC solutions that offer software defined disaster
recovery or replication functionality, enterprises are able to dynamically adapt to
infrastructure outages and recover critical applications at a much lower cost than
traditional DR solutions.
The Regulatory Landscape
Increasingly, businesses are affected by the regulatory landscape. Governments
and regulatory agencies create regulations to ensure that businesses act in such a
way that protects the safety and interests of their customers and employees.
These regulations can put tremendous pressure on IT groups to maintain data
security, retention, and accounting in very specific ways to comply with
regulations.
With today’s constantly changing and growing data, ensuring compliance with
such regulations while using traditional data center infrastructure can be next to
impossible. What companies need is the ability to know what data they have,
where it is, and whether it’s compliant with the regulations.
Just as important, businesses also need to be able to ensure that that data
maintains its own policy compliance as the data moves throughout the
company’s infrastructure, mobile devices, and the public cloud.
For example, virtual machines may be stored in a format that makes them
difficult to convert and move to another hypervisor, a cloud storage provider
might charge you an exorbitant fee to move your data out of their cloud, or a
hyperconverged server vendor might sell you a solution that requires you to
continue purchasing their equipment as you need more capacity in the data
center.
Certainly there are different levels of lock-in. On one hand, some lock-in may
just be an inconvenience that requires you to go through additional required
steps to leave that vendor’s solution. On the other hand, other levels of lock-in
may have massive financial impacts that require the company to continue paying
a large monthly payment, whether they use the solution or not. For example,
cloud or hosting providers might require you to continue paying a monthly fee or
lose your data in a service that has a large barrier to exit.
That being said, most modern data center solutions are becoming so open that
workloads and data can be dynamically moved from one solution to another,
even without downtime. For example, if a solution uses vSphere you can usually
utilize vSphere’s Storage vMotion (svMotion) to move from one storage solution
to another but keep in mind that deduplicated data will have to be rehydrated.
Still, you must be ever vigilant for solutions that try to lock you in, and evaluate
your level of comfort with lock-in. Remember, every solution you choose has
some level of lock-in, even if it is a minor one.
The risk here is that if IT is not an active participant in the business processes,
it’s unlikely that they will be seen as relevant to the business. It’s also unlikely
that they will be able to help the company leverage technology to their
advantage, and it’s unlikely that they will receive accurate business requirements
and project scopes for new initiatives.
Are willing to take the time to understand the business and its
technology needs.
Have technology that can help the business to be agile, efficient, and
scalable.
In other words, IT must not only “talk the talk” (of being a business-enabler) but
also “walk the walk,” or have the technology to back it up.
When put that way, IT is one of those ugly overhead expenses that it doesn’t
want to be. So, what can we in IT do to restructure our budgets, our technology,
and our staff to achieve the position of “business-enabler” that we so desperately
desire?
Help to drive down the cost of manufacturing below the cost of its
competitors. You then create a competitive advantage and are a
business enabler.
What the OpEx model brings is the ability to simply pay for the infrastructure
that they need to run the applications they require, solely based on resource
consumption — even down to a per-minute consumption basis.
Executives and financial people in enterprises around the world are asking IT to
adopt an OpEx, pay-as-you-go pricing model to provide the infrastructure
needed to run the company’s applications and store their data. These pricing
models certainly apply to public cloud infrastructure but they can also apply to
on-premises infrastructure, either through the vendor who supplies the
technology or through creative financing sources.
When it comes to selecting modern infrastructure for the data center, selecting
infrastructure that is designed in such a way that it can easily scale up or down
with the greatest efficiency and lowest cost possible may make the difference
between being able to provide a reasonable OpEx pricing model or not. Modern
software defined infrastructure and virtualization allow companies to maximize
their infrastructure purchases and more easily scale their infrastructure when
needed.
The Changing Career Landscape
As the infrastructure in the data center changes to become more intelligent and
efficient, and as the business expects more out of IT practitioners, the job duties
and roles in IT organizations are changing as well.
To some in IT, their immediate concern is that less complexity in the data center
will lead to less demand for their skills. All IT professionals should be assured
that technology skills will always be in demand. However, as you know, the
challenge with technology skills is to always keep them up to date.
But, when have you ever met an IT professional who felt absolutely up to date
on all the latest technology?
In our recent survey, 70% of respondents said that they had only “some knowledge”
of SDS and 64% had “some knowledge” of hyperconvergence. Thus, the vast majority
of IT Pros need to improve their knowledge of these technologies to be prepared to
manage the datacenter of the future.
The reason that there are no IT professionals who absolutely, always feel current
on all the latest technology is because there’s always something new in
technology and the landscape is so wide. In other words, think of the mantra
“evolve or die.” If you don’t adapt your skill set to the new and more efficient
technology in the data center, then you will undoubtedly become obsolete. For
example, there aren’t many fax server administrators or mainframe architects
around these days because those technologies are becoming more and more
obsolete.
With that being said, if data centers adopt SDS, many administrators question
whether the role of the storage administrator will still be necessary. If someone
is thinking about the SAN administrator who sits around provisioning LUNs all
day or monitoring Fibre channel switch ports, then yes, that type of storage
administrator may become obsolete. However, if the storage administrator was
to elevate him-or herself to a higher level and think of themselves as being
responsible for all data across the company, then that type of storage
administrator (who administers the SDS infrastructure, cloud storage, and data
security/protection infrastructure) would be very valuable to any company —
today and in the future.
Taking that to a higher level, the architect who leads the company from
traditional storage to SDS could help the company to achieve dramatic cost
savings and gain business efficiencies (and earn a large bonus in the process).
Another way to think of future data center roles is that all roles need to be
elevated to a higher level. For example, no longer do you have server
administrators, storage administrators, network administrators, and virtualization
administrators. Instead, with smarter infrastructure requiring less deep
knowledge, the traditional data center silos can be broken down and the
traditional data center positions/roles go along with them. Instead of those
specialized administrative roles who had deep product knowledge, all of those
roles in the future (as long as newer, more efficient infrastructure is used) could
be consolidated into a role simply called “infrastructure administrator” or
“infrastructure architect.” People in this new role would have a greater breadth
of infrastructure knowledge but less depth because it’s no longer needed. With
their time freed from constantly having to maintain deep product knowledge,
they could be a much more valuable asset to the company by spending their time
finding ways to help the company use technology to its advantage to become
more competitive or increase bottom-line profitability.
Most would agree, that with the state of today’s technology we have finally
gotten to a point where, when given enough hardware, software, and IT experts,
we can meet the needs of the business in terms of agility, performance,
availability, and scalability. However, what’s lacking for most companies is that
they simply can’t afford or can’t justify the costs required for the infrastructure
and personnel in order to achieve their utopian vision of IT services. It’s
complexity, in most cases, that is the barrier we must overcome in order to meet
or exceed the business requirements.
More modern SDDC solutions are able to eliminate the complexities that we’ve
become accustomed to from more proprietary hardware-defined systems. These
hardware-defined systems were only created originally, because commodity
servers with a software layer simply couldn’t provide the resources or
performance required to run the services needed by the business. For example,
storage area network solutions were only created when it became obvious that
servers with local storage simply couldn’t provide the performance or capacity
necessary to store all of the company’s files.
Today, commodity hardware and software defined solutions are able to provide
all of the performance and capacity necessary for just about any enterprise. Plus,
by running all infrastructure services in software so that many of the complex
server, storage, and network complexities can be eliminated, the daily life of the
infrastructure administrator can be made far easier (and make the role of the
infrastructure administrator even possible, as we discussed above).
However, software defined solutions don’t stop with simply making the life of
the administrator easier. Modern software defined systems with less complexity
and greater efficiency go a long way to making that utopian vision of IT services
possible by making it easier (and less costly) for IT to provide the agility,
performance, availability, and scalability that the business truly requires.
Agility and Performance
To truly compete in today’s business environment, companies must be agile.
What that means is they must be able to very quickly adapt to the ever-changing
business and competitive environment. In order to do that, companies must be
able to perform, or execute, what needs to be done exactly when it needs to be
done, and exactly in the way that it needs to be done. In most cases, the
company’s agility and performance are the determining factor between success
or being put out of business by the competition.
In terms of the modern infrastructure, agility and performance can be tied (at a
high level) to being able to deploy new applications and application functionality
very quickly, or being able to scale-up existing applications very quickly.
At a lower level, you could tie the term performance to the actual performance
of an application. This, in turn, ties back to the actual performance of the servers,
storage, and network that make that application possible. In order to achieve
high performance at the application level, there is an abundance of hardware
resources available, at lower costs than ever before (faster CPUs, faster data
paths, flash storage, and greater densities).
No longer do large companies always have the edge over smaller companies.
Today, it’s the agile companies, performing what’s needed and when needed,
which are the most successful in the business world.
In our recent survey, the single most important decision criteria (at 72%) when
considering software defined storage was the performance of the that could be gained
in their infrastructure, once implemented. For those who had implemented SDS, half
of all respondents said that the performance was higher than they anticipated.
Automation and Orchestration
One way that companies are becoming more agile is by automating and
orchestrating current business processes and the deployment of new IT services.
There’s always great confusion on the difference between automation and
orchestration (see Figure 4-2).
Orchestration
Orchestration, on the other hand, works at a higher level and leverages
automated tasks. One example of orchestration is deploying a new 3-tiered
application environment with a Web server, middleware server, and database
server. To accomplish this level of orchestration, the orchestration engine would
leverage prebuilt automated tasks (which you can think of as building blocks).
One of those tasks would be the same example of automation previously
discussed where we have scripted the deployment of a new Windows server. The
orchestration engine would use that automated task to deploy three new Window
servers, all from the same automation scripts, and then use other automation
scripts that perform the configuration of these three different types of servers.
The entire process that is pulled together by multiple automated task, is
orchestration.
Keep in mind that self-service doesn’t necessarily mean that every employee in
the company has access to an application catalog and can deploy whatever
applications they fancy or whatever virtual machines their heart desires. For
example, self-service can be performed in such a way that only trained
developers or trusted power users in the company have access to perform the
deployment of fully orchestrated environments.
Here are two statistics that anyone planning for the future of enterprise data
center storage should be alarmed about:
When it comes to finding a solution for the data growth challenge utilizing SDS
is the best place to start. Most SDS solutions include data reduction services in
the form of deduplication and compression. You’ll want to fully understand the
details of how the SDS solution performs these data reduction services because
not all deduplication and compression are created equal.
So, what can enterprise IT do to ensure that applications and data are both
resilient to failure and completely available in order to meet business
requirements and customer expectations?
Traditional storage systems had zero application awareness and local high-
availability for such storage systems required fully redundant, costly, and
complex designs. Additionally, high-availability between data centers for the
storage infrastructure not only required a duplication in cost and complexity of
the storage but also expensive licenses and plentiful bandwidth between the
sites.
SDS systems also work in a policy-based manner such that they can increase the
number of replicas (on the fly, without downtime) for the most critical
applications, and use the minimum number of replicas for less critical
applications. They can also replicate only the changed blocks of the most critical
applications to remote sites to ensure the data is protected and that applications
can remain highly available in the event of the unexpected (while maintaining
data efficiency). By utilizing application oriented policies, SDS systems are able
to provide application resiliency and availability where it’s most needed and
when it’s most needed.
After all, if the Internet-based SaaS applications have set the bar for your end-
users in terms of high availability, shouldn’t enterprise IT organizations also use
the same type of advanced software defined intelligence used in that
“HyperScale” world?
Chapter Recap
In this chapter, you learned that modern IT business requirements are more
demanding than ever. IT organizations must “adapt or die” in order to remain
relevant to the business. You also learned about the numerous facets of business
that IT is a part of — risk, cost, complexity, agility, performance, resiliency, and
availability — and how modern software defined solutions can help IT to rise to
the challenge of modern business requirements.
Up next, you’ll learn some principles that can help you make your own data
center more agile. Get ready to learn how to make a big impact on the business.
[3] UNECE Stat: www1.unece.org/stat/platform/display/msis/Big+Data
The modern enterprise is a very different beast than the enterprise of days past.
In the modern enterprise, IT and the data center play pivotal roles in the
operation and success of the business. In fact, the role that IT and the data center
plays in helping the business achieve its goals is becoming increasingly critical.
IT is not just an expense that impacts the bottom line. For many businesses, IT is
a key top-line revenue driver.
However, the old ways of managing these critical resources are no longer
sufficient. Rather than build complex conglomerations of components that are
carefully crafted into crazy combinations, the emerging enterprise must create
environments that are agile and that imbue the business with ferocious
flexibility.
Thinking big
Starting small
Moving fast
Each of these three principles is discussed in this chapter, along with strategies
that can help you forge your own path through the modern data center’s forest of
options.
Think Big
Small thinkers need not apply. Today’s technology environments demand big
thinking from people who have deep understanding of both the business and
technology.
Be bold.
In practical terms, here’s what this means: it’s time to look at the whole
technology environment and re-evaluate everything.
Does the data center lend itself to emerging constructs, such as pay-as-you-go
adoption? Data centers of days past required massive initial investments, which
were neither economical, nor particularly practical. Organizations then spent
years attempting to recoup these large investments, only to find that they may
have never really realized a maximum return.
New thinking may require that you jettison your old ideas around how IT
systems are procured, deployed and supported. Your efforts, though, will result
in a streamlined IT function that is laser-focused on the needs of the business.
Are you still providing break/fix services for devices like desktop computers and
printers? Anyone can do that. Outsource it. You will likely save money and end
up getting better service.
It’s easy to say that you should outsource something, but it’s important to
understand the reason why you should outsource these kinds of commodity
services. That reason is summed up by two words: opportunity cost.
Although agile IT principles demand that you consider all aspects of the IT
organization, since the focus of software defined storage (SDS) and
hyperconverged infrastructure (HCI) is on the data center, let’s start there.
If you have users who would be better served through the implementation of
self-service tools, implement them. Just remember, doing so requires that you
have the aforementioned automation and orchestration tools in place.
36% of survey respondents indicated that their organization will add more software
defined storage systems in the next 12 to 36 months. 35% of the same group indicated
that they plan to deploy more hyperconverged infrastructure in the next 12 to 36
months.
Start Small
With an understanding for what agile IT demands, the next question you may
have is, “How do I do it?”
Even the creators of the most ambitious plans need to start somewhere. It’s not
generally feasible to simply throw away everything that already exists and
replace it with something new. At the very least, new initiatives need to be
staged for implementation in such a way that they minimize impact to
production.
So, start small. Find something that needs to be fixed and, well, fix it. Perhaps
the physical server environment is not well-managed and you have a desire to
increase your use of virtualization. At the same time, part of your “big thinking”
plan is to move the entire data center to an HCI solution or one that leverages
SDS.
Unless you’re a very small company, you can’t do that all at once. So, use the
server virtualization initiative as your stepping stone. Implement your brand new
virtual desktop environment on an HCI or SDS solution. Then, during the
replacement cycle for your server environment, transition that environment over
to the new one. In this way, you’ll maximize your investment in your legacy
infrastructure while eventually gaining the benefits of the new one, such as easy
scale, fast deployment, and simple administration.
Your projects should be use-case driven and tied to clear business outcomes.
Even achieving a simpler, more cost effective data center is a business outcome
if it means that the IT department as a whole can now be more responsive to
business needs.
Beyond the infrastructure though, starting small may mean finding a place to
begin automation. Many organizations, for example, continue to manually
provision user accounts. You don’t necessarily have to adopt an expensive and
complex identity management solution, but with just a few scripts, you can
probably automate a good chunk of the process. The role of the person who used
to provision accounts then moves to one of oversight rather than action, freeing
that person up to service other higher priority needs.
Now, with that extra time that you gain in IT, what do you do with it? You
improve the business. Meet with the business unit leaders in your company and
help them discover and implement new services that increase customer
satisfaction or revenues Help them find technology-driven ways to reduce costs.
Figuring out how to provide analytics that helps a company market better to their
customers. For example, if you’re a bank, helping to implement the ability to
scan checks to deposit them via smartphones is a great example of how
technology can drive customer satisfaction. This service changed the retail
banking industry and it’s primarily a technology solution.
Move Fast
If you’re asking yourself when you should get started, that’s an easy answer:
Now! Don’t wait.
— Scott D. Lowe
4. Identify Core Services and Gaps. At this point, you should have a
good understanding for how the various products and services present
in the organization actually meet business needs. You should also have
a good understanding for where gaps exist in the environment.You
should be able to answer the following questions:
Which products and services are core to the mission and
vision of the business?
The core services list may not mean that you need to maintain the
status quo. It simply means that the particular service is required to be
provided in some way. It’s here that you begin taking concrete action
to correct deficiencies and help propel IT forward.
People have been really wary of the word outsource because it can imply
that people will lose their jobs. But you have to look beyond all of this
and figure out what is best for the business. Outsourcing is not just about
saving money; it’s also about creating opportunity for the business.
Every time you outsource a function, the internal people that were
handling that function can redirect their efforts to more value-add
projects.
As you review the list of services being supported by IT and the list of
support gaps that you have identified, you need to look for the best way
to address the core services. Sometimes, this will mean outsourcing a
core service. For example, if Microsoft Exchange is a core service, does
it make more sense to keep it in-house or should you move it to Office
365?
6. Decide What to Improve. With what’s left, you must now decide
where you want to improve. The hard part can be figuring out where to
begin. With that in mind, here’s a tip:
Look at the items on the list and, for each one, assign two metrics (a
value of 1 to 5 for each):
The difficulty level for addressing that need. For the difficulty
metric, a 1 indicates that it’s low difficulty and a 5 means that it’s a
high level of difficulty.
Now, immediately attack the items that have a 1 or 2 difficulty level and
a 4 or 5 payoff level. Get them done and done fast.
If it looks like this advice is 100% focused on the business rather than IT, that’s
because it is. However, as you’re making your way through these six steps, look for
opportunities to improve IT itself.
Would a move from a traditional data center environment to one based on
hyperconverged infrastructure be able to address the complexity issue that we discussed?
Would assigning staff to begin automating routine tasks be of benefit to the business?
These kinds of initiatives can absolutely be beneficial to the business as they can free up scarce IT
personnel to tackle those complex business challenges that were identified when you began your journey.
Sometimes, by addressing internal IT needs, you’re also making it easier to address the critical goals of the
business.
Chapter Recap
In this chapter, you learned that time is not on your side. But by leveraging the
tools and services available to you, it’s possible to move beyond just “keeping
the lights on.” Following are a few key terms and concepts to take away from
this chapter.
Agile IT is a set of principles that require that people think big, start
small, and move fast.
The key to success is to make sure that you can align your IT function
with what the business needs.
Up next, it’s time to talk about action. Chapter 6 will discuss how to begin the
actual transformation of your data center.
6
Jack Welch, most known for his role as CEO of General Electric for two
decades, is quoted as saying, “An organization’s ability to learn, and translate
that learning into action rapidly, is the ultimate competitive advantage.” Starting
with the principles from Chapter 5, it’s time begin the transformation of your
data center. This must be done immediately!
Look at what Jack says: the competitive advantage comes from translating
knowledge into action rapidly. Data center transformation sounds like a lofty
goal, but it needn’t be quite so scary. There are plenty of opportunities to begin
effecting radical change without tearing everything down and starting from
square one.
Whether the data center is the product, or the data center supports the people who sell
the product, or some combination of both, the only reason that a data center exists is
to make the business money. Keep this perspective when assessing opportunities for
data center transformation. You’re sure to get stakeholder buy-in when the
transformation will have a measurable impact on the bottom line.
— James Green
With that in mind, the first step to transformation is to take a hard look at which
transformation choices will affect the bottom line.
For example, a radical overhaul to turn all of the blue LEDs in the storage arrays
to newer, sleeker green LEDs are not likely to be well received by the board.
However, if these transformations lower operational expenses by reducing
administrative complexity, they will be better received. Or do these
transformations increase the accuracy of the final product, reducing the number
of products that are discarded as faulty or returned? If so, that’s another way to
garner support. Perhaps the transformations make another business unit happy;
that’s always a good way to find support for a project! If the finance team needs
updates to their portion of the website to be completed more quickly, changing
the development workflow and using automation and orchestration to increase
the speed of iteration on development projects will make them happy.
Regardless of what the benefit to the business is, you must have a clear goal in
mind before beginning this transformation. Implementing a hyperconverged
solution to aid in building an SDDC simply for the sake of having a software
defined data center is missing the point and is liable to get someone fired.
So, what’s the best way to make sure that a project looks good from the start?
Get some easy wins right out of the gate. This makes the project look good to
stakeholders and increases support for the project moving forward. Let’s look at
some ways to get started on the right foot.
Where to Address the Low-Hanging Fruit
No business is exactly the same as any other, so there can be no conclusive
blueprint for completing the transformation to a modern data center. However,
there are a number of technology use cases that apply to a great number of
businesses. It’s quite likely that one of these use cases applies to your business in
one way or another. Any one of these use cases can be the perfect opportunity to
show the value of the software defined approach by taking a technology and
business process that the organization is familiar with and streamlining it.
Plus, the more accurately the environments are reproduced each time, the less
error prone the Test/Dev process, and finally, the more automated the creation
and destruction of environments, the less time operations-and-development staff
waste performing repetitive operations. Their attention can then be directed to
other important tasks.
Test/Dev environments are low-hanging fruit for many due to the fact that
development staff can immediately see the benefit of the work being done, and
sometimes are even eager to help. Getting input from the developers can help
create an agile infrastructure that caters perfectly to their needs.
SDN can allow the developers to sandbox their work with no additional effort
required, which leads to less frustration and quicker, more accurate testing.
82% of survey respondents indicated that they support at least one remote/branch
office. A full 19% of them support 25 or more sites! A software defined approach to
managing this expansive environment could dramatically cut costs in many
environments.
The SDDC transformation has already begun in most data centers, but if it hasn’t
in yours, it must begin immediately. That value of intelligent, optimized,
automated server virtualization is huge and provides operational and financial
benefits in almost every case. If this process has begun with some basic server
virtualization, automating deployments and configurations in the next big step.
Leveraging configuration management tools like Puppet or Chef to ensure
continuity throughout the environment and radically increase the speed of
provisioning will pay dividends.
While compute agility is important in big data, the name “big data” also implies
there’s a lot of it, and all of it must be stored. Storage agility is also critical.
Thanks to the abstraction that SDS can provide, however, scaling the
performance and capacity requirements of the big data workloads is simple. SDS
also allows the workloads to be controlled by policy, allowing the precise needs
of the workload to be met automatically.
53% of survey respondents indicated that “Big Data” was one of their primary uses cases for using Software
Defined Storage or Hyperconverged Infrastructure. With the massive amount of data storage involved, it’s
no surprise that these organizations are looking to leverage SDS to keep a handle on things.
Software Defined Networking
SDN enables big data analytics by exposing APIs that can be
used to automate the creation and destruction of job processing
environments. Because big data environments often contain
sensitive information, SDN can provide micro-segmentation for enhanced
security as well as automated, policy-based access to the data.
Disaster Recovery
While disaster recovery (DR) is last on the list of low-hanging fruit use cases, it
is certainly not the least. DR is important to environments of all sizes.
When it comes to the SDDC, the first step in the transformation is to gain
knowledge about specific products and technologies that will enable the
transition. We’ll explore a few ideas for specific technologies that you could dig
into implementing SDC, SDS, and SDN.
Then, armed with some knowledge and excitement, the next step is to begin
testing a very simple abstraction: a “Hello, World!” sort of experiment with
software defined solutions. This will cement a better understanding of how the
tools that enable the SDDC to work.
When adopting an SDDC mentality, there’s a critical step that many organizations
overlook: begin viewing everything through the lens of the SDDC vision.
If you miss this step, your SDDC vision will never be realized, despite doing all the
other steps correctly. It’s all too common for an organization to complete an SDN
project and then attempt to retrofit the use of the SDN tools and infrastructure in their current operational
model. At this point, the value of SDN is drastically diminished and this effort is doomed to fail. To
successfully complete the SDDC portion of the transformation to the modern data center, SDDC must be
the new operational lens moving forward.
— James Green
Software Defined Compute
Practically speaking, how do you go about adopting SDC technologies and
practices into your business? The answer to that depends largely on which use
case (VDI, Test/Dev, etc.) will be the target of the SDC initiative.
There are several questions you’ll need to answer as shown in Figure 6-2.
Cloud or On-Premises?
Depending on the workload target, the first decision to be made is whether on-
premises resources or public cloud resources will be used. If public cloud is the
chosen direction, SDC is likely already provided. The task, then, is simply to
integrate this resource with the management systems and workflows of the rest
of the SDDC. If the compute platform is to be built on-premises, there’s many
more decisions to make and more implementation work to do.
At this point, the choice comes down to one of three basic options: virtual
machines, containers, or both. For many organizations, the choice for now will
be “both.”
Hypervisor/OS of Choice?
Hypervisor/OS of Choice?
Assuming your organization has chosen to move forward with both virtual
machine and container-based virtualization, the first step of real action is to
begin building hypervisors and/or host machines to perform the virtualization.
For virtual machines, choosing and installing a hypervisor will be the first step.
For containers, choosing and installing the Linux distribution for the host
operating system will be the step there. Either way, a system now exists that has
the following software defined features: it abstracts the compute workload from
the physical resource running it, it provides a programmatic interface to allow
for automation and orchestration, and the compute workloads can be controlled
by policy. And just like that, the compute is software defined.
SDS will allow one or both of these deployment models to be used, but knowing
which model fits the application will help with making SDS decisions moving
forward. Also at this stage, the type or storage medium (or the ratio in the case of
mixed medium) needs to be determined. Will the performance and capacity
requirements necessitate flash storage? Disk? DRAM?
Hardware or Software?
There’s a dichotomy in SDN that also exists in SDS and SDC, but is not as
prevalent. That is that some solutions, while enabling the software defined
vision, actually require certain pieces of hardware, while other solutions are
completely software. Neither model is right nor wrong in a universal sense; the
challenge is to choose the one that is right for a given situation.
When adopting an SDDC approach to transforming your data center, you can’t
expect to change everything at once. The SDDC transformation is often a
gradual process, starting with compute, then moving to storage, and finally
networking and security. However, the most common process isn’t the best for
every organization, and there may well be organizations that should transform in
the exact opposite order. No matter what, getting measurable improvements to
business processes, in terms of reliability and speed, is the end goal. Always aim
for that, and the project will likely be successful.
Simplification
As important of a transformation as any other in the journey to the modern data
center, the simplification of data center systems and processes has the potential
to revolutionize the way an IT organization operates and in the end, how it
spends money. The simplification process needs to be a broad and sweeping one,
yet inspected at the most granular level possible. Leave no stone unturned in the
quest for removing complexity.
Complexity may, in fact, be the single costliest attribute of a data center. Think
about the fallout from complexity in the data center: troubleshooting is a total
disaster because the system has to be reverse engineered before troubleshooting
can even begin; operating expenses increase as more staff is required to maintain
the complex systems; new systems take ages to implement because of the
rigidity of the environment. It’s plain to see that attention paid to the process of
simplification in the data center can return immediate benefits.
In the modern data center, there is a better way to solve the storage problem. The
solution is hyperconvergence. While hyperconverged infrastructure (HCI) is not
a panacea, it’s quite a good fit for solving many of the problems exhibited by
traditional data center architectures. Chapter 8 will discuss hyperconvergence in
great depth, but for the purposes of this section, just understand that
hyperconvergence is the “pooling of direct attached storage, flash and potentially
local DRAM to create a distributed storage system.”
Rather than many servers pointing at a single storage target, the storage is spread
throughout the servers. Software-defined storage (SDS) tools allow that direct
attached storage to be protected and managed as if it were one big array.
Alternatively, it seems that finding a certain project that’s a good fit and
deploying a hyperconverged system (rather than upgrading or scaling an existing
legacy system) is a successful way for many IT organizations to begin the
transformation. This strategy of finding a specific, well-fitting project and using
it as a way to open the door for hyperconvergence can be referred to as
opportunistic hyperconvergence. In other words, rather than a rip-and-replace
transformation of the data center, you would do a new implementation or make
the move to hyperconvergence as new systems are built and old systems that
aren’t supported any more need to be replaced.
Implement Opportunistic Hyperconvergence
Opportunistic hyperconvergence comes in a few different flavors. The first is the
one previously discussed — leveraging hyperconverged infrastructure for a
specific project to prove its potential. A very common example of this is VDI.
Because the nature of VDI workloads is so different from that of server
workloads, it is preferred that they run in segregated infrastructures so that they
don’t cause each other performance problems.
Keep in mind that VDI is only an example. Any project where the intention is
already to deploy separate infrastructure is a perfect candidate for opportunistic
hyperconvergence.
However, if the business is hesitant to try this new direction and a new remote
office is being built, why not use that limited-risk opportunity to give
hyperconvergence a shot? This outside-in approach is surprisingly easy to grow
as internal support for the technology increases. Because of the way most HCI
platforms are designed, adding systems in the main data center down the road
and connecting them up with the nodes out in the remote office is a trivial
process.
Management
It makes little sense to transform the details of the data center for the better if the
big picture remains blurry. What will eventually make the SDDC shine in the
eyes of the business is having a robust yet nimble grip on the entire data center
by using a set of management tools that monitor and control the big picture.
Insight is sought after more than gold in organizations today, but providing it is
tricky. Taking appropriate action based on that insight is trickier still. The final
component to transforming an old, tired data center into a modern data center is
to bring new life to the management systems.
It’s critical when managing a data center to be able to get a top-to-bottom view
of the entire infrastructure. All aspects of operating a data center are made more
difficult by not having complete visibility. Being able to manage all the way
through the infrastructure stack makes troubleshooting, maintenance, and design
more fluid. It’s also important to begin to shift toward a hands-off approach
where systems function without the need for IT’s intervention. This means
investing in automation, workflow orchestration, and self-service provisioning.
The modern data center accomplishes far more than the data center of the past
but with less manual work required of the IT administrators. This frees up staff
resources to keep innovating and revitalizing the data center.
Full Stack Management
Because visibility through the stack is so important to the overall management
picture, it’s vital that your hyperconvergence vendor of choice for transforming
the data center provides the tools needed to get this visibility. The more parts of
the stack that are under their control, the more insight can be gained. This is the
challenge with traditional data centers. The storage system is completely
unaware of the network system which is completely unaware of the compute
system. Making decisions without all the relevant information is nearly
impossible. The only way to make truly beneficial decisions regarding workload
optimization or failure protection is to have all the details.
Today, there seems to be two methods of full stack management, neither being
more preferable than the other:
Automation
Orchestration
Self-Provisioning
If it can accomplish these three things, the modern data center will have been
realized. It would be preferable that the management system itself provides these
functions, but if it doesn’t, it’s acceptable that it simply exposes APIs to allow
other, better suited systems to interface with it in order to perform those roles in
the environment.
Automation
Streamlined automation is the hallmark of the modern data center. There are two
key reasons why automation is critical:
Self-Service
The final thing a full stack management system should provide or enable is a
self-service provisioning model.
Transforming a data center is only valuable for one reason, and it’s the
same reason why the data center exists in the first place: the data
center makes the business money. With that in mind, the first step to
transformation is to take a hard look at which transformation choices
will affect the bottom line.
Address low-hanging fruit first. It’s wise to begin the data center
transformation with a technology that is most familiar to the team, has
specific value to the business, and is extensible into other areas of the
data center.
One of the major players in the data center transformation is software defined
storage (SDS). The next chapter will take an in-depth look at what SDS is and
how it works.
7
It’s undeniable that all electronic storage is accessed and managed through some
type of computer software. So aren’t all storage systems built on software? The
answer is, yes, of course. All of the traditional storage systems that are in use in
data centers today are built on software.
The SDS landscape is a bumpy one. Let’s answer some of the most common
questions about SDS to help clear things up.
What Is Software Defined Storage?
The exact definition of SDS is still evolving, but the generally accepted
definition is that software defined storage is “where the management and
intelligence of the storage system is decoupled from the underlying physical
hardware.”
This means that the SDS software is then able to provide policy-based
management and provisioning of the data being stored, regardless of the storage
hardware that’s being used.
Most SDS systems create a file system overlay on top of the physical hardware.
That filesystem is then utilized by virtualization hosts (to store virtual machines)
or by physical servers — both of which provide application services.
The SDS storage system is typically distributed across the multiple physical
hosts that provide the actual physical storage capacity. The distribution offers
both high availability for the data (in the event of failure) as well as performance
(by having multiple copies of the data available). That physical storage capacity
can be a mix of traditional spinning disk or flash storage (which can be used for
caching or for data storage, depending on the SDS design).
Some might think that SDS and hyperconvergence are very new and not used in the
field but in our recent survey we found that 36% of companies are already using SDS
or hyperconvergence and 35% say that they will be increasing that usage.
Today’s SDS systems include storage virtualization that provides this abstraction
of the storage intelligence away from the underlying storage. SDS systems
typically allow consumers great flexibility in the underlying storage hardware
and advanced storage features. For example, they may provide deduplication,
compression, thin provisioning, snapshots, replication, caching, and other
advanced storage functionality.
More and more, SDS systems are fully accessible thought a RESTful API
(Figure 7-1) so that they can participate in automation and orchestration
processes which allow the storage to dynamically adapt, as the business demands
change.
SDS may be sold separately from the underlying storage, or it may be included
with the underlying storage. The most important thing is that the software that
makes the storage possible can be separated from the storage hardware. In many
cases, SDS runs on top of different operating systems and is compatible with
multiple types of underlying storage hardware.
Along with software defined networking (SDN), SDS is a critical piece of the
software defined data center (SDDC) ideal that most companies are striving to
adopt.
What Isn’t Software Defined Storage?
Traditional storage area networks (SAN) and network attached storage (NAS)
systems that are packaged as a single solution where the storage intelligence and
the hardware are coupled so tightly that neither of them can be interchanged, are
not SDS.
The reason that these so-called “hardware-defined storage systems” were created
in the first place was because, at the time, the server hardware that could have
run this storage intelligence simply wasn’t adequate to provide the processing
and throughput that the applications required of it. The rest, dedicated SAN and
NAS hardware, was created in order to tightly couple the software with
specialized hardware to provide high-performance.
Today with server CPU, bus, and I/O channels being able to offer such high
performance, there is no reason that intelligent software can’t run on just about
any X86 server in the modern data center and provide the storage services that
are required by the applications.
If you were to compare SDS to traditional SAN/NAS solutions, you would find
the following differences:
SDS can pool numerous types of hardware (and even cloud) resources into a
single pool that can then be managed using software-based policies which are
created and applied based on the needs of the company and their applications.
The most common types of storage resources that can be pooled by software
defined storage include:
Flash Storage. With the lower costs, higher capacities, and the
incredible performance of flash memory, flash storage is being used
more and more in just about every SDS solution. Flash storage is
commonly used in SDS for performance acceleration through caching
or, in some cases, as primary storage in all-flash storage solutions.
There are multiple form-factors of flash that you have to select from
including solid state drive (SSD), memory channel storage (MCS), and
PCIe-based flash. Any of these flash storage options can be added to
existing servers and, along with traditional spinning disks, can be used
to provide advanced performance (though caching) or advanced
efficiency (through deduplication and compression).
Types of Flash
Many may assume that there is but a single type of flash storage and that any variation
in price is all a marketing-scheme. However, as of 2015, there are actually five types
of flash storage that you need to be aware of. The technical differences cause not only
variations in price but variations in life span, endurance, performance, and capacity.
SLC. Single level cell (SLC) flash is the highest performance and likely
the highest cost flash available – typically used only in enterprise
applications. It has the highest endurance at around 100,000 program
erase cycles per cell as well as the highest performance.
eMLC. Enterprise multi-level cell (eMLC) flash offers good
performance and good endurance at 20,000 to 30,000 program erase
cycles per cell. Used by many enterprises, eMLC many enterprises and
higher-end consumer products, eMLC sits between SLC and MLC in
price, performance, and endurance.
MLC. Multi-level cell (MLC) flash is lower-grade than eMLC and its
endurance reflects that at roughly 10,000 to 30,000 program erase cycles
per cell. Used by many consumer-grade applications, it’s not
recommended for applications that perform many writes.
TLC. Triple level cell (TLC) flash offers the maximum density possible
but it’s endurance is typically only between 1,000 to 5,000 program
erase cycles per cell; however, it’s going to be the lowest cost flash
storage available. Used for low-end consumer-grade electronics, TLC
flash is also the lowest performance flash storage.
3D NAND TLC. A new type of flash storage, 3D NAND, is when TLC
flash is configured to use larger bit cells. The result is a low cost flash
alternative that provides similar performance and similar durability as
MLC flash. The relatively new flash offering is reported to offer roughly
the same number of P/E cycles as MLC but at a lower cost.
RAM. Almost every virtual host, where SDS runs, has some excess
RAM capacity. Many SDS solutions will leverage a portion of that
RAM memory to improve data efficiency with the end result being
improved storage performance (Figure 7-2).
You might be familiar with some of these advanced data services based on
features that are already available in traditional storage systems. The difference
with SDS is that these advanced data services are commonly included with the
SDS solution at no additional cost.
Also, SDS makes certain that new types of data services possible, such as data
mobility across the data center, thus opening up new opportunities for more agile
data centers.
Data Reduction
With data at enterprises and across the cloud growing exponentially, enterprises
need every form of data reduction available. The two most common forms of
data reduction are deduplication and compression.
Deduplication occurs when a storage system reduces the size of the data by
eliminating the redundancies (copies of data) over a large data set.
For example, consider for a moment all of the copies of the Windows or Linux
OS that are used in an enterprise data center. A very large percentage of the
operating systems for servers and virtual machines stored in enterprise data
centers is duplicate data. Deduplication allows you to store just a single instance
of each block of data in your environment. By enabling deduplication in your
storage environment, you can save tremendously on capacity. For example,
imagine if you had 100 absolutely identical servers. With deduplication, you
would in essence store just a single copy of that server and would not need to
store the other 99.
Deduplication Extends the Life of Flash
You know that deduplication saves space. But even better, when well-implemented
inline by the vendor, data deduplication can even help you get more life out of your
flash storage.
As you probably know, flash storage devices have a finite lifespan, measured in the
number of program erase (P/E) cycles that the device can withstand. Every time data is written to a device,
the device must perform a P/E cycle on the target cells.
Now, imagine you have 100 identical servers and deduplication means that you only have to store one of
them. That means that you can avoid having to write the other 99 copies, thereby forgoing the need to put
the flash media through the P/E cycles necessary to write those other 99 copies. This technique, also called
write avoidance, is one of the primary methods by which flash storage vendors are able to ensure that flash
media can last for the long term.
As shown in Figure 7-3, there are different types of deduplication designs such
as in-line, or pre-process deduplication (performed at the time the data is written
to the storage system), and post-process deduplication (performed after the data
has been written). Each type has its own set of pros and cons. See the sidebar
entitled “Pros and Cons for Deduplication Types” to learn more.
Figure 7-3: Post-Process vs. Inline Deduplication
There are numerous, highly technical, methods for performing data deduplication, but
when viewed from a higher level, the two choices that most IT organizations choose
are either pre-process or post-process deduplication. Each of these has its own pros
and cons.
Inline Deduplication
With pre-process deduplication, data is deduplicated before it is written to disk (usually in memory or in a
flash tier). The upside to this is that there is never any duplication of the data on disk so disk space is not
consumed for duplicate data and an I/O is not consumed in the storage infrastructure.
The downside to pre-process deduplication is that there is may be a storage latency penalty and associated
with performing deduplication of data before it is written. However, many deduplication solutions are able
to eliminate the overhead of pre-process duplication with caching.
With any pre-process deduplication solution, a potential concern is whether or the solution can provide
global pre-process deduplication (on all data) or if the pre-process deduplication system just works on a
small subset of data and then the remaining data still has duplication (which could be eliminated by post-
process deduplication).
Post-Process Deduplication
With post-process deduplication, data is deduplicated after it is written to the disk, by a separate process,
after the fact. The upside to this is that there is no performance overhead for the deduplication process as it
is typically done during non-peak hours when there are free I/O and CPU cycles anyway.
Another upside is that post-process deduplication has a better chance of performing global deduplication,
across all data in the data center, than pre-process as it typically has access and resources to do so.
The downside to post-process deduplication is that it occurs after data is already written, and that means
that it happens with the penalty of a write I/O and requires storage capacity on disk (until the post-process
deduplication process happens).
Many advanced storage systems give enterprises flexibility by being able to perform both pre-process and
post-process duplication.
Also of special consideration when considering deduplication systems are the block size used when the
deduplication is performed as well as the type of media that the deduplicated blocks are stored upon. Both
of these factors can be significant factors in the performance of the deduplicated data, whether it is pre-or
post-process duplicated.
Compression, on the other hand, also eliminates duplicate data but does so in
small data sets such as in a file or a block of data. Like deduplication,
compression requires CPU processing cycles and doing real-time compression
(comparable to pre-process deduplication) has its trade-offs between data
reduction and performance throughput.
While both compression and deduplication can consume resources if done in real
time, most modern SDS solutions will eliminate that performance penalty via the
use of intelligent caching, providing both storage reduction and high
performance.
I/O Acceleration
You previously learned about pooling resources and how RAM and flash storage
can be part of those pools. SDS solutions often use high-speed RAM and flash
storage, not just to mitigate the performance impact of software defined data
services but also to accelerate storage throughput.
Because SDS allows you to manage your storage using policies applied per
virtual machine or application, this type of I/O acceleration can also be applied
on a per-virtual machine or per-application basis instead of across an entire array
or an entire storage LUN.
In our recent survey, respondents rated advanced data services as either important or
very important when it came to selecting SDS or hyperconvergence solutions. For
example, 90% said data reduction was either important or very important and 92%
said that replication was important or very important.
Snapshots
As with virtualization hypervisors and traditional storage systems, you can take
snapshots with SDS solutions.
The downside to traditional storage snapshots is that they usually require you to
snapshot the entire storage LUN. This requires much more capacity than is
needed.
Depending on how the storage is presented, SDS lets you take storage-or virtual
machine-level snapshots that are virtualization-aware and, in some cases,
deduplication-aware. Essentially, what this means is that you get the best of both
worlds. You get to offload the resources necessary to take the snapshot onto the
storage (where it belongs), but the preserved data only pertains to the virtual
machine.
Cloning
Almost every virtualization administrator has used clones in their virtual
infrastructure. Clones duplicate a virtual machine in the case of a virtualization
hypervisor. Clones can save administrators a tremendous amount of time,
because when they need to create new virtual machines, they can simply clone
an existing virtual machine.
The concern that administrators have always had with clones is the performance
impact that they will have depending on what type of clone is being performed.
With SDS cloning, like snapshots, enterprises receive the best of both worlds
where a clone of a virtual machine can be taken, with virtually no performance
penalty, and be used immediately.
Replication
Data replication is used heavily by companies of all sizes for data protection in
the event of a localized data center outage (such as failure of a SAN or NAS) or
in the event of a large scale outage (e.g., the company headquarters and data
center were destroyed by a tornado). No matter the issue, when data is not
available, applications are going to be down — the company will start to lose
money.
Replication essentially copies data on disk to another disk, and that other disk
could be local to the same data center, or in another data center across the state
or country. In the event of data loss, that data can be used to bring up, for
example, the virtual machines running the company’s point-of-sale system that
were replicated to a backup data center.
With traditional storage, replication was typically only enabled when a separate
license key was purchased from the storage vendor. The most granular data
replication that could be done was on a per-LUN basis, which likely included
many virtual machines.
With SDS, however, replication is yet another of the advanced data services that
can be enabled on a per-virtual machine basis and it’s included in the SDS
software (no extra key or licenses are needed). SDS-enabled replication can be
used to protect data both in a local data center cluster, across data centers, or
across the cloud.
Data Mobility
Once SDS abstracts away the physical storage hardware, your data is now
mobile and can be moved across any of the various forms of storage — flash,
HD, SAN/NAS, cloud, and more (Figure 7-4). What this means is that if, for
example, you replace your SAN and move to I/O-accelerated local storage, your
data can move dynamically, from one form of storage hardware to another
without any downtime for the applications using that data.
However, storage mobility isn’t just for hardware replacement. For example,
let’s say that you want to provide a virtual machine with a higher tier of storage
(from silver to gold). With storage mobility, you could change the storage policy
on a virtual machine and, with no downtime for the virtual machine or its
applications, the virtual machine storage could be migrated from one type of
storage (perhaps SATA “bronze tier” storage) to another type of storage
(perhaps I/O-accelerated “gold tier” storage).
Figure 7-4: Tiered Storage Systems
Encryption
Just like other advanced data services enabled all in software, data encryption
can be enabled to ensure that all data (or just an individual virtual machine, for
example) is encrypted and secure when stored.
Thin Provisioning
When it comes to SAN LUN provisioning or virtual machine disk provisioning,
LUNs and virtual disks are filled with whitespace in order to reserve their total
capacity. However, in most cases, the storage provisioned is only fractionally
used. Thin provisioning tells the device which requested the storage that the total
capacity is available. However, in reality, the only storage capacity that has
actually been used has been reserved (not the total).
As shown in Figure 7-5, with SDS, thin provisioning is enabled by default and
is done across the entire storage pool, not just on individual LUNs.
For example, what if you lose a disk? A group of disks? A caching disk? Or a
host in the storage cluster?
When analyzing SDS solutions, there are several types of data resiliency and
data protection capabilities to ask about.
SDS Checklist
When considering SDS options, there are many traits that SDS should display. They are:
Software-only
High availability
Data protection
Simplified configuration
Policy-based management
Storage virtualization
Hardware flexibility
I/O acceleration
Snapshots
Cloning
Replication
Data mobility
Encryption
Thin provisioning
Failure Scenarios
First, consider all the different ways that an SDS system could eventually fail.
What will be the effect on the data and applications using that data should the
storage system fail?
Recovery
When failure does occur in the storage system, storage administrators should be
notified so that they can repair or replace the failed hardware, at which time the
SDS system can rebuild the distributed data, across all nodes in the storage
cluster.
High Availability
With traditional SAN/NAS storage, creating a fully redundant storage
infrastructure is complex and expensive. To create true storage high availability,
you must either purchase a second, redundant SAN/NAS or, at minimum,
configure your SAN/NAS with fully redundant everything.
SDS systems must provide high availability both at the disk level and host level.
With a solid understanding of what SDS is and how it works, the next step is to
explore the topic of hyperconvergence and see how SDS enables
hyperconverged infrastructure (HCI).
8
Hyperconverged Infrastructure
With the rise of server virtualization, companies are realizing the improved
efficiency, productivity, and return on investment (ROI) that server
virtualization can bring to the data center. Specifically, many virtual machines
can be run on each host, administrators can “do more with less” by administering
more servers per administrator, and companies can gain a greater financial return
on investment in every single server once the virtualization hypervisor is
installed.
When companies first started virtualizing servers, they realized how inefficient
and costly the storage silo of the data center was. It slowed the provisioning of
new virtual machines, it was complex to maintain, and required dedicated
administrators and experts. Worst of all, it ate up the majority of infrastructure
budgets while providing a much lower return than other data center
infrastructure. But what could be done about the storage?
The next solution to the storage problem came in the form of hyperconvergence.
What is hyperconvergence, exactly?
Hyperconvergence
As you can see (Figure 8-1), the servers run a hypervisor, and that hypervisor
either runs the virtual storage appliance (VSA) or uses a hypervisor-integrated
storage management module to provide the distributed storage layer for the
virtual machines. That distributed storage layer is spread across the servers that
make up the storage and compute cluster.
Distributing the storage across the compute layer using software defined storage
(SDS) and running your business workloads alongside the VSA is the minimum
requirement to dub something “hyperconvergence.” However, most modern
hyperconvergence solutions also provide:
“Single pane of glass” management. This means a single, integrated
management interface is available for both the virtualization and
storage resources, all in one place, for administration,
performance/capacity monitoring, and troubleshooting.
For example, when learning about and shopping for SDS solutions, you should
ask, “How can the SDS solution…”
When we asked respondents to tell us what their future plans are regarding storage,
the responses paint a bleak future for disk-based storage. A full 19% of respondents –
almost 1 in 5 – say that they will fully decommission their disk-based storage systems
over the next two to three years. The primary gainers in the same timeframe will be all
flash arrays and hybrid storage arrays, but 35% each also say that they will expand
their use of software defined storage and hyperconverged infrastructure.
Hypervisor/kernel-integrated
Hypervisor/Kernel-Integrated Storage
With hypervisors/kernel-integrated storage, the SDS solution runs inside either
the kernel of a virtualization hypervisor or inside the kernel of another operating
system.
The benefits of this type of SDS deployment are that the storage software has
direct access to the operating system and thus, the underlying storage hardware.
Some vendors claim that hypervisor/kernel-integrated storage offers a slight
performance benefit over the VSA approach (discussed more later); however,
the real performance difference in SDS is determined by a combination of the
SDS software’s speed, the storage media used for caching/acceleration, the
storage media used as the primary storage tier, and the networking connecting all
of these components.
The drawback to using this type of SDS deployment is that the SDS solution is
typically only compatible with the kernel of the virtualization hypervisor or
operating system that it’s integrated with. What that means to enterprises is that
hypervisor/kernel-integrated storage solutions are version-locked, and in some
cases feature-constrained, due to this level of deep integration.
Another drawback is that these types of SDS deployments may require the
purchase of specific and costly virtualization hypervisor licenses that really are
required to run the SDS solution, and that don’t offer immediate value to the
enterprise. All of these drawbacks are typically summed up into the single term,
hypervisor lock-in.
Different VSA solutions have different requirements for the minimum number of
nodes necessary to achieve high availability. Some can operate with as few as
two nodes and, as such, are the most economic. Others require a minimum of
three or four nodes.
Flexibility. The VSA approach offers the flexibility to use the lowest
cost virtualization hypervisor or operating systems to run the SDS on
top of, as makes sense for the business.
Hybrid or All-Flash
Of particular importance when choosing a hyperconvergence solution are the
types of storage media and how they are used in the HCI nodes. Hyperconverged
solutions can use all hard drives, hybrid (mix of flash and hard drives), all-flash,
or DRAM with flash.
The number of hard disks versus flash disks is going to have a number of
business impacts, because it determines:
For the HCI solutions that are offered as software-only, many of those offer a
reference architecture. Reference architectures tell enterprise architects and
administrators how many virtual machines or what type of performance the
enterprise will receive if they configure the hyperconvergence software with a
specific configuration of CPU, memory, and storage.
In other words, you can think of reference architectures as blueprints for an HCI
that are defined by the hyperconvergence vendor in cooperation with hardware
server vendors.
Both of these approaches have their own sets of pros and cons.
The downside with the packaged hyperconverged appliance approach is that they
are essentially locked in to only specific, static, server configurations from their
hyperconvergence vendor with few options for price negotiation.
Software-Only Approach
With the software-only, reference architecture approach, enterprises likely have
the option to use their existing servers or at least obtain servers from their
existing vendor/VAR at previously negotiated discount levels. This maintains
the enterprises’ comfort-level while gaining the best of both worlds: leveraging a
pre-architected, pre-sized, and fully supported blueprint to build their own HCI,
while using their own hardware.
Enterprises also have the option to either scale up or scale out by either adding
more CPU and memory to existing hosts or by adding more hosts of varying
sizes (discussed more in the “Question of Scale” section later).
Hypervisor Choice
Another aspect of hyperconvergence selection is whether the solution offers a
choice in hypervisor support. Any solution that is hypervisor-integrated is only
going to be offered on the hypervisor that the integration was done on.
While the hypervisor that you are using today may be supported by the
hyperconvergence solution you are considering, it’s important to remember
support for multiple hypervisors in your selection criteria, because these
hyperconvergence solutions will give you the greatest leverage when it comes
time to negotiate software maintenance and support on your existing hypervisor
(whether you choose to exercise that choice or not).
Server Hardware Choice
As mentioned previously with the comparison between “Appliance vs.
Software/Reference Architecture,” when considering hyperconvergence
solutions, it’s important to ensure that they also offer you hardware choice. You
may have little choice in hyperconvergence solutions that are sold or packaged
as an integrated appliance (some vendors offer you the choice between two
brands of appliances), but will have ultimate flexibility with hyperconvergence
solutions that are sold as software-only, reference architecture.
While some form of lock-in is inevitable when choosing hardware and software
solutions, enterprises want to ensure that the solution that they choose offers
them choice or the flexibility to easily move to the alternative (whether they ever
choose to exercise that choice, or not). In other words, very rarely is any
enterprise ever really “locked in” and unable to make a change in hardware or
software. So the real question is, “What is the price to change?” The level of
lock-in is very different if the price to move is $0 versus $10 million.
When considering choice in the context of server hardware; however, there are
many organizations that want someone to tell them what to do. It’s easier for
these companies to simply buy a complete hardware and software bundle that
they can put it a rack, turn on, and immediately begin using. While they may
become locked in to that platform, that is by choice.
The Question of Scale
Before we go into details on the question of scale, let’s define the difference,
between scale up and scale out (Figure 8-3).
For example, with traditional storage arrays that were scaled up, you had to add
more disk shelves to the array to add more capacity and throughput (that is, until
you hit bottlenecks with the storage controllers in the array head).
With scale up hyperconvergence solutions, you have the option to add more
CPU, memory, storage capacity, or storage throughput to your existing servers in
the hyperconvergence cluster.
Scale Out (Linear Scalability) Defined
With scale out solutions, or solutions that scale linearly (like a horizontal line),
when you need to add more resources (such as CPU, memory, storage capacity,
or storage throughput), you only have the option to “scale out.”
What that means is that, when you need more capacity, you add another node
complete with more CPU, memory, storage capacity, and storage throughput
even if all you needed was, for example, just more memory.
Thus, with the scale out hyperconvergence design using the appliance-approach,
you will always be over-provisioned.
Do You Have Scale Up and Scale Out Options?
Another consideration when selecting a hyperconvergence solution comes back
to the question of how much scalability is available.
In the end, neither the applications nor administrators will think about
this design choice on a daily basis. What matters is the efficiency of the
hyperconverged system in terms of how many compute and storage
resources are available to workloads.
In the next chapter, you’ll read about some real world experiences with the
software defined data center (SDDC), SDS, and HCI. Chapter 9 will focus on
case studies and interviews detailing experiences with these technologies.
9
This chapter provides the main highlights around people’s existing data center
and storage environments with a focus on software defined storage and
hyperconverged infrastructure. If you’d like to read the full results and learn
even more about how your peers feel about these technologies, download the full
report from www.atlantiscomputing.com/survey2016.
Technology Domain Knowledge
With that in mind, we begin our analysis with a look at how the IT pros that
responded to our survey view their own knowledge of various data center
elements. Figure 9-1 shows the breakdown of survey results. As is very obvious,
the only area in which respondents believe that they have expert level knowledge
is server virtualization, with 55% responding as such. For two primary emerging
technologies – software defined storage and hyperconverged infrastructure –
12% and 18%, respectively, of respondents feel that they have little to no
knowledge of the subject matter. Only 18% of respondents feel that they have
expert-level mastery of each these topics. Given the relative age of these
technologies when compared to other data center technologies – server
virtualization, datacenter networking, and enterprise storage – it’s not that
surprising that knowledge level is quite a bit lower. Over time, we expect to see
people’s comfort level with software defined storage and hyperconverged
infrastructure approach that of enterprise storage, which 39% of respondents
have mastered. You will notice that, overall, people are more comfortable with
software defined storage over hyperconverged infrastructure. 12% say that they
have no knowledge of software defined storage while 18% say the same about
hyperconverged infrastructure. This is likely due to the fact that many software
defined storage systems more closely resemble traditional storage arrays
whereas hyperconverged infrastructure is quite different.
Figure 9-1: Knowledge levels of various data center technologies
Virtualization Penetration
Particularly with hyperconverged infrastructure, virtualization penetration is a
key indicator for just how much of the existing environment can be migrated.
Hyperconverged infrastructure deployments require that applications run
virtualized. With that in mind, gaining an understanding for a respondent’s level
of virtualization is important to learn just how successful that deployment might
be. We learned from respondents that most are at least 71% virtualized on the
server front, but that desktop virtualization is truly still in its infancy or, at the
very least, not of interest to many organizations. Only 19% of respondents are
more than one-half virtualized on the desktop.
In Figure 9-2, you can see the virtualization penetration rate for those that have
deployed either software defined storage or hyperconverged infrastructure. The
results aren’t radically different, but you can see that 75% are at least half
virtualized. The most interesting item here really revolves around the desktop. In
the total population, a full 23% have done no VDI. For those that have deployed
either software defined storage or hyperconverged infrastructure, only 10% have
not deployed VDI. This suggest that virtual desktops are of more interest to
SDS/HCI adopters. Figure 9-3 shows this correlation.
Figure 9-2: Virtualization penetration for servers and desktops
Figure 9-3: Virtualization penetration for servers and desktops – SDS/HCI deployers
Hypervisor Usage
Given the magnitude of virtualized applications – people are virtualizing more
and bigger workloads all the time – hypervisor choice is a critical issue. Not
every hyperconverged infrastructure solution is able to support every hypervisor
available on the market. It’s with that in mind that it comes as no surprise that
VMware vSphere remains the dominant choice in the hypervisor market (see
Figure 9-4). It’s also no surprise to see that, over the next 24 to 36 months,
many vSphere administrators intend to migrate to the latest version of VMware’s
hypervisor. Hyper-V will be the likely recipient for much of vSphere’s loss.
XenServer 6 looks to hold pretty steady as well. However, for those on
XenServer 5, it looks like they will abandon the platform for other options.
We were not surprised to see Docker more than doubling in adoption in the next
few years. Container technology is getting more attention and, much as was the
case in the early days of virtualization, we expect to see container adoption start
to become more interesting to people as they learn more about the technology
and as it expands to support more workload types.
What is surprising is what happens when the hypervisors are grouped into their
respective vendor categories and then analyzed. As you can see in Figure 9-5,
VMware will maintain its market share. This may seem insignificant, but when
considering the market as a whole, there is a story there, especially as the
virtualization market is likely to continue to grow overall. What that means is
that VMware is likely to simply maintain share in a growing market while those
that are abandoning other platforms – such as Citrix – are more likely to jump to
to Hyper-V rather than VMware.
Figure 9-5: Current and future hypervisor/container breakdown by product
Now, let’s look at the same information, but this time just for those that have
already adopted software defined storage (see Figure 9-6). The information here
suggests those that deploying software defined storage will do so at VMware’s
expense, with that company dropping from 81% of market share among SDS
adopters to 76%. Likewise, Citrix will drop from 30% share to 24% and Oracle
will lose 4% of its share. The gainers will be Microsoft, KVM, and Docker.
Microsoft is poised to gain 6% share among SDS adopters, while KVM will see
3%, and Docker a large 7% increase, almost doubling its market penetration.
Figure 9-6: Current and future hypervisor/container breakdown by product – SDS adopters
Among hyperconverged infrastructure adopters, the trends are similar, but with a
somewhat different magnitude. See for the breakdown of this. Here, VMware’s
market drops from 85% to 77%, a full 8% drop, which is substantial. Microsoft’s
Hyper-V starts today at 42% and is expected to jump to 47% among our HCI
adopter population. Citrix only loses a single point of their adopter base, and
KVM jumps a full 3%, to 18%. We do expect to see an increase in KVM
adoption among hyperconverged infrastructure users as the KVM-based HCI
options continue to penetrate the market. Among HCI users, Oracle usage is
poised to drop 6%, which is interesting since Oracle has their own converged
infrastructure solution. And, again, Docker looks to gain significant followers as
that product continues to improve.
Figure 9-7: Current and future hypervisor/container breakdown by product – HCI adopters
Storage Capacity
Being able to support workloads means having sufficient storage capacity in
your organization across both your primary location as well as any remote or
secondary locations. Both hyperconverged infrastructure and software defined
storage solutions have the capability to support both very small as well as very
large deployment scenarios and either one can support centralized or distributed
storage needs. As you can see in Figure 9-8, storage capacity varies widely and
there are substantial storage resources housed at remote locations. From this
chart, you can see that about 16% of respondents are running 20TB to 50TB of
storage at their primary location. The most surprising piece of information here
is just how much storage is present across remote and distributed sites. Only
18% of respondents indicate that they have no storage outside the headquarters
location.
When breaking the data down by our four primary verticals, it’s really easy to
see that the 20 TB to 50 TB storage range is the sweet spot for our overall
respondent group (see Figure 9-10). It’s also easy to see that different verticals
have very different average storage needs. For example, only 4% of those in the
education vertical are running 200 TB to 500 TB of storage whereas 21% from
finance have that level of capacity. Given the data-driven nature of financial
companies, this comes as no big surprise, but is interesting nonetheless. By
comparing the individual bar sizes in Figure 9-10, you can begin to see where
each vertical ranks with regard to storage capacity. Here are the major ranges for
each vertical (again, this is storage capacity at the primary location only):
Education: 20 TB to 50 TB
Finance: 200 TB to 500 TB
Government: 20 TB to 50 TB
Healthcare: 50 TB to 100 TB
Now, let’s look at the storage capacity breakdown across the aggregate of all
respondent remote sites. Figure 9-11 excludes storage at the primary location.
The data here is slightly more mixed than we see with capacity figures at the
primary location, with a large number of respondents having no remote storage
capacity. However, for those that do have storage resources in remote sites, the
20 TB to 50 TB range is once again the leader of the pack, but we also see a
jump in the number of overall organizations that have more than 1 PB spread
across remote storage systems. As mentioned earlier, this situation reinforces the
need for hyperconverged infrastructure and software defined storage solutions
that focus on ROBO use cases. Here are the major ranges for each vertical (this
time, this is storage capacity at remote sites):
Education: 20 TB to 50 TB
Government: 5 TB to 10 TB
Healthcare: 50 TB to 100 TB
With ROBO being a key use case for hyperconverged infrastructure, we wanted
to look at overall capacity at remote locations for organizations that deployed
one of these technologies. There were a total of 342 respondents that have
undertaken such deployments. In Figure 9-12, you can see the remote storage
capacity breakdown for each technology. Earlier, we learned that storage
capacity and company size are linked to one another; bigger companies have
more storage. From Figure 9-12, it’s clear that some very large companies have
deployed both software defined storage and hyperconverged infrastructure since
the choice “More than 1 PB” garnered the greatest number of respondents.
Figure 9-12: Storage capacity for SDS/HCI adopters (remote locations only)
Data Growth
Perhaps one of the most serious technology challenges facing organizations is
keeping up with the sheer growth of data. Figure 9-13 shows you that most
organizations are seeing a 10% to 30% annual data growth rate. However, a
number of companies see much higher rates, even 50% or 100%. For these
respondents, finding a storage solution that can scale easily and inexpensively is
absolutely critical to maintaining reasonable level of expense and application
availability.
In the four charts below, we can get a look at the data growth patterns for the
four primary verticals under scrutiny for our survey. As you can tell, in general,
the data growth patterns are all pretty similar; most organizations, regardless of
vertical, fall primary into the 10% to 30% data growth range and have some kind
of peak around the 50% data growth rate. Here, though, finance is something of
an outlier, with its “secondary peak” coming in at around 45% with a smaller
third peak coming at 65%.
It’s a similar story when considering this information just for those that have
deployed hyperconverged infrastructure or software defined storage. However,
while the peaks are in similar places – in the 10% to 30% data growth range,
fewer software defined storage users report these levels of growth.
Storage Performance
While storage capacity is absolutely critical to consider, storage performance is
also a key success factor for workloads. Over the years, storage performance
challenges have become severe, leading to the rise of flash-based storage
solutions and a renaissance in the overall storage market. Software defined
storage and hyperconverged infrastructure are two rising markets that have
emerged as a part of this renaissance. But, just how well are these newer entries
in the storage market meeting performance goals?
As it turns out, pretty well. Figure 9-14 shows that only 16% of respondents
have solutions that are slower than their disk-based storage systems. A full 50%
say that their solutions are faster than their disk-based systems, with 14% saying
that it’s as fast as an all flash system. Overall, from a performance perspective,
these newer storage options are holding their own.
When we asked respondents to tell us what their future (future = 2-to-3 years
out) plans are regarding storage, the responses paint a bleak future for disk-based
storage. A full 19% of respondents – almost 1 in 5 – say that they will fully
decommission their disk-based storage systems over the next two to three years.
Figure 9-16 shows that the primary gainers in the same timeframe will be all
flash arrays and hybrid storage arrays, but 35% also say that they will expand
their use of software defined storage and hyperconverged infrastructure.
None of this comes as a major surprise. Flash storage has been plummeting in
price for quite some time and is expected to hit price parity with disk within the
next few years. Once raw price parity is achieved, expect to see spinning disk
quickly fall off in terms of usage. Flash simply carries with it far too much
performance potential when compared to disk.
50% of respondents, though, say that they intend to deploy cloud-based storage
services. For the foreseeable future, we expect that most deployments will be of
the hybrid variety in which organizations combine cloud-based storage with
local storage. Over time, as more companies seek to further simplify their data
center environments, many are turning to the public cloud, which eases
deployment. However, because of security concerns, locality concerns, and even
cost challenges, many companies are discovering that keeping things private
makes more sense. We’ll see how this plays out in the coming years, but for
now, cloud is still a big plan for many.
This information would seemingly contradict what you just learned – that 19%
of people currently using disk-based storage arrays intend to decommission
them. However, bear in mind that, for those that intend to “remain the same” on
disk-based storage, that means that they will ultimately need to replace them,
which we believe is the reason that we see strong results for Traditional
SAN/NAS devices in Figure 9-17. Also note that the response categories are
slightly different, since we add cloud storage as an adoption option to this
question.
In terms of systems performance, data center space, and power and cooling
costs, there have been tremendous gains for implementers of software defined
storage and hyperconvergence. On the performance front, it’s more than likely
that the gains have come from the fact that the previous storage was more disk-
focused while the new solution is either hybrid or all flash. Data center space is
much improved in hyperconverged infrastructure scenarios since compute and
storage are integrated together into server nodes. Further, less equipment in the
data center overall translates to lower power and cooling costs.
You will notice that direct costs – acquisition and support costs – stay relatively
constant. About the same number of respondents experienced higher costs as
lower costs. While still providing a lot of value, software defined storage and
hyperconverged infrastructure solutions have not yet helped companies reduce
initial expenditures on infrastructure, but have helped when considering the total
cost of ownership. This leaves a major opportunity for companies in the
emerging storage space that can reduce both acquisition cost and TCO.
Hyperconverged Infrastructure or Software Defined
Storage Deployment Timeframe
Deployment of software defined storage and hyperconverged infrastructure is
happening in waves and is more than likely taking place based on existing
hardware replacement cycles. Figure 9-24 shows that over the next year or so,
17% of respondents say that they will undertake deployments. Over the next two
years, that number jumps to a total of 62%. 27% of respondents say that they are
uncertain as to their deployment plans, which could mean that they are still not
sure whether they will definitely deploy or they truly don’t know when they
might plan a deployment.
IT Transformation
On the Toyota production line, one operator can be responsible for many
machines because his or her job is to monitor and provide the “human touch,”
not to sit there and operate a single machine. In the same way, less engineers
will be required to operate a world-class data center if many of the processes and
operations are left to computers and the humans just provide oversight.
The Challenge
The number of disparate systems in the data center has grown to an
unmanageable level. There’s a product (or a few products) for nearly every
problem. In many environments, the burden of the number of moving pieces
causes serious inefficiency on the part of the IT staff.
This is a simple example, but imagine how much time the staff saves managing
the data center each year by not needing to dedicate time to this task. All they
need to do at this point is monitor the orchestration system to be sure tasks are
completed as expected, and fix it if something is not working correctly.
The Solutions
Due to the way that cloud-based IT infrastructure tends to be more suited for
ephemeral workloads, automation and orchestration will become increasingly
important. The amount of resource creation and destruction mandates this. Since
the primary purpose of data center infrastructure is to serve the application that
runs on it, a new culture is emerging that emphasizes the community and
teamwork of the development staff and the operations staff as a single unit.
These roles have been segregated in the past few years, but experience has
shown that there’s much more value in the teams working together. As such, the
DevOps methodology will experience dramatic uptake in the coming years.
These two ideas were briefly covered in Chapter 6. Feel free to jump back to that
section for a quick review before diving in deeper here. This section focuses on the
future and potential long term impact of automation and orchestration as compared to
the basic overview in Chapter 6.
Automation
Automation has always been attractive to the IT organization, because many
tasks that IT administrators perform are iterations of the same task for different
users, sites, applications, and so on. However, it wasn’t until recently that this
concept has started to become mandatory.
At a certain scale, it’s no longer possible for humans to maintain the level of
accuracy and speed needed to meet the requirements of the business. Driven by
the cloud, automation of repetitive or broad modifications to the environment is
becoming commonplace, and will become far more commonplace in the future.
Why Automate?
There are two main purposes for implementing IT automation: speed and
accuracy. Both of these requirements ultimately line up to the needs of the
business as it relates to provisioning time and downtime aversion.
The first purpose is speed. Computers can issue commands much faster than
humans can manually enter them, thus even the most adept operator can be
easily outdone by a machine. Therefore, as the scope of the change or
implementation grows, so does the advantage to the machine over a human.
When even a slight misconfiguration can lead to downtime, this is a big deal. An
IT organization would likely be interested in automation for either one of these
purposes, but when both are at stake, automation is a no-brainer.
On the other hand, the data center is all about applications, and it’s high time
that infrastructure-focused staff spent a little more time focusing on
development.
Operations staff will be encouraged over the next few years to hone their
development chops and get more comfortable with code. By doing this, they’ll
not only become more efficient and effective, but have a better understanding of
the needs and methodology of the development staff.
In the data center over the next few years, users will be blessed with (and come
to expect) self-service provisioning of resources in a way that streamlines their
jobs and allows them to be more productive. This is a win for the internal
customers as they don’t have to wait on IT, and it’s a win for IT because they
won’t be burdened by the request and can continue to focus on being proactive.
Justifying Automation
A key challenge with IT automation is justifying the front-end investment of
time. It will be vital to every organization putting serious effort into automating
IT processes that they quantify and measure the return on investment (ROI) that
they’re getting from automation projects. At the end of the day, automation is
only valuable if the change it creates is directly reflected in the bottom line.
Following an automation project, the business will need to be able to say that,
“by implementing this automation project (in which we have invested 60 hours),
the business saved $120,000 this year.”
Orchestration
Orchestration can be seen as either an extension of or a complement to
automation. The idea of orchestration is that a single platform would control
many disparate systems to accomplish a defined business function.
67% of survey respondents said that relative to SDS and HCI, the REST API was one
of their key decision criteria. This clearly shows the importance current data center
architects and CIOs are placing on orchestration and the ability for these systems to
interact with other systems.
DevOps is the idea of pulling the operations team into the development
methodology to use their unique expertise to help the development process be
more rapid. It also includes implementing tools and processes that leverage all
the different skill sets available.
The development process isn’t the only way DevOps brings developers and
operations staff together. The entire life cycle of a product is given the DevOps
treatment; this means, for example, that operations and infrastructure staff are
involved in the application design process as well. Their understanding of the
infrastructure can lend a better perspective regarding server architecture,
performance, high availability, and so on.
Benefits of DevOps
Developers making a decision regarding application architecture without
understanding the underlying infrastructure is always a bad thing; and it is
equally unfortunate when systems architects make infrastructure decisions
without an understanding of the way an application functions. By placing the
two groups together on one team, collaboration is dramatically enhanced, which
leads to better decision making and project planning.
A byproduct of increased collaboration (and one of the main drivers for DevOps
adoption) is the elimination of a phenomenon known as technical debt.
Cutting corners.
The principle of technical debt says that these shortcuts only “borrow” time or
efficiency. However, the price is paid elsewhere.
For example, as more of these shortcuts and workarounds show up in the code
base or in the infrastructure, it becomes increasingly frustrating to work on. This
means IT staff can become unmotivated at best and disgruntled and malicious at
worst.
Technical debt is just like credit card debt; used responsibly, it can be a tool, but
it is more frequently used irresponsibly, which leads to more cost than if it had
not been used.
Placing the development, QA, operations, and other relevant team members in a
collaborative environment leads to better planning and more accurate
expectations. This, in turn, can and should directly lead to less technical debt
over time. This alone is reason enough for an organization to move toward a full
adoption of DevOps methodology.
Future IT organizations, once shifted to this methodology, will increase their rate
of integration to an alarming pace. For an example of an organization that is
already there, look no further than Amazon Web Services (AWS). They deploy
multiple times per minute!
Compare this to organizations using the waterfall methodology who can deploy
as rarely as every few weeks. When iterating at these dramatically different
intervals, imagine the impacts of a bad deployment. On the AWS side, a very
small section of the code base was impacted and the negative impact of the bad
commit is likely very small. DevOps made this possible. On the legacy
development side, a huge multi-week deployment going bad leaves a steaming
crater in the data center and leaves everyone scrambling to put out fires. The
advantage to the DevOps methodology in regard to this problem is pretty clear.
Industry Needs
As much as IT organizations are being forced to change by advancing
technology, there are other variables. Business is the motivation that drives the
change in the data center, but there are business requirements that are also
challenging the status quo. Performance and capacity can only go so far in
making a product a good fit for an organization. Beyond that, it needs to fit in to
the existing ecosystem of the IT environment. It needs to be a sensible business
decision from a budgetary perspective based on the cost of the product and the
desired outcomes to the business.
With these needs in mind, there are a minimum of three objectives that the IT
industry as a whole must meet to move toward.
In the case of open standards, this means that a few collaborative entities work
together to define a standard that all of their products will conform to. This
common blueprint ensures compatibility of components in an environment, even
if they’re coming from different manufacturers.
There’s more fuss in the IT industry now regarding open standards and
interoperability, and there’s a perfectly good reason for that. The various layers
of abstraction present (and being developed) in the data center are making
certain components a commodity. Decision makers have less loyalty to any one
manufacturer than was previously common, and instead choose based on
features and cost. In essence, vendors lose the ability to “own the whole stack”
and bake-in their proprietary intellectual property.
In sticking with the discovery theme, you could contrast an open standard like
LLDP with a proprietary standard like Cisco Discovery Protocol (CDP).
A proprietary standard like CDP has a major advantage to the manufacturer, but
a major disadvantage to the IT organization. The benefit of such a standard is
that the manufacturer (Cisco) can do whatever they want to with the technology.
This means they can develop it more rapidly, expand in any direction they want,
and discontinue support for the protocol if they want to. That all sounds pretty
good, so what’s the problem? Unfortunately, in a mixed environment like the
one above, the Brocade, Juniper, and HP switches may not be able to discover
the Cisco switches. You can clearly see the problem here for the IT organization:
vendor lock-in. Unless the organization buys all Cisco switches, the networking
picture becomes disjointed.
One final word on open standards: trust is harder to come by now than ever, and
in the future it will be more so. One of the benefits of open standards is that
they’re driven by a committee, anyone can participate, and anyone is free to
review the specifications. This has a major security implication, because hiding
malicious code or an “accidental” backdoor will be much harder when the whole
world can peer at the specifications.
Each level needs to make a profit, so they all purchase at a certain discount from
list price before the product ultimately makes it to the customer. The channel
partner ecosystem is immensely valuable to the IT industry, but this method of
selling can sometimes cause unintended harm.
There are multiple reasons why hiding the true cost of products can seem to
provide leverage. Many vendors don’t even publish a meaningless list price;
even worse, they don’t publish any price at all! Why? Because then a customer
has to ask for the price.
As soon as a customer asks, they’re paired with a salesperson who can engage
them and guide them through the buying process. Hiding list prices is a lead
generation tactic. While that may have worked well in the past, in the future, it’s
going to frustrate a generation of people who are accustomed to having an
overwhelming amount of information at their fingertips. Searching for
information and being unable to find it is very frustrating to this new generation.
Beyond list prices, the vendors who sneakily require certain important features
to be purchased as an add-on, and charge for expansion or software upgrades,
are expecting that each of these additions is an extra dose of revenue. This is a
fundamentally short-sighted approach. In the same way as the list price issue,
this practice will frustrate the customer (or potential customer) and ultimately
lead to a loss of business over time, even if not immediately.
This all seems pretty confusing, and that’s because it is. The buying process is
ultimately very complicated and slow due to all these details. Negotiating the
discounts to be applied between the vendor and the partner, and then between the
partner and the customer, can take time. The complicated ordering process can
cause confusion because sorting through 65 cryptic line items on a quote proves
to be too daunting for the potential customer. Future IT customers will not put
up with the productivity and time lost playing this game. This way of doing
business in IT needs to change, and a number of industry leaders are stepping up
to blaze a new trail in this regard. They’re posting true costs right on their
website. They’re creating an ordering process that involves two line items on a
quote instead of two pages’ worth. They’re making it easy for the customer to
buy the product.
Benefits of Transparency
What does the hardware or software manufacturer stand to gain by creating
transparency and clarity in their buying process? A whole list of words than any
company would love to be associated with like trust, loyalty, and
expeditiousness on the part of the buyer to name a few.
Customers who don’t feel like they’re having the wool pulled over their eyes are
likely to want to develop a healthy working relationship and are likely to
purchase right away.
If the buying process is made clear by a quoting process that is fast and simple,
the customer is likely to purchase quickly.
Manufacturers must start doing these two things in order to stay relevant:
Show the true costs and remove the games from the buying process,
while keeping in mind that each business needs to make money.
Make the purchasing process (quotes) simple and clear. Manufacturers
that are succeeding in this area commonly have a mere two or three
line items on a quote. These quotes might include a product (the single
line item which includes everything) and a couple of years of support,
for instance. Without all the SKUs that look like hieroglyphs and vague
descriptions, it’s easy for the customer to see what they’re getting into.
While it’s true that many products are being developed today that can achieve
things we barely thought possible only a few short years ago, it’s also still a
fairly common occurrence that a product doesn’t live up to the hype and
promises made by marketing folks and sales people.
61% of survey respondents indicated that they expect performance when deploying
SDS and HCI to be higher than that of their legacy infrastructure. This expectation
makes it critical that performance tests are performed in an acceptably unbiased and
realistic way.
Storage Benchmarking
One of the most obvious areas this problem shows up in is enterprise storage.
The tools commonly used to benchmark storage systems are confusing and
potentially ambiguous at best, and downright inaccurate if configured
improperly.
Part of the problem here is that many of the benchmarking tools generate a
synthetic workload pattern.
This means that in benchmarking, the storage platform will be serving a different
I/O characteristic than it will serve when it’s in production. This leads to
benchmark numbers being unreliable unless the benchmarking tool can
accurately simulate a production environment.
Other common pitfalls in storage benchmarking are: testing with a workload that
fits entirely in cache and building systems specifically for benchmarking that
aren’t the same as the system a customer will buy, even though they have the
same model number. The unfortunate fact is that many tools simply can’t
simulate a production environment with any degree of realism.
In the same way that open standards need to be adopted for the sake of
interoperability and the development of platforms, open benchmarking standards
need to be created and adopted. Some areas of IT already have these standards
while others don’t.
In the case of storage, this would be something like the specifications developed
by the Storage Performance Council (known as SPC-X where X represents a
certain performance test). SPC benchmarks don’t accurately test all types of
workloads, but they do provide common ground upon which any manufacturer
can test their product and compare it apples to apples with another vendor’s
competing product.
In the future of IT, publishing results based on standards like this will be a
requirement. To combat all the distrust and cynicism the industry has developed
in regards to performance in recent years, objective assessments are the only
way to gain trust in a claim moving forward.
Chapter Recap
In this chapter, you learned about the way the data center industry is being
transformed by a focus on automation, orchestration, and the DevOps
methodology. You also learned about some of the changing needs in the IT
industry. The following are a few key terms and concepts to take away from this
chapter.
DevOps is the idea of pulling the operations team into the development
methodology to use their unique expertise to help the development
process be more rapid. It also includes implementing tools and
processes that leverage all the different skill sets available.
In the next and final chapter, we’ll discuss the future of the data center. We’ll
take a look at what technologies and business considerations are shaping the
coming years in the data center.
11
This final chapter will expound upon the same principle that started this book in
Chapter 1: that computing power doubles every few years. This principle, better
known as Moore’s Law, has, and will continue to, radically change the data
center.
As computing power reaches new heights in the coming years, the most
successful organizations will be empowered by taking advantage of this.
Because of the exponential nature of Moore’s Law, failing to stay current will
more quickly result in obsolescence than ever before. That means that if the
technologies a competitor uses are a year or two newer than yours, they will
provide much greater performance and insight which could result in rapid loss of
your market share.
This chapter will explore the trends and technologies that are on the horizon and
will revolutionize data center architecture.
Keep in mind that as the cost of NVMe-based drives and Non-Volatile RAM in
general comes down, performance potential in the data center will skyrocket.
Also, as Intel continues to perfect the Rack Scale Architecture model, new
possibilities will emerge from a disaggregation and abstraction standpoint. This
means the data center will look at whole lot different in a few short years.
Future Trends
The recent evolution of the data center and IT as a business unit has been more
exciting than ever. However, the coming few years are likely to bring even more
excitement. Some of the technologies that are currently in development (and
could be in data centers within a few years) are simply mind boggling.
Some of the trends that will grow in the next few years include:
These will make the data center 3 to 4 years from now look quite different than it
does today.
For example, enabled by ever-increasing compute power, the amount of data that
will be stored and acted upon in the next decade huge. New application
architectures will be required, and dramatically enhanced hardware will be
created to support the new requirements.
Plus, the types of storage that have changed the data center in the past few years
(namely NAND flash) will likely be displaced by a successor even more suited
to the performance and capacity requirements of future IT.
Migrating to the cloud has garnered a lot of lip service lately, but relatively few
organizations are actually doing it compared to the number that are talking about
it. Soon, most organizations will adopt a cloud-focused model for provisioning
services. There will be abundant opportunities in the services industry for
helping customers migrate to the cloud.
However, LXC is not a new technology; it was first released in 2008 and is
viewed as relatively mature. In fact, containerization as a concept existed earlier
in Solaris Zones (2005) and AIX Workload partitioning (2007).
Containers are easy to deploy and consume very little resources as compared to a
virtual machine. Where virtual machines abstract operating systems from the
underlying hardware, containers abstract applications from the operating system.
From an efficiency standpoint, this is obviously a huge win; it also has the
benefit of ensuring consistency across the applications since they’re sharing the
dependencies rather than each application having a potentially disparate version
of the dependency.
LXC developers are now expanding on the platform with LXD, a new way of
interacting with containers which exposes a REST API. This will allow much
greater orchestration of containers moving forward.
As containers become more popular, the distribution of data center software will
likely focus more on containers than on virtual machines.
Tip from the Field by James Green
LXC is respected as a great project, but also has the potential to be circumvented by
projects like libcontainer due to their efficiency and potential for going cross-platform
in the near future.
— James Green
The OVA file contains a pre-built virtual machine with the vendor’s application
pre-installed. Compared to downloading the binaries and installing them
manually, downloading the full OVA carries the benefit of allowing the vendor
to make additional modifications such as including dependencies and tuning the
underlying operating system to suit the application.
Because of how lightweight and portable containers are, the future of the data
center will contain a shift toward applications being shipped as containers rather
than OVA/virtual machines like they are today.
The first obvious solution to this problem is to have at least one of each type of
guest OS required to run all the application. But this immediately becomes a
management nightmare if each one is a management silo.
In this way, any given container can be run on a Linux machine where
appropriate or on a Windows machine where appropriate, but all management is
done from a single platform.
The next logical step in solving this problem is to actually develop containers
that can run on multiple platforms. Meaning, a certain application could be run
on a Window host or a Linux host interchangeably. This is far away from reality
in common practice, but Microsoft and Docker have already demonstrated a
.NET application being run on both a Linux and Windows machine.
This endeavor will characterize the next phase of containerization in the data
center: allowing containers to be compatible across multiple OS platforms.
Open Source Tools
The pervasiveness of open source software in the modern data center is growing
very rapidly, as of late. But if open source software is so great, why isn’t it just
used for everything?
Well, that’s exactly what many of the web giants of the past decade have asked
as well. Google, Facebook, and the like are responsible for some of the most
popular open source projects being developed at present.
Open source software, due to its collaborative nature which spans verticals and
geographies, is commonly produced at a much faster pace and with higher
quality than a competitive proprietary project started around the same time.
For this reason, many IT organizations are choosing to focus their new initiative
on open source first, and proprietary software second if no open source option
fits the bill. Because of the highly competitive environment at the top of any
industry, being on the bleeding edge with open source software can give a non-
trivial advantage over a competitor who is at the mercy of a proprietary software
developer.
Moving into the future, open source software will be the default mode of most
organizations. It’s easy to see from examples like Docker and Hadoop that the
power of the community that can form around an open source project can extend
well beyond the capabilities of any single organization. That community of
contributors and committee members not only helps to develop the project with
exceptionally high quality, but they act in the best interest of the user base.
The final differentiator in the open source versus proprietary software debate in
many organizations turns out to be cost. No big surprise there. Community-
supported editions of open source software are commonly available for free or
for a nominal fee. At scale, this difference can really add up and make open
source a no brainer.
Flash Capacity
NAND flash has set new expectations for performance in the data center, but has
left something to be desired in terms of capacity. When NAND solid-state drives
(SSDs) first became available to consumers, a 30 GB SSD wasn’t small.
However, that’s far too small to be of much practical use to enterprise data
centers storing (potentially) petabytes of data.
Over the last five years, the capacity issue has been addressed to some degree,
but the breakthrough in flash capacity is just happening now.
Unfortunately, at this point the cells have become so small that creating smaller
ones is becoming quite challenging.
In the future of the data center, technology advances will allow the size of flash
cells to continue to decrease. This, in combination with the 3D design, will allow
SSDs to have acceptable capacities to all but eliminate the need for spinning,
magnetic disks.
It’s too fast in the sense that many SSDs can perform at substantially higher
read/write speeds than the bus and access specification is capable of providing.
The Past
In the recent past, storage has typically used the advanced host controller
interface (AHCI) standard for communicating with the rest of the system. The
problem is that this standard for accessing storage was designed with mechanical
hard drives in mind. Many SATA-based SSDs could perform higher than the
maximum throughput limit of the bus if the bus and interface standard could
handle it.
SSDs need faster access to the CPU and memory, and the best place to find that
is on the PCI Express (PCIe) bus.
The low latency and parallelism of the PCIe bus allow SSDs using NVMe to
achieve orders of magnitude higher performance. While AHCI has a single
command queue with a depth of 32 commands, NVMe has 216 (65,536) queues
with a depth of 216 commands per queue.
Also, AHCI allows devices to allocate a single interrupt, where NVMe allows
(via the MSI-X extension) devices up to 2,048 interrupts. Especially in the case
of random reads, this parallelism allows IOPS performance approaching double
that of the same drive using AHCI at certain queue depths.
The Future of NVMe
NVMe as a standard is still relatively new to the manufacturers working to
implement it. Development of the specification has been taking place for some
time; the project was underway by late 2009, the spec was released in early
2011, and the first commercially available drive to leverage NVMe was released
in 2012.
The year 2016 will likely be the year where mainstream adoption of the
technology takes place, and from there the price will start to come down to a
place that is more approachable for the average data center. Because the
economy of flash storage is largely related to cost per gigabyte, other
technologies like 3D NAND will make devices leveraging NVMe a more viable
option by increasing density.
Beyond Today’s Flash
While NAND flash has been in development for decades and is very mature at
this point, engineers are realizing that there are physical limitations to the
density NAND flash can be driven to.
There are several options that will likely emerge as the leader to replace NAND
down the road. Each of these technologies may sound like they’re straight from
Star Trek, but they have the potential to be in enterprise data centers in only a
few short years.
PCM has some of the same density challenges NAND does, but it has
substantially better endurance.
Resistive RAM
Another future non-volatile memory technology that could replace NAND is
called Resistive RAM (RRAM). This may also be referred to as ReRAM or
Memristor technology.
This type of memory functions on the concept of resistive switching, wherein the
resistance of a dielectric material changes based on electric current being
applied.
RRAM is possibly the most likely candidate to replace NAND due to very low
power consumption combined with high durability and performance.
Magnetroresistive RAM
A third contender for the future of flash is called magnetoresistive RAM
(MRAM). This technology is similar to PCM and RRAM in that it measures
resistance to determine the cell state (1 or 0).
What’s different about MRAM is that, rather than using an electrical charge to
represent state, MRAM uses a magnetic charge. The resistance of the material
will change based on the presence or absence of a magnetic charge.
However, there are some potential scalability problems with this technology, so
it’s not the strongest candidate.
Other Options
PCM, RRAM, and MRAM are far from an exhaustive list of the technologies
being explored to replace NAND flash in the future. It won’t be useful to
expound on all of the other options here, but research on developing non-volatile
memory technologies might also include:
Ferroelectric RAM
Racetrack Memory.
Each has its merits and challenges, but there is certainly no shortage of options
for the future of non-volatile memory.
Pooling of Resources
Non-volatile memory isn’t the only place where massive change is on the
horizon. Especially in large public cloud data centers, the standard container of
computing power could change in the near future. The unit of computing (as far
as a data center operator has been concerned) in recent years has been a server.
As you know, a server contains a CPU or two, a bit of memory, some network
interfaces, and perhaps some internal storage. A rack is loaded full of servers to
provide the cumulative computing power that the data center needs. But what if
the container used to contain the computing power was bigger than a server?
In massive data centers, this will allow for more efficiency and agility in exactly
the same way server abstraction via hypervisor did just a few years ago.
Imagine a standard data center rack. At the top are some very dense network
switches (perhaps 64 ports each) using the small and high-speed MXC
connector. This fabric uses silicon photonics technology to achieve multiple
terabits of throughput in and between racks. Below that sits a number of trays of
CPUs. Not whole servers, just the CPUs themselves. Below that, trays and trays
of memory in the same way. And at the bottom of the rack, some NVMe-
connected storage. Figure 11-3 illustrates a disaggregated architecture like the
one just described. This is what the future data center may look like. But why?
Figure 11-3: Disaggregated resources at the rack scale
Benefits of Disaggregation
What is to be gained by the disaggregation and pooling of components that were
previously contained to a server? Scalability and efficiency are certain to
increase. Plus, this new architecture would change the way that upgrades are
done.
Today, if an IT organization wants to use the latest and greatest CPUs, the server
is replaced with a new one that supports them. In an architecture where CPUs
were disaggregated from the rest of the system, the CPUs could be replaced
without having to replace the motherboard, memory, storage controllers,
network cards, and so on. This potentially makes upgrades quite a bit cheaper if
only a certain component needs to be upgraded. It will also make it much easier
for the IT staff doing the upgrade.
The Future
The key to achieving a mainstream uptake of the disaggregation of compute
resources into a larger pool will be manufacturer adoption and interoperability.
Just like with the other awesome technologies Intel develops, it only becomes
commonplace once other ecosystem vendors invest the time and effort into
creating compatible products. Since this architecture is vastly different than the
architecture that has defined data centers for years, this one could take a while.
There are still many phases of the migration to the cloud that businesses are still
working through. This is likely to be ongoing for a number of years because, just
like any major transition in architecture, it’s not typically a lift-and-shift sort of
move. It’s a gradual, opportunistic process where the changes in infrastructure
design begin to incorporate cloud components where it makes sense.
Hybrid Cloud
As a part of this gradual shift, a large number of data center infrastructures in the
coming years will take a hybrid cloud stance. Hybrid cloud means that a portion
of the IT resources exist in an on-premises data center and a portion exists in a
third-party cloud data center. But even in that hybrid approach, there are
questions yet to be answered.
How do you handle security in the cloud? How about making sure that data is
protected and can be restored in the event of a loss? Answering questions like
these will help you plan for and feel prepared to make the leap.
One of the ways this will take place is by abstracting the cloud resources in such
a way that the application or user doesn’t know or care where they’re coming
from.
Another way it will take place is by implementing solutions that leverage cloud-
based solutions to manage on-premises equipment. Some good examples of this
are managed OpenStack tools where the control mechanism lives in a public
cloud and manages on-premises equipment. Is that public or private cloud? Both,
perhaps? Similarly, network equipment that exists physically on-premises by is
managed entirely by a controller in a public cloud. This is neither an entirely
public nor a private architecture.
The future growth in this area will allow more agility and flexibility than has
been available in the past. By choosing the best location for any given
infrastructure component without being constrained to a certain environment,
more creative and available designs can take shape. In some multi-tiered
application use cases, the most optimal deployment involves running some front-
end components in a public cloud for scalability and availability purposes and
running the database tier in an on-premises data center.
Where a tiered application would likely run web servers in the cloud and keep
the storage in the on-premises data center, there are also certain use cases where
the path that makes the most sense is to run user-facing workloads in the data
center while the storage back-end lives in the public cloud.
Naturally, there are some concerns with this architecture. The primary concern
would be security, of course.
More than ever, companies are sensitive about their data’s security. Placing data
in someone else’s data center necessarily turns over control of security to them
as well. With the ever-growing number of data security breaches in a year,
organizations may be hesitant to adopt this sort of model.
What many of those organizations should consider is that the public cloud data
centers where their data would reside are likely held to even higher standards
than they are themselves. The public cloud may be safer than many perceive it to
be, due to providers being held to higher standards for things like encryption,
patch compliance, change control, and governance.
This choice and flexibility based on requirements at any one moment will be the
push that many organizations need to begin moving critical workloads to the
cloud.
There are various reasons why you might want to move a workload from an on-
premises data center to the cloud and back. For example, many data centers
charge based utilization and peak hours much like an energy company does
(cloud resources are commonly viewed as a utility, just like energy). It may
make sense to run a workload in an on-premises data center when peak hours
make running in the public cloud expensive, but then migrate the workload to
the public cloud during off-peak hours for better performance.
There are two ways to get past this in the future: either the applications will need
to be redesigned to be disaggregated in such a way that which data center they’re
running in is irrelevant, or manufacturers will create better shims.
In reality, both of these options will probably be utilized in the future. Innovators
will create applications that can move seamlessly; manufacturers will create
ways of abstracting the move from the application so that IT organizations stuck
on legacy software for various business reasons will still be able to access the
operational and financial advantages of seamlessly moving workloads between
the public and private cloud.
Whatever the reason, multiple cloud resources which providers use complicates
things to say the least. For this reason, in the future, more attention will be paid
to brokering and abstracting cloud resources. In the same way that data centers
today run workload placement engines to determine the best hypervisor and
storage volume to place a virtual machine on, data centers in the future will run
advanced workload placement engines to determine which cloud among the
organization’s multiple options is the best one to run a given workload in.
Rather than interacting with all the separate cloud options the organization has
available, the cloud abstraction platform will expose a set of APIs that other
infrastructure components can interface with to provide a single point of control
across all of the cloud platforms. The high-level overview gained by designing
the environment this way will also give organization insight into the health and
performance of their entire enterprise as opposed to distinct silos of cloud
resources.
Determining who is responsible for what, and making sure that all the gaps in
protection are covered between the two sides, is half the challenge of a
successful cloud data protection strategy. Another part of the challenge is
determining exactly what it is that must be protected.
The future of data protection in the cloud will involve customers answering a
very important set of questions. What infrastructure and application components
must be backed up? Who is responsible for backing up each of those
components? What is the method for backing them up, and where does the
backup data reside?
Obviously, data loss is never trivial. No matter how small the event is, losing
control of company data has an impact on the business. So caution in certainly
warranted. However, there are substantially more cases of data loss each year
where the breach was due to improper configuration by the customer rather than
a vulnerability that was the fault of the cloud provider.
The largest improvement in cloud security over the next couple of years will
likely not come from technology, but rather the people using it.
What is required for an enterprise to secure the resources running in the public
cloud is educating the teams that build, implement, and maintain the cloud
systems. From a technology perspective, the systems in use today are actually
quite secure. So, if the individuals and teams working on an IT organization’s
cloud resources understand the design considerations and technical
configurations required to secure their data in the cloud, security is increased
substantially.
Unfortunately, people are imperfect and mistakes will be made. Pride will cause
people to configure things they don’t know how to configure and leave gaping
security holes. This is why education will be so important.
The IT industry analyst firm Gartner speculates that by the year 2020, a full 95%
of cloud security incidents will be the fault of the customer and not the cloud
provider. This means that hesitant organizations who would love to take
advantage of a public cloud’s operational and financial advantages should
actually turn their focus inward. They’ll need to consider the policies and
practices that will make their teams successful at implementing their cloud
adoption strategy securely.
This isn’t to say that cloud security won’t improve from a technology standpoint
as well. One of the fastest growing security technologies in the coming years
will be the cloud control and accounting tools that help a large organization audit
and control access to their cloud resources. One of the challenges with software
as a service (SaaS) and platform as a service (PaaS) services is accurately
understanding who has access to what across the entire organization. These tools
will help organizations understand that in order to stay secure.
Chapter Recap
In this chapter, you saw a glimpse of what the future holds for the data center.
This is so important, because the changing data center will change the way the
rest of the business operates. Fortunately, the future looks very exciting! The
following are a few key terms and concepts to take away from this chapter.
No doubt, the blurring lines between public and private cloud will
change the way IT does business in the near future. There will be new
markets for cloud brokering and abstraction.
Due to increased cloud adoption, there will be a growing need for data
protection and security in the cloud.
Now that you’re equipped to transform your own data center, don’t wait to start.
Remember what Jack Welch said, “An organization’s ability to learn, and
translate that learning into action rapidly, is the ultimate competitive advantage.”
That exact advantage can come to your organization, but only if you start today!