Data Center Principles and Strategies 6
Data Center Principles and Strategies 6
presentation
BITS Pilani WILP
Pilani Campus
BITS Pilani
Pilani Campus
CS 06
Books
Text Book(s)
No Author(s), Title, Edition, Publishing House
T1 Building a Modern Data Center: Principles and Strategies of Design
by Scott D. Lowe (Author), David M. Davis (Author), James Green (Author),
Seth Knox (Editor), Stuart Miniman (Foreword)
• First and foremost, SDS is an abstraction from the physical storage that
is being managed.
• It includes a type of storage virtualization akin to the way compute
virtualization makes virtual machines independent of the underlying
physical hardware.
• This is very important, because the strength of SDS is its flexibility.
• That flexibility is made possible by abstraction.
• The requirement to provide abstraction does not mean that SDS can’t be
a way of consuming storage from a more traditional, monolithic storage
array.
• SDS is commonly associated with hyperconvergence; however, that’s
only one of many ways that SDS can be leveraged.
• An SDS layer can provide the method for managing, automating, and
scaling an already specialized storage solution.
• That abstraction is typically found in one of two implementation types.
• The first type is a virtual appliance deployed in the infrastructure.
• This virtual appliance contains the software that provides and manages the
SDS platform and abstracts the storage behind it from the workloads in
front of it.
• The other method is kernel-level storage virtualization.
• Rather than running in a virtual machine, software runs on the hypervisor
itself to provide the storage features of the SDS platform.
• It might say that the workload must reside on Tier 2 storage (the
qualifications for Tier 2 having been previously defined by the administrator).
• Imagine applying these specific settings to one virtual machine a single time.
• The task is not incredibly daunting, given the right software.
• However, imagine applying these same settings to 1,000 virtual machines in
an environment where six new virtual machines are provisioned each week.
• It’s only a matter of time before mistakes are made, and with each new
virtual machine an administrator will burn time setting it up.
• With policy-driven SDS, simply by having applied the policy (created once),
the virtual machines will be treated exactly as desired with accuracy and
consistency over time.
• APIs are necessarily present in SDS because the SDDC as a whole uses
some sort of orchestration engine to make all the pieces work together.
• Each node contains compute, storage, and networking; the essential physical
components of the data center.
• From there, the hyperconvergence platform pools and abstracts all of those
resources so that they can be manipulated from the management layer.
BITS Pilani, Pilani Campus
Simplicity
• Makers of hyperconverged systems place extreme amounts of focus on making
the platform simple to manage.
• If managing compute, storage, and networking was complicated when they were
separate, imagine trying to manage them at the same complexity but when
they’re all in one system.
• It would be a challenge to say the least.
• This is why the most effective hyperconvergence platforms take great care to
mask back-end complexity with a clean, intuitive user interface or management
plugin for the administrator.
• By nature, hyperconvergence is actually more complex than traditional
architecture in many ways.
• The key difference between the two is the care taken to ensure that the
administrator does not have to deal with that complexity.
• Or the platform may use specialized hardware to provide the best reliability or
performance.
• Neither is necessarily better, it’s just important to know the tradeoffs that come
with each option.
BITS Pilani, Pilani Campus
Software or Hardware?
• If special hardware is included, it dramatically limits your choice with regards to
what equipment can be used to run the platform.
• But it likely increases stability, performance, and capacity on a node (all else
being equal).
• The opposite view is that leveraging a VSA [ Vector Signal Analysis] and no
custom hardware opens up the solution to a wide variety of hardware
possibilities.
• While flexible, the downside of this approach is that it consumes resources from
the hypervisor which would have served virtual machine workloads in a
traditional design.
• This can add up to a considerable amount of overhead.
• Which direction ends up being the best choice is dependent on myriad variables
and is unique to each environment.
• It’s important to realize how much software defined storage (SDS) technology
makes the concept of hyperconvergence infrastructure (HCI) possible.
• If SDS didn’t exist to abstract the physical storage resource from the storage
consumer, the options left would be the architectures that have already been
shown to be broken.
• Namely, those architectures are silos of direct attached storage and shared
storage in a monolithic storage array.
• Pooled local storage has advantages over both of those designs, but would
not be possible without the help of SDS which performs the abstraction and
pooling.
• One of the main advantages of pooled local storage is a highlight of the
hyperconvergence model in general: the ability to scale the infrastructure with
building blocks that each deliver predictable capacity and performance.
• Hyperconvergence has SDS to thank for the fact that as this infrastructure
grows over time, the storage provided to workloads is a single distributed
system (an aggregation of local storage) as opposed to an ever-growing stack
of storage silos.
• Most hyperconverged platforms offer the ability to apply data protection and
performance policies at a virtual machine granularity.
• This capability is also a function of the SDS component of the hyperconverged
system.
• Policy from the management engine interacts with the SDS interface to apply
specific changes to only the correct data.
• This granularity, again, would not be possible without software defined storage.
• There are many things that go into making a hyperconverged model successful,
but one component that hyperconvergence absolutely could not be successful
without is flash storage.
• The performance capabilities of modern flash storage are the only reason it’s
possible to attain acceptable performance from a hyperconverged platform.
• In a legacy monolithic storage array, there was one way of achieving additional
performance for quite some time: add more disks.
• Each disk in a storage array can serve a certain amount of data at a time.
• This disk performance is measured in I/O Operations per Second (IOPS).
• In other words, how many individual I/O requests (reads or writes) can the disk
complete in one second.
• As spinning disks have ceased to increase in rotational speed, the fastest
spinning disks topped out somewhere between 160 and 180 IOPS.
• There are a number of different disk configurations that you might see used in
hyperconvergence (Figure 3-1):