EMC ProtectPoint For VMAX3 Overview - SRG
EMC ProtectPoint For VMAX3 Overview - SRG
Click the Notes tab to view text that corresponds to the audio recording.
Copyright © 1996, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009, 2010, 2011, 2012, 2013 ,
2014 EMC Corporation. All Rights Reserved. EMC believes the information in this publication is accurate as of its
publication date. The information is subject to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO
REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION,
AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR
PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable software
license.
EMC2, EMC, Data Domain, RSA, EMC Centera, EMC ControlCenter, EMC LifeLine, EMC OnCourse, EMC Proven, EMC
Snap, EMC SourceOne, EMC Storage Administrator, Acartus, Access Logix, AdvantEdge, AlphaStor,
ApplicationXtender, ArchiveXtender, Atmos, Authentica, Authentic Problems, Automated Resource Manager,
AutoStart, AutoSwap, AVALONidm, Avamar, Captiva, Catalog Solution, C-Clip, Celerra, Celerra Replicator, Centera,
CenterStage, CentraStar, ClaimPack, ClaimsEditor, CLARiiON, ClientPak, Codebook Correlation Technology,
Common Information Model, Configuration Intelligence, Configuresoft, Connectrix, CopyCross, CopyPoint, Dantz,
DatabaseXtender, Direct Matrix Architecture, DiskXtender, DiskXtender 2000, Document Sciences, Documentum,
elnput, ECS, E-Lab, EmailXaminer, EmailXtender, Enginuity, eRoom, Event Explorer, FarPoint, FirstPass, FLARE,
FormWare, Geosynchrony, Global File Virtualization, Graphic Visualization, Greenplum, HighRoad, HomeBase,
InfoMover, Infoscape, Infra, InputAccel, InputAccel Express, Invista, Ionix, ISIS, Max Retriever, MediaStor,
MirrorView, Navisphere, NetWorker, nLayers, OnAlert, OpenScale, PixTools, Powerlink, PowerPath, PowerSnap,
QuickScan, Rainfinity, RepliCare, RepliStor, ResourcePak, Retrospect, RSA, the RSA logo, SafeLine, SAN Advisor,
SAN Copy, SAN Manager, Smarts, SnapImage, SnapSure, SnapView, SRDF, StorageScope, SupportMate,
SymmAPI, SymmEnabler, Symmetrix, Symmetrix DMX, Symmetrix VMAX, TimeFinder, UltraFlex, UltraPoint,
UltraScale, Unisphere, VMAX, Vblock, Viewlets, Virtual Matrix, Virtual Matrix Architecture, ViPR, Virtual
Provisioning, VisualSAN, VisualSRM, Voyence, VPLEX, VSAM-Assist, WebXtender, xPression, xPresso, YottaYotta,
the EMC logo, and where information lives, are registered trademarks or trademarks of EMC Corporation in the
United States and other countries.
All other trademarks used herein are the property of their respective owners.
© Copyright 2014 EMC Corporation. All rights reserved. Published in the USA.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 1
This course covers the new ProtectPoint solution, including the customer challenges
ProtectPoint addresses, and the benefits ProtectPoint provides. The course also presents
the licensing, packaging, and ordering considerations for the ProtectPoint solution.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 2
Shown here is a generic, yet traditional, enterprise-class backup environment. In such an
environment, an application server will have its data stored on an enterprise class storage
array, often via a Storage Area Network (SAN) connection. The storage array may provide
benefits such as:
• Higher I/O – via striping of data
• Guarantees against single drive failure – via mirroring or parity striping
• Disaster Restart – via replication to another site (remote site not shown)
• Local replicas – via internal array copies, replicas may be used for testing of
software patches, for example
• And other potential benefits, which may be array specific
Though these features do quite a bit to keep data safe, they are not a complete data
protection solution. Near instant local and remote replicas may still become as flawed as
the original copy if, for example, a junior database administrator accidentally drops the
wrong database table. In such a case, the local and/or remote replicas would also have
dropped the table. In cases of data corruption, such as that which may be caused by a
virus infecting the application server, the near instant local and remote replicas would do
their jobs and copy the corruption, bit for bit. These are just two examples to show that
use of a storage array, by itself, is unlikely to be enough to fully protect enterprise data.
For these and other reasons, enterprises will make use [click] of a backup server. The
server will schedule backups and may coordinate with [click] storages nodes (not shown)
and/or media servers to migrate full, incremental, and differential backup data to, for
example, [click] a tape library or virtual tape library. Depending upon the media library in
use, backup data may be shipped or replicated to an off site location. In the event that the
backed-up data is needed, the same backup server will be used to locate and place restore
data onto the application server. As the application server is also the recovery target, and
a host to the SAN storage array, [click] it shall be referred to as the Application/Recovery
(A/R) Host .
Many enterprises will also have regulatory reasons to maintain backup copies.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 3
A problem with the typical backup scenario is that the A/R Host must be quiesced so that a
consistent backup may be made. This results in a state of reduced productivity or
[click]even a complete productivity pause for users of the application. Because a backup
may take quite some time, they are often scheduled for low-usage hours. This may
minimize, but not completely alleviate, the cost of making the backup. Put another way:
[click] backups take time, and time costs money.
To further help mitigate the negative effects of making a backup, the power of the storage
array may be leveraged. In this manner, the A/R Host only needs to be paused (users lose
access) long enough [click] for an internal copy to be made. Then [click] the backup can be
created from the internal copy while the application can be made available for its users.
There remain difficulties with this scenario. In many cases, the backup administrator must
coordinate backup timing with not only the application administrator but also with the
storage administrator. This is to guarantee that the application is quiesced (for data
consistency) , and that the internal copy is made, and presented for backing up. These
backups become known good, point-in-time copies. So, in the case of an accidental
deletion of data, one may use the backup data to restore data which was lost.
If a full backup is made of the copied volume, the backup process can take quite some
time, and quite an amount of storage space will be needed to store the backup. Performing
differential backups (where only the data since the last full backup is copied), or
incremental backups (where only the data since the last incremental or full backup is
copied) can help reduce the time for backup creation, and reduce the storage space
needed. However, both differential and incremental backups come at the cost of slower
restore operations. This is because both backup types require that a full backup image
must be loaded, and then the incremental steps (this could be several backup sets) be
applied, or in the case of a differential backup, the full backup must have the single
differential changes applied before performing the restore operation.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 4
ProtectPoint leverages capabilities of the VMAX 3 and Data Domain storage systems to
eliminate most of the challenges found in a traditional enterprise backup environment.
ProtectPoint also puts control of the backup into the hands of the A/R Host administrator.
Later in this course, we will take a deeper look at the specific storage system capabilities
that synergistically allow creation of the ProtectPoint solution. This will also explain how
two data devices shown here [click] are really the same device.
Data Domain storage systems perform inline deduplication. This means that data is
deduplicated as it enters the storage system and before it is written to disk. Deduplication
of data eliminates redundancies in the data by storing only a single copy of redundant data
and using pointers to that copy for any later duplicates.
The VMAX3 storage systems are able to maintain tables of disparity between associated
devices. This is known as changed block tracking. Once a device is associated as a replica
of a production device, changes to the production device are tracked, and the replica may
be [click] updated by scanning through the table and only copying those blocks that were
changed on the production device. Note that this is not device mirroring, as the production
device may be updated independently of the replica device.
To perform a ProtectPoint backup:
• The A/R Host administrator initiates a backup at a consistent checkpoint
• VMAX3 copies changed blocks directly to Data Domain device
• Data Domain creates and maintains full backups in native format
Because the backup data resides in a Data Domain storage system, the data is
automatically deduplicated and the storage-space pain experienced in a traditional
environment is greatly alleviated.
Some other notable benefits the ProtectPoint solution brings:
• A reduction by 20 times in the time needed to create a full copy of the production
data
• Near zero disruption to A/R Host
• Reduced infrastructure cost to achieve backup Service Level Agreements (SLAs)
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 5
As most Data Domain storage systems operate on the same basic principles, think of the
model shown as being a Generic Data Domain storage system [click] capable of being used
by either Linux/Unix or Windows hosts. The Data Domain storage system has its own
processors and memory, for ingesting and deduplicating data, as well as numerous hard
drives for storing that data on.
For storing data, a file system is built upon the drives within the Data Domain storage
system. Data Domain storage systems use Redundant Array of Independent Disks (RAID)
and a Data Invulnerability Architecture (DIA) to provide the stored data with additional
layers of protection. More details on RAID and DIA are outside the scope of this course.
A Linux/Unix host also uses a file system (for example: ext2, ext3, ext4, Reiser4, ZFS,
etc.), but its details are unimportant to this discussion. [click]Typically, the file system is
organized in a tree structure, branching out from a root directory, represented by a forward
slash ( / ). Adding another disk drive, or CD-ROM, or other storage becomes a simple
matter of logically attaching the new storage to the tree structure. This is called mounting.
For a Linux/Unix host to use a Data Domain storage system, the Data Domain storage
system [click] must present a folder across the Local Area Network (LAN) for access. This
folder [click] may then be mounted to the Linux/Unix file system tree, and used as though
it was native, local storage.
A Windows host also uses file systems (for example: FAT, FAT32, NTFS, etc.), but typically
has one independent file system per logical hard drive., i.e. per hard drive letter. In
Windows, adding another disk drive, or CD-ROM, or other storage usually involves creating
a [click] new drive letter for the Windows host to access. This is called drive mapping. If
the Windows host desires access to a Data Domain storage system, the Data Domain
storage system [click] must present a folder across the LAN for access. This folder [click]
may then be mapped to the Windows host, and used as though it was a native, local hard
drive already formatted with a file system.
Two different protocols are used for exporting a folder to be mounted, and exporting a
folder to be mapped. Linux/Unix hosts most often make use of the Network File System
(NFS) protocol. Windows hosts most often use the Common Internet File System (CIFS)
protocol. Details of these protocols are outside the scope of this course.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 6
Like a Data Domain storage system, a VMAX3 storage system has its own processors and
memory, as well as numerous hard drives for storing data on. A typical VMAX3 storage
system also uses RAID to protect stored data against single hard drive failure, as well as to
provide faster access to the data than what a single hard drive could offer.
Unlike a Data Domain storage system, a VMAX3 storage system does not build a file
system upon its hard drives. Instead, it uses [click] portions of each drive to form a new,
logical device. That device, [click] via the Fibre Channel protocol, can be presented to a
variety of hosts (Linux, Unix, Windows, etc.) for their storage use.
While this can be accomplished by directly cabling a host to the VMAX 3 storage system, it is
more often accomplished through the use [click] of a Storage Area Network (SAN). In a
SAN, switches (not shown) are configured with zones in such a way that [click] presented
logical devices are only visible to the designated hosts. In this way, a Linux host is kept
completely unaware of a device that is used by a Windows host, and therefore can not
cause any accidental damage to the data on that device. Except in special circumstances,
the SAN zoning allows only a single designated host to “see” any one particular logical
device presented by the VMAX3.
So, similar to a Data Domain storage system exporting a folder across a LAN for a host to
use, a VMAX3 presents a logical device across a SAN for a host to use. [click]The host
which is given access to the logical device thinks the device is a raw hard disk. The host
may then use the disk device however is desired. (Examples: typically an NTFS file system
will be build by Windows hosts, and perhaps an ext3 file system by Linux hosts might be
built upon a different logical device.)
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 7
An A/R host that is to participate in a ProtectPoint environment will have [click] at least one
Host Bus Adapter (HBA) for connecting [click] to a SAN. The A/R Host will also have [click]
a Network Interface Card (NIC) for connecting [click] to a LAN.
Through the SAN, the A/R host can make use of [click] a logical device for some of its
storage needs. This device is presented through the SAN [click] by a VMAX 3 storage
system. The A/R Host, especially if running a large application [click] like Oracle, will likely
need more [click] storage than will be provided by a single logical device.
Through the LAN, the A/R Host can communicate with [click] users or administrators,
and[click] the Data Domain storage System.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 8
Now that we have seen, from a high level, the components which participate in the
ProtectPoint solution, let us look at how those components are brought together to
cooperate in providing the ProtectPoint solution.
Within the Data Domain storage system a disk image [click] is created and stored in the file
system. Because this disk image is stored in the Data Domain file system, it is
automatically deduplicated, and changes or updates to it will also be deduplicated. This
disk image is called a Data Domain vdisk, and the Data Domain storage system can use a
SAN attachment [click] to present the image through the SAN in much the same manner as
a VMAX3 presents a logical device. In fact, like the VMAX3, the Data Domain storage
system [click] can present numerous vdisks through the SAN.
The VMAX3 storage system has an ability to connect to the SAN, but acting as a typical
host. In this way, the VMAX3 storage system can make use of logical devices presented
through the SAN by a separate storage system, [click] in this case a Data Domain storage
system. Of course, like any other host, it treats these logical devices as native, local
storage[click]. This is known as device encapsulation, as the external logical device
becomes logically encapsulated within the VMAX 3.
To prepare these newly encapsulated devices for use by ProtectPoint, each should be
associated with the production device that it is to backup. Once associated, a consistent
copy of the production devices can be made onto the encapsulated devices. This is known
as Backup Zero.
After preparing the encapsulated device with an initial copy of the application data, the
VMAX3 change block tracking mechanism may be leveraged. Then the application
administrator can initiate a backup, at a consistent checkpoint [click], and changes made by
the A/R Host can be copied to the encapsulated devices. As these devices actually reside
through the SAN, and on a Data Domain file system, the data is automatically deduplicated.
Backups performed by the application administrator, using ProtectPoint, only need to move
the blocks of data that changed since the previous backup. This gives the speed of a more
traditional incremental backup. As the full encapsulated device exists, it is truly a full
backup. By leveraging the FastCopy technology within the Data Domain storage system,
these full backups remain available for later restore operations. Any redundant data
spanning the various full backup images is automatically deduplicated and compressed by
the Data Domain storage system, thereby granting an incredible storage space savings.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 9
Once configured, the process for achieving a full backup is fairly simple. [click]An
application owner initiates a backup at an application consistent checkpoint. ProtectPoint
instructs the VMAX3 to [click]copy changes from the protected devices to the encapsulated
devices each is associated with. For simplicity, only one device is shown here, in a real
environment a group of devices is a more likely scenario.
Remember, as the encapsulated device actually exists inside the Data Domain storage
system, this copying of data really sends the data through the SAN to the Data Domain
storage system, where the changes are deduplicated and applied to a FastCopy (pointer-
based) image of the previous backup for retention.
Later, when another backup is initiated, [click] the process repeats. In this manner, the
Data Domain storage system maintains multiple copies of full image backups in their native
format, each backup results in a complete static image of the production volume(s) at the
time the backup was made. Any of these copies may be selected, and presented through
the SAN, for data recovery purposes at a later time.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 10
With a remote, or secondary, site [click] similarly configured to the primary site[click], Data
Domain storage systems can use IP via LAN or Wide Area Network (WAN) [click] to
efficiently replicate the backup images[click]. For data recovery or disaster restart testing
purposes, [click]any of the replicated images can be presented to the remote VMAX 3 using
the SAN at the remote site.
Whether or not the VMAX3 system also replicates its primary production (non encapsulated)
devices is immaterial to, and outside the scope of, the ProtectPoint solution.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 11
Backup images may reside, at either a primary or a remote site. A data restore may be
accomplished at either site. This scenario will be the same for restore of production data
that was lost, or for restore done for testing purposes.
• A backup static image to restore from is selected [click] and presented to the VMAX 3,
through the SAN, for encapsulation
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 12
Listed here are some things to keep in mind regarding ProtectPoint workloads:
• If using an application that spans multiple devices, as Oracle does, be sure to group the
devices together.[click] This way, the inter-dependent device consistencies will be
maintained correctly in the backup set.[click] As the application initiates the backup at a
consistent checkpoint (at the direction of the application administrator, of course), the
various inter-dependent devices should be consistent as a set.
For more information about the ProtectPoint solution, its management, restrictions and
limitations, please refer to the official documentation:
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 13
There are a few key points to be aware of regarding the packaging, licensing, and ordering
of ProtectPoint.
[click]
• The cost for a ProtectPoint license is based on registered capacity, as like other offerings
from EMC’s Core Technologies Division (CTD). This simplifies license management for
customers.
[click]
• ProtectPoint may be added to an order for new storage (Data Domain or VMAX3).
[click]
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 14
There are new fixed services for VMAX3 and ProtectPoint. These are offered through both
the CTD and VMAX3 DxP/CxP paths, and they are:
• Implementation services (required)
and
• Oracle Integration services (recommended)
• Oracle Integration is offered in Base + Add-On model
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 15
From a high level, here are the steps for the Professional Services Implementation Process:
• To install and test a ProtectPoint backup solution, [click]there must be application
servers, using storage on a new VMAX3, which was designed and ordered for a
ProtectPoint implementation.
• As defined, the ProtectPoint installation is incremental
– [click]Install the Data Domain storage system
LAN Connectivity
SAN Connectivity
– [click]Setup VMAX3 to Data Domain relationship
Create needed zoning in the fabric(s)
Create vdisks
Present vdisks from the Data Domain storage system to the VMAX 3 storage
system
– Remember: the VMAX3 DX ports should have been defined and zoned.
Encapsulate the vdisks
– [click]Setup production and backup device relationship(s)
– [click]Create: Backup Zero
– [click]Test ProtectPoint backup
– [click]Customer Transfer of Information (TOI) – commands, usage, etc.
– Turn the environment over to the customer for production use
• Customer now has ProtectPoint Solution operational; Professional Services’
engagement is complete.
A Greenfield account would follow the same basic process, though, obviously, the initial A/R
Host using VMAX3 storage would become a very early part of the implementation.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 16
This course covered the new ProtectPoint solution, including the customer challenges
ProtectPoint addresses, and the benefits ProtectPoint provides. The course also reviewed
the licensing, packaging, and ordering considerations for the ProtectPoint solution.
This concludes the training. Proceed to the course assessment on the next slide.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 17
Do not record. Feedback Link will be inserted by Production Team.
Copyright 2014 EMC Corporation. All rights reserved. EMC ProtectPoint for VMAX3 Overview 18