Module4
Module4
Contents:
and takeaways
Module overview
1 of 93 3/12/2019, 12:38 PM
about:blank
Objectives
After completing this module, you will be able to:
• Describe and implement the Storage Spaces feature in the context of enterprise
storage needs.
Lesson objectives
After completing this lesson, you will be able to:
2 of 93 3/12/2019, 12:38 PM
about:blank
• Describe common usage scenarios for storage spaces, and weigh their benefits
and limitations.
Enterprise needs
3 of 93 3/12/2019, 12:38 PM
about:blank
When planning your balanced storage deployment approach to meet your storage
needs, you will need to assess your capacity and performance requirements in
relation to your cost. For cost efficiency, your storage environment should utilize
solid-state highly active data (higher cost) and
hard data accessed infrequently ( the cost).
4 of 93 3/12/2019, 12:38 PM
about:blank
higher performance, but with lower capacity. As a result, your balanced storage
deployment approach will most likely include a mix of HDDs and SSDs to achieve the
best performance and capacity at the appropriate cost.
Included in your storage planning, you should consider whether your storage solution
needs to support the common capabilities of most storage products, such as:
• Mirror/parity
• Data
• Enclosure
• Storage tiering
• Storage replication
• Data
• Data
• Performance
Note: This list is only meant to provide suggestions and is not an exhaustive
list of the common capabilities of most storage products. The storage
requirements of your organization might differ.
5 of 93 3/12/2019, 12:38 PM
about:blank
Discovery
Which factors should you consider when planning your enterprise storage strategy?
Discovery
What your organization
Show solution
6 of 93 3/12/2019, 12:38 PM
about:blank
• Storage pools. Storage pools are a collection of physical disks aggregated into a
single logical disk, allowing you to manage the multiple physical disks as a single
disk. You can use Storage Spaces to add physical disks of any type and size to a
storage pool.
• Storage spaces. Storage spaces are virtual disks created from free space in a
storage spaces have attributes such storage
tiers, precise administrative advantage
of no longer need to Instead, you
can unit. Virtual disks are logical unit
number ( SAN.
Note: The virtual disks that you create with the Storage Spaces feature are
not the same as the virtual hard disk files that have the .vhd and .vhdx file
extensions.
7 of 93 3/12/2019, 12:38 PM
about:blank
o Disks must be blank and unformatted, which means no volumes can exist on
the disks.
o Disks can be attached using a variety of bus interfaces including SAS, SATA,
SCSI, Non-Volatile Memory Express (NVMe), and universal serial bus (USB). If
you plan to use failover clustering with storage pools, you cannot use SATA,
• Virtual disk or storage space. This is similar to a physical disk from the perspective
of users and applications. However, virtual disks are more flexible because they
include both fixed provisioning and thin provisioning, also known as just-in-time
(JIT) allocations. They are also more resilient to physical disk failures with built-in
functionality mirroring and parity. These Array of
Independent technologies, but Storage data
differently
Note: When planning your Storage Spaces deployment, you need to verify
whether the storage enclosure is certified for Storage Spaces in Windows
Spaces to identify array’s
failure lights, the array must Enclosure
8 of 93 3/12/2019, 12:38 PM
about:blank
You can format a storage space virtual disk with an FAT32 file system, New
Technology File System (NTFS) file system, or Resilient File System (ReFS). You will
need to format the virtual disk with NTFS if you plan to use the storage space as part
of a Clustered Shared Volume (CSV), for Data Deduplication, or with File Server
Resource Manager (FSRM).
Feature
• Simple. A simple space has data striping but no redundancy. In data striping,
logically sequential data is segmented across several disks in a way that enables
9 of 93 3/12/2019, 12:38 PM
about:blank
Feature Description
Note: The number of columns for a given storage space can also impact the
number of disks.
10 of 93 3/12/2019, 12:38 PM
about:blank
Feature Description
• If the list contains at least one 4-kilobyte (KB) drive, the pool sector size is set to
4 KB.
Cluster disk Failover clustering prevents work interruptions if there is a computer failure. For a pool to
requirement clustering, all drives in the pool
11 of 93 3/12/2019, 12:38 PM
about:blank
Feature Description
Stripe You can increase the performance of a virtual disk by striping data across multiple
parameters physical disks. When creating a virtual disk, you can configure the stripe by using two
parameters, NumberOfColumns and Interleave.
• Interleave represents the amount of data written to a single column per stripe.
The NumberOfColumns and Interleave parameters determine the width of the stripe (e.g.,
stripe_width = NumberOfColumns * Interleave). In the case of parity spaces, the stripe
width determines how much data and parity Storage Spaces writes across multiple disks
to increase performance available to apps. You can control the number of columns and
when creating a new virtual PowerShell
New-VirtualDisk with the NumberOfColumns parameters.
• Fault tolerance. Do you want data to be available in case a physical disk fails? If
so, you must use multiple physical disks and provision virtual disks by using
mirroring
12 of 93 3/12/2019, 12:38 PM
about:blank
• Reliability. Virtual disks in parity layout provide some reliability. You can improve
that degree of reliability by using hot spare physical disks in case a physical disk
fails.
• Extensibility. One of the main advantages of using Storage Spaces is the ability to
expand by adding physical physical disks
to after you create it to capacity or to
provide
• Create volume.
Demonstration steps
1. In the VIRTUAL DISKS pane, create a New Virtual Disk with the following
13 of 93 3/12/2019, 12:38 PM
about:blank
settings:
Thin
Simple Volume
14 of 93 3/12/2019, 12:38 PM
about:blank
File and storage services includes technologies that help you deploy and manage
one
• Storage Spaces Direct. This feature enables you to build highly available storage
systems by using storage nodes with only local storage. You will learn more about
this module.
15 of 93 3/12/2019, 12:38 PM
about:blank
• Storage Quality of Service (QoS). With this feature, you can create centralized
QoS policies on a Scale-Out File Server and assign them to virtual disks on
Hyper-V virtual machines. QoS ensures that performance for the storage adapts
to meet policies as the storage load changes.
• Data Deduplication. This feature was introduced in Windows Server 2012 and is
improved 2016 in the following information about
Data covered later in this module):
o up to 64 terabytes ( been
Server 2016 and is able to
CPU’s per volume to increase throughput rates on
volume sizes up to 64 TB.
o Support for file sizes up to 1 TB. With the use of new stream map structures
and other improvements to increase optimization throughput and access
performance, deduplication in Windows Server 2016 performs well on files up
o Support for Nano Server. A new deployment option in Windows Server 2016,
Nano Server fully supports Data Deduplication.
• Support for cluster rolling upgrades. You can upgrade each node in an existing
Windows Server 2012 R2 cluster to Windows Server 2016 without incurring
downtime nodes at once.
16 of 93 3/12/2019, 12:38 PM
about:blank
Note: The registry values for these settings aren’t present by default;
however, the hardening rules still apply until Group Policy or other registry
Windows Server 2012 R2 and Windows Server 2012 offered several new and
improved file and storage-services features over its predecessor, including:
• iSCSI Target Server. The iSCSI Target Server provides block storage to other
servers and applications on the network by using the iSCSI standard. Windows
Server 2012 R2 also includes VHDX support and end-to-end management by
using Management Initiative Specification.
17 of 93 3/12/2019, 12:38 PM
about:blank
• Unified remote management of File and Storage Services in Server Manager. You
can use the Server Manager to manage multiple file servers remotely, including
their role services and storage.
• Windows PowerShell cmdlets for File and Storage Services. You can use the
Windows PowerShell cmdlets for performing most administration tasks for file and
storage
• Server Message Block (SMB) 3.0. SMB protocol is a network file-sharing protocol
that allows applications to read and write to files and request services from server
programs on a network.
18 of 93 3/12/2019, 12:38 PM
about:blank
When considering whether to use Storage Spaces in a given situation, you should
weigh and limitations. The was
designed administrators to:
• Aggregate individual drives into storage pools, which are managed as a single
entity.
• Use different types of storage in the same pool ( SATA, SAS, USB, SCSI).
• Grow required.
19 of 93 3/12/2019, 12:38 PM
about:blank
• Use the existing tools for backup and restore and Volume Shadow Copy Service
(VSCS) for snapshots.
• The contents of a drive are lost when you introduce that drive into a storage pool.
o a minimum of two
20 of 93 3/12/2019, 12:38 PM
about:blank
• Storage layers that abstract the physical disks are not compatible with Storage
Spaces, including:
o deployed in a separate
Mirror mirror)
mirror)
Parity units of
21 of 93 3/12/2019, 12:38 PM
about:blank
Storage Spaces Direct removes the need for a shared SAS fabric, simplifying
deployment and configuration. Instead, it uses the existing network as a storage
fabric, leveraging SMB 3.0 and SMB Direct for high-speed, low-latency CPU efficient
storage. To scale out, you simply add more servers to increase storage capacity and
I/O performance.
You also can deploy Storage Spaces Direct in support of SQL Server 2012 or newer,
22 of 93 3/12/2019, 12:38 PM
about:blank
which can store both system and user database files. SQL Server is configured to
store these files on SMB 3.0 file shares for both stand-alone and clustered instances
of SQL Server. The database server accesses the SOFS over the network using the
SMB 3.0 protocol. This scenario requires Windows Server 2012 or newer on both the
file servers and the database servers.
You can use Storage Spaces inside an Azure virtual machine to combine multiple
virtual hard drives, creating more storage capacity or performance than is available
from a single Azure virtual hard drive. There are three supported scenarios for using
Storage Spaces in Azure virtual machines, but there are some limitations and best
practices follow, as described below.
Multi-tenant scenarios
Storage Spaces can be made visible only to a subset of nodes in the file cluster. This
23 of 93 3/12/2019, 12:38 PM
about:blank
can be used in some scenarios to leverage the cost and management advantage of
larger shared clusters and to segment those clusters for performance or access
purposes. Additionally, you can apply ACLs at various levels of the storage stack (for
example, file shares, CSV, and storage spaces). In a multitenant scenario, this
means that the full storage infrastructure can be shared and managed centrally and
that you can design dedicated and controlled access to segments of the storage
infrastructure. configure a particular customer storage pools,
storage volumes, and file and
ACLs tenant has access
24 of 93 3/12/2019, 12:38 PM
about:blank
Discovery
What are the advantages of using Storage Spaces compared to using SANs or NAS?
Check
Discovery
What are the disadvantages of using Storage Spaces compared to using SANs or NAS?
25 of 93 3/12/2019, 12:38 PM
about:blank
Discovery
In what scenarios would you recommend each option?
Lesson objectives
After you will be able to:
• Describe how to use event logs and performance counters to monitor Storage
Spaces.
Managing Spaces
26 of 93 3/12/2019, 12:38 PM
about:blank
Storage Spaces is integrated with failover clustering for high availability, and
integrated volumes (CSV) for can
manage using:
• Server
• Windows PowerShell
Manage Manager
Server Manager provides you with the ability to perform basic management of virtual
disks and storage pools. In Server Manager, you can create storage pools; add and
27 of 93 3/12/2019, 12:38 PM
about:blank
remove physical disks from pools; and create, manage, and delete virtual disks. For
example, in Server Manager you can view the physical disks that are attached to a
virtual disk. If any of these disks are unhealthy, you will see an unhealthy disk icon
next to the disk name.
more information, in
http://aka.ms/po9qve
cmdlets in Windows download
module for use in Windows more
“Storage Spaces Cmdlets PowerShell” at:
ms/M1fccp
28 of 93 3/12/2019, 12:38 PM
about:blank
When planning for storage tiering, you should assess the workload characteristics of
your storage environment so that you can store your data most cost-effectively
depending on how you use it. In Windows Server 2016, the server automatically
optimizes your storage performance by transparently moving the data that's
accessed more frequently to your faster solid state drives (the SSD tier) and moving
less active data to your less expensive, but higher capacity, hard disk drives (the
HDD
Additionally, when planning for storage tiering, you should assess if there are
situations in which a file works best when placed in a specific tier. For example, you
need to place an important file in the fast tier, or you need to place a backup file in
the slow tier. For these situations, your storage solution might have the option to
assign a file to a particular tier, also referred to as pinning the file to a tier.
29 of 93 3/12/2019, 12:38 PM
about:blank
• Don't allocate all available SSD capacity for your storage spaces immediately.
Keep some SSD capacity in the storage pool in reserve, so you can increase the
size of an SSD tier when a workload demands it.
• Don't pin files to storage tiers until you see how well Storage Tiers Optimization
can optimize storage performance. When a tenant or workload requires a
particular performance, you can pin files ensure that all
I/O that tier.
You should use the Storage Tier Optimization Report when observing or monitoring
your used to check the storage tiers
and might optimize their the
performance report provides data for such as,
“How and “How much do capacity?”
30 of 93 3/12/2019, 12:38 PM
about:blank
Before deployment, you should plan Storage Spaces to handle disk and JBOD
enclosure impact on service loss. With
any storage should expect that hardware is
especially storage solution. To caused by
failing plan should account number of
failures, occur in your environment. You for how your
solution should handle each fault without service interruption.
• Design a complete, fault-tolerant storage solution. For example, if you want your
storage solution to be able to tolerate a single fault at any level, you need this
minimum
31 of 93 3/12/2019, 12:38 PM
about:blank
• Deploy a highly available storage pool. Using mirrored or parity virtual disks in
Storage Spaces provides some fault tolerance and high availability to storage
resources. However, because all physical disks connect to a single system, that
system single point of failure. physical
disks access to the storage exist.
Storage Server 2016 supports storage
pool spaces, parity spaces, cluster
Storage environment must meet requirements:
o All storage spaces in the storage pool must use fixed provisioning.
Note: The SAS JBOD must be physically connected to all cluster nodes
that will use the storage pool. Direct attached storage that is not
connected to all cluster nodes is not supported for clustered storage
pools with Storage Spaces.
32 of 93 3/12/2019, 12:38 PM
about:blank
o Use only certified SAS-connected JBODs. These enclosure models have been
tested with Storage Spaces and enable you to identify the enclosure and slot
for a physical disk easily.
o Install the latest firmware and driver versions on all disks. Install the firmware
version that is listed as approved for the device in the Windows Server Catalog
or is recommended by your hardware vendor. Within a JBOD, it's important that
all disks of the same model have the same firmware version.
• Always replace the physical disk before you remove the drive from the storage
pool. Changing the storage pool configuration before you replace the physical disk
33 of 93 3/12/2019, 12:38 PM
about:blank
in the enclosure can cause an I/O failure or initiate virtual disk repair, which can
result in a “STOP 0x50” error and potential data loss.
• As a general rule, keep unallocated disk space in the pool for virtual disk repairs
instead of using hot spares. In Windows Server 2016, you have the option to use
available capacity on existing disks in the pool for disk repair operations instead of
bringing This enables Storage automatically repair
storage disks by copying data pool,
significantly time it takes to recover when
compared spares, and it lets you disks
instead aside hot spares.
o To correct a failed disk in a virtual disk or storage pool, you must remove the
disk that is causing the problem. Actions such as defragmenting, scan disk, or
using chkdsk cannot repair a storage pool.
o To replace a failed disk, you must add a new disk to the pool. The new disk
automatically when disk maintenance daily
Alternatively, you can trigger manually.
34 of 93 3/12/2019, 12:38 PM
about:blank
spaces return to health after the initial disk repairs, you should replace the failed
disk as soon as possible to avoid the risk of additional disk failures, which might
compromise storage health and availability and risk data loss. If you want to be
able to delay disk repairs safely until your next scheduled maintenance, configure
your storage spaces to tolerate two disk failures.
35 of 93 3/12/2019, 12:38 PM
about:blank
One of the main benefits of using Storage Spaces is the ability to expand your
storage additional storage. Occasionally,
investigate storage is being used pool
before the storage. This is your various
virtual across the physical disks
configuration based on the storage layout selected when
creating the pool. Depending upon the specifics, you might not be able to extend the
storage, even if there is available space in the pool.
Example
Consider example:
36 of 93 3/12/2019, 12:38 PM
about:blank
• However, if you attempt to extend vdisk2, you can do so because that disk is
currently three devices and across
those extend it.
37 of 93 3/12/2019, 12:38 PM
about:blank
• Vdisk2 might just be a virtual disk that used two-way mirroring. This means that
data on disk1 is duplicated on disk2 and disk3. If you wish to expand a virtual disk
with two-way mirroring, it has to have the appropriate number of columns available
to accommodate the needs of the virtual disk.
Determining
Expanding
• Server Manager. Open Server Manager, select File and Storage Services, and
then click Storage Pools. You can add a physical disk by right-clicking the pool,
and then click Add Physical Disk.
• Windows PowerShell. You can use the Windows PowerShell cmdlet Add-
PhysicalDisk physical disk to the storage
38 of 93 3/12/2019, 12:38 PM
about:blank
Demonstration
View the Properties of a Storage Pool
3. commands:
Get-StoragePool
StoragePool1 |
39 of 93 3/12/2019, 12:38 PM
about:blank
Get-VirtualDisk | fl
d. To return a list of physical disks than can be pooled, run the following
command:
| Where {$_.canpool
–StoragePoolFriendlyName StoragePool1
Data -Size 2GB
b. To add a list of physical disks that can be pooled to the variable, run the
following command:
Get-PhysicalDisk –CanPool
40 of 93 3/12/2019, 12:38 PM
about:blank
-StoragePoolFriendlyName StoragePool1
With any storage technology, it is important that you monitor storage behavior and
function to ensure ongoing reliability, availability, and optimal performance.
Using
41 of 93 3/12/2019, 12:38 PM
about:blank
100 Physical drive %1 failed to read the A physical drive can fail to
configuration or returned corrupt data read the configuration or
for storage pool %2. As a result, the in- return corrupt data for a
memory configuration might not be the storage pool for the following
most recent copy of the configuration. reasons:
Return Code: %3.
physical drive
requests
I/O
insufficient
physical
online and
updated with their
latest configurations.
42 of 93 3/12/2019, 12:38 PM
about:blank
200 Windows was unable to read the drive Windows was unable to read
header for physical drive for a
know the drive is still usable,
resetting the drive health
command line or GUI might
failure condition and enable
reassign the drive to its storage
Return Code: %2.
43 of 93 3/12/2019, 12:38 PM
about:blank
300 Physical drive %1 failed to read the A physical drive can fail to
configuration or returned corrupt data read the configuration or
for storage space %2. As a result, the return corrupt data for the
in-memory configuration might not be following reasons:
the most recent copy of the
The physical drive
configuration. Return Code:
requests
I/O
physical drive
might contain
corrupted storage
space configuration
data.
drives might
corrupted
pool
configuration data.
44 of 93 3/12/2019, 12:38 PM
about:blank
might contain
insufficient memory
resources.
• Insufficient number of
physical drives have
storage space
physical drive
contain
insufficient memory
resources.
303 Drives hosting data for storage space This event can occur if a
have failed or are missing. As a result, drive in the storage pool fails
no copy of data is available. Return or is removed.
Code: %2.
45 of 93 3/12/2019, 12:38 PM
about:blank
Performance Monitoring
There are multiple components that handle storage requests within your storage
architecture, including:
• File
• File
• Volume
46 of 93 3/12/2019, 12:38 PM
about:blank
You can use Windows PowerShell and Performance Monitor to monitor the
performance of your storage pools. If you want to use Windows PowerShell, you
must install the Storage Spaces Performance Analysis module for Windows
PowerShell.
Measure-StorageSpacesPhysicalDiskPerformance
-StorageSpaceFriendlyName StorageSpace1 60
-SecondsBetweenSamples -ReplaceExistingResultsFile
-ResultsFilePath StorageSpace1.blg -SpacetoPDMappingPath PDMap.csv
This cmdlet:
• Monitors the performance of all physical disks associated with the storage space
named StorageSpace1.
47 of 93 3/12/2019, 12:38 PM
about:blank
You can use Performance Monitor to view the data collected in the two files specified
in the cmdlet above, named StorageSpace1.blg and PDMap.csv.
Scenario
Adatum purchased a number of and you
have a storage solution devices
to the requirements in A. Datum redundancy,
you must have a redundancy solution that does not
require fast disk read and write access. You also must create a solution for data that
does require fast read and write access.
You decide to use Storage Spaces and storage tiering to meet the requirements.
Objectives
After will be able to:
• Create space.
Lab setup
Estimated
User Adatum\Administrator
Password: Pa55w.rd
48 of 93 3/12/2019, 12:38 PM
about:blank
For this lab, you need to use the available virtual machine environment. Before you
begin the lab, you must complete the following steps:
4. credentials:
Administrator
• Password: Pa55w.rd
• Domain: Adatum
5. for 20740C-LON-SVR1
Scenario
Your server does not have a hardware-based RAID card, but you have been asked to
configure redundant storage. To support this feature, you must create a storage pool.
49 of 93 3/12/2019, 12:38 PM
about:blank
1. Create a storage pool from six disks that are attached to the server
2. Create a three-way mirrored virtual disk (need at least five physical disks)
5. available
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps
Detailed Steps
Detailed Steps
Result: After completing this exercise, you should have successfully created a
storage pool and added five disks to it. Additionally, you should have created a
three-way mirrored, thinly-provisioned virtual disk from the storage pool. You also
should have copied a file to the new volume and then verified that it is accessible.
Next, after removing a physical drive, you should have verified that the virtual disk
was you could access it. added
another storage pool.
Scenario
50 of 93 3/12/2019, 12:38 PM
about:blank
Management wants you to implement storage tiers to take advantage of the high-
performance attributes of a number of SSDs, while utilizing less expensive hard disk
drives for less frequently accessed data.
2.
3.
4. Specify the media type for the sample disks and verify that the media type is
changed
7.
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps
Detailed Steps
Detailed Steps
Detailed Steps ▼
Result: After completing this exercise, you should have successfully enabled and
51 of 93 3/12/2019, 12:38 PM
about:blank
Review Question(s)
Discovery
At a must you add to a three-way
mirrored
Show solution
Discovery
You have four SAS disks, and attached to a
Windows You want to provide a users that
they What would you use?
Show solution
Data Deduplication is a role service of Windows Server 2016. This service identifies
and removes duplications within data without compromising data integrity. It does this
to achieve storing more data disk space.
This implement Data Deduplication Server 2016
storage.
Lesson objectives
After completing this lesson, you will be able to:
52 of 93 3/12/2019, 12:38 PM
about:blank
53 of 93 3/12/2019, 12:38 PM
about:blank
54 of 93 3/12/2019, 12:38 PM
about:blank
schedule one to run in the future. Installing the Data Deduplication feature and
enabling deduplication on selected volumes can also be accomplished by using
the Unattend.xml file that calls a Windows PowerShell script and can be used with
Sysprep to deploy deduplication when a system first boots.
After deduplication, files are no longer stored as independent streams of data, and
they are replaced with stubs that point to data blocks that are stored within a common
chunk store. Because these files share blocks, those blocks are only stored once,
which reduces the disk space needed to store all files. During file access, the correct
blocks transparently assembled to serve the data without the application or the
user the on-disk transformation enables you
to apply without having to worry behavior
to the users who are accessing
Deduplication storage scenarios with that are not
modified
Windows Server 2016 includes several important improvements to the way Data
Deduplication Windows Server 2012 R2 2012,
including:
55 of 93 3/12/2019, 12:38 PM
about:blank
and able to utilize multiple CPUs per volume to increase optimization throughput
rates on volume sizes up to 64 TB. This is a limitation of VSS, on which Data
Deduplication is dependent.
• Support for file sizes up to 1 TB. In Windows Server 2012 R2, very large files are
not good candidates for Data Deduplication. However, with the use of the new
stream other improvements optimization
throughput performance, deduplication 2016
performs TB.
• Support for cluster rolling upgrades. Windows servers in a failover cluster running
deduplication can include a mix of nodes running Windows Server 2012 R2 and
nodes running Windows Server 2016. This major enhancement provides full data
access to all of your deduplicated volumes during a cluster rolling upgrade. For
example, upgrade each deduplication existing
Windows cluster to Windows incurring
downtime nodes at once.
56 of 93 3/12/2019, 12:38 PM
about:blank
Server 2012 R2 deduplication nodes and are blocked from running on the
Windows Server 2016 deduplication nodes until the cluster rolling upgrade is
complete.
• Volumes must not be a system or boot volume. Because most files used by an
operating system are constantly open, Data Deduplication on system volumes
would negatively affect the performance because deduplicated data would need to
be you could use the files.
• Volumes must be attached to the Windows Server and cannot appear as non-
removable drives. This means that you cannot use USB or floppy drives for Data
Deduplication, nor use remotely-mapped drives.
• Volumes can be on shared storage, such as Fibre Channel, iSCSI SAN, or SAS
array.
57 of 93 3/12/2019, 12:38 PM
about:blank
58 of 93 3/12/2019, 12:38 PM
about:blank
59 of 93 3/12/2019, 12:38 PM
about:blank
Note: You should be cautious when using the unoptimization job because all
the deduplicated data will return to the original logical file size. As such, you
should verify the volume has enough free space for this activity
data to allow the job successfully.
Data
60 of 93 3/12/2019, 12:38 PM
about:blank
• Replaces all duplicate chunks with a reference, or stub, to a single copy of the
chunk in the chunk store.
• Replaces the original files with a reparse point, which contains references to its
data chunks.
The Data Deduplication process works through scheduled tasks on the local server,
but you can run the process interactively by using Windows PowerShell. More
information about this is discussed later in the module.
Note: The three main types of data deduplication target (or post-
and in-line (or transit
61 of 93 3/12/2019, 12:38 PM
about:blank
After you enable a volume for deduplication and the data is optimized, the volume
contains the following elements:
• Unoptimized files. Includes files that do not meet the selected file-age policy
setting, alternate data streams, with
extended smaller than 32 KB, or
• Additional free space. The optimized files and chunk store occupy much less
space than they did prior to optimization.
Deploying Deduplication
62 of 93 3/12/2019, 12:38 PM
about:blank
o VHD libraries. Virtual hard disk (VHD) file storage for provisioning to
63 of 93 3/12/2019, 12:38 PM
about:blank
hypervisors.
For example, files that frequently change and are often accessed by users or
applications are not good candidates for deduplication. In these scenarios,
deduplication might not be able to process the files, as the constant access and
change to the data are likely to cancel any optimization gains made by
other hand, good candidates deduplication
64 of 93 3/12/2019, 12:38 PM
about:blank
• Evaluate savings with the Deduplication Evaluation Tool. You can use the
Deduplication Evaluation Tool, DDPEval.exe, to determine the expected savings
that enable deduplication
DDPEval. evaluating local drives remote
shares.
o You have directories that you do not want to deduplicate. Add a directory to the
65 of 93 3/12/2019, 12:38 PM
about:blank
After completing your planning, you need to use the following steps to deploy Data
Deduplication your environment:
ServerManager
Add-WindowsFeature -Name FS-Data-Deduplication
Import-Module Deduplication
66 of 93 3/12/2019, 12:38 PM
about:blank
iii. Enter the minimum number of days that should elapse from the date of file
creation before files are deduplicated, enter the extensions of any file types
that should not be deduplicated, and then click Add to browse to any folders
with files that should not be deduplicated.
• Configure Data Deduplication jobs. With Data Deduplication jobs, you can run
them manually, on demand, or use a schedule. The following list are the types of
67 of 93 3/12/2019, 12:38 PM
about:blank
–Volume VolumeLetter
–Volume VolumeLetter
68 of 93 3/12/2019, 12:38 PM
about:blank
• Enable Deduplication.
69 of 93 3/12/2019, 12:38 PM
about:blank
Demonstration steps
Install the Data Deduplication Role Service
Enable
2. Storage Services.
3. Click Disks.
5. Enable Data Deduplication, and then click the General purpose file server
6. settings:
c. Exclude: D:\shares
Check Deduplication
1. PowerShell.
a.Get-DedupStatus
70 of 93 3/12/2019, 12:38 PM
about:blank
b.Get-DedupStatus | fl
c.Get-DedupVolume
d.Get-DedupVolume |fl
e.Start-DedupJob D: -Type Optimization –Memory 50
3. and 2c.
The following table highlights typical deduplication savings for various content types.
Your data storage savings will vary by data type, the mix of data, and the size of the
71 of 93 3/12/2019, 12:38 PM
about:blank
volume and the files that the volume contains. You should consider using the
Deduplication Evaluation Tool to evaluate the volumes before you enable
deduplication.
• User documents. This includes group content publication or sharing, user home
folders (or MyDocs), and profile redirection for accessing offline files. Applying
Data shares might save percent of your
system’s
• Virtualization libraries. This includes virtual hard disk files (i.e., .vhd and .vhdx
files) storage for provisioning to hypervisors. Applying Data Deduplication to these
libraries might be able to save you up to 80 to 95 percent of your system’s storage
space.
Based on observed savings and typical resource usage in Windows Server 2016,
deployment deduplication are ranked
• Ideal deduplication
72 of 93 3/12/2019, 12:38 PM
about:blank
o VDIs)
o Line-of-business servers
o Web servers
o computing (HPC)
• Not deduplication
o WSUS
73 of 93 3/12/2019, 12:38 PM
about:blank
In Windows Server 2016, you should consider the following related technologies and
potential deploying Data Deduplication:
• Failover Clusters. Windows Server 2016 fully supports failover clusters, which
means volumes will failover gracefully the cluster.
Effectively, volume is a self-contained e., all of
the information that the requires that
each accesses deduplicated running the
Data feature. When a cluster is Deduplication schedule
information is configured in the cluster. As a result, if a deduplicated volume is
taken over by another node, the scheduled jobs will be applied on the next
74 of 93 3/12/2019, 12:38 PM
about:blank
• FSRM quotas. Although you should not create a hard quota on a volume root
folder enabled for deduplication, using File Server Resource Manager (FSRM),
you can create a soft quota on a volume root which is enabled for deduplication.
When FSRM encounters a deduplicated file, it will identify the file’s logical size for
quota Consequently, quota usage ( thresholds)
does deduplication processes quota
functionality, volume-root soft quotas will work
as deduplication.
75 of 93 3/12/2019, 12:38 PM
about:blank
76 of 93 3/12/2019, 12:38 PM
about:blank
primarily on how much and how frequently the data on the volume changes and the
data access throughput rates of the disk storage subsystem.
• Deduplication optimization must be able to keep up with the daily data churn.
• The total amount of churn scales with the size of the volume.
77 of 93 3/12/2019, 12:38 PM
about:blank
types, for your estimations. However, the most accurate method of assessing the
appropriate volume size is to perform the measurements directly on your
deduplication system based on the representative samples of your data, such as data
churn and deduplication processing speed.
You should consider using the following options to monitor deduplication in your
environment its health:
o Get-DedupStatus. The most commonly used cmdlet, this cmdlet returns the
deduplication status for volumes which have data deduplication metadata,
which includes the deduplication rate, the number/sizes of optimized files, the
last run-time of the deduplication jobs, and the amount of space saved on the
volume.
container.
store.
container.
78 of 93 3/12/2019, 12:38 PM
about:blank
▪ Hotspots in a container.
• Event Viewer logs. Monitoring the event log can also be helpful to understand
deduplication events and status. To view deduplication events, in Event Viewer,
navigate to Applications and Services Logs click Microsoft click Windows,
and Deduplication. For example, Event you with
the deduplication job and the
79 of 93 3/12/2019, 12:38 PM
about:blank
Read Bytes/sec, Disk Write Bytes/sec, and Average Disk sec/Transfer. Depending
on other activities on the server, you might be able to use the data results from
these counters to get a rough estimate of the saving ratio by examining how much
data is being read and how much is being written per interval. You can also use
the Resource Monitor to identify the resource usage of specific programs/services.
To view disk activity, in Windows Resource Monitor, filter the list of processes to
locate examine the I/O on tab.
• File Explorer. While not the ideal choice for validating deduplication on an entire
volume, you can use File Explorer to spot check deduplication on individual files.
In viewing the properties of file, you notice that Size displays the logical size of
the displays the true physical For an
optimized is less than the actual because
deduplication contents of the file to and
replaces an NTFS reparse
With the data that is collected by monitoring, you can use the following Windows
PowerShell optimal efficiency of environment.
80 of 93 3/12/2019, 12:38 PM
about:blank
• Use
81 of 93 3/12/2019, 12:38 PM
about:blank
o Priority, which sets the CPU and I/O priority for specific deduplication jobs.
82 of 93 3/12/2019, 12:38 PM
about:blank
Some of the most common causes for deduplication to report corruption are:
Note: For more information, refer to “FSRM and Data Deduplication may
be adversely affected when you use Robocopy /MIR in Windows Server
2012” at: http://aka.ms/W0ux7m
83 of 93 3/12/2019, 12:38 PM
about:blank
• General corruption. You can use the steps below to troubleshoot most general
causes for deduplication to report corruption:
84 of 93 3/12/2019, 12:38 PM
about:blank
You can use the following command in Windows PowerShell to initiate a deep
Scrubbing job:
85 of 93 3/12/2019, 12:38 PM
about:blank
• Individual
• Full backup/restore
On the other hand, the following backup and restore scenarios are not supported with
deduplication Server 2016:
86 of 93 3/12/2019, 12:38 PM
about:blank
Note: New chunks are appended to the current chunk store container. When
its size reaches approximately 1 GB, that container file is sealed and a new
container file is created.
Restore Operations
1. The complete set of data deduplication metadata and container files are
restored.
3. are restored.
As with any product from a third-party vendor, you should verify whether the backup
solution supports Data Deduplication in Windows Server 2016, as unsupported
backup introduce corruptions after common
methods support Data Deduplication 2016:
• Some backup vendors support optimized backup for a full volume backup, which
backs up the deduplicated files as-is; i.e., as a reparse point stub with the chunk
87 of 93 3/12/2019, 12:38 PM
about:blank
store.
The backup vendor should be able to comment on what their product supports, the
method version.
Discovery
Can you enable Data Deduplication on a drive with storage tiering enabled?
Show solution
Check
Discovery
Can you enable Data Deduplication on ReFS formatted drives?
Check
Discovery
Can you Deduplication on volumes in running
and apply virtual machines?
88 of 93 3/12/2019, 12:38 PM
about:blank
Scenario
After you have tested the storage redundancy and performance options, you decide
that it also would be beneficial to maximize the available disk space that you have,
especially servers. You decide to test solutions to
maximize for users.
Objectives
Lab
Estimated
89 of 93 3/12/2019, 12:38 PM
about:blank
Password: Pa55w.rd
For this lab, you must use the available virtual machine environment. These should
already be running from Lab A. If they are not, before you begin the lab, you must
complete the following steps and then complete Lab A:
Password: Pa55w.rd
5. for 20740C-LON-SVR1
Scenario
90 of 93 3/12/2019, 12:38 PM
about:blank
Detailed Steps ▼
Detailed Steps
Detailed Steps
Scenario
3. optimized
4. again
5. module
91 of 93 3/12/2019, 12:38 PM
about:blank
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps ▼
Detailed Steps
Review Question(s)
Discovery
Your the impact that using have on the
write servers’ volumes. Is
Show solution
Common Troubleshooting
Review Question(s)
92 of 93 3/12/2019, 12:38 PM
about:blank
Discovery
You attach five 2-TB disks to your Windows Server 2012 computer. You want to simplify
the process of managing the disks. In addition, you want to ensure that if one disk fails,
the failed What feature can you accomplish these
goals?
Show solution
Check Knowledge
Discovery
Your manager has asked you to consider the use of Data Deduplication within your
storage architecture. In what scenarios is the Data Deduplication role service
particularly useful?
Show solution
93 of 93 3/12/2019, 12:38 PM