Dell Compellent Best Practices With VMware VSphere 5.x
Dell Compellent Best Practices With VMware VSphere 5.x
Dell Compellent Best Practices With VMware VSphere 5.x
Revisions
Date
Description
September 2011
Initial Release
October 2012
January 2014
May 2014
Added SCOS 6.5 updates, including compression, multi-VLAN, Sync Live Volume
September 2014
THIS WHITE PAPER IS FOR INFORMATIONAL PURPOSES ONLY, AND MAY CONTAIN TYPOGRAPHICAL ERRORS AND
TECHNICAL INACCURACIES. THE CONTENT IS PROVIDED AS IS, WITHOUT EXPRESS OR IMPLIED WARRANTIES OF
ANY KIND.
2014 Dell Inc. All rights reserved. Reproduction of this material in any manner whatsoever without the express
written permission of Dell Inc. is strictly forbidden. For more information, contact Dell.
PRODUCT WARRANTIES APPLICABLE TO THE DELL PRODUCTS DESCRIBED IN THIS DOCUMENT MAY BE FOUND
AT: http://www.dell.com/learn/us/en/19/terms-of-sale-commercial-and-public-sector Performance of network
reference architectures discussed in this document may vary with differing deployment conditions, network loads, and
the like. Third party products may be included in reference architectures for the convenience of the reader. Inclusion
of such third party products does not necessarily constitute Dells recommendation of those products. Please consult
your Dell representative for additional information.
Trademarks used in this text:
Dell, the Dell logo, Dell Boomi, Dell Precision ,OptiPlex, Latitude, PowerEdge, PowerVault,
PowerConnect, OpenManage, EqualLogic, Compellent, KACE, FlexAddress, Force10 and Vostro are
trademarks of Dell Inc. Other Dell trademarks may be used in this document. Cisco Nexus, Cisco MDS, Cisco NX0S, and other Cisco Catalyst are registered trademarks of Cisco System Inc. EMC VNX, and EMC Unisphere are
registered trademarks of EMC Corporation. Intel , Pentium, Xeon, Core and Celeron are registered trademarks of
Intel Corporation in the U.S. and other countries. AMD is a registered trademark and AMD Opteron, AMD
Phenom and AMD Sempron are trademarks of Advanced Micro Devices, Inc. Microsoft , Windows, Windows
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Server, Internet Explorer, MS-DOS, Windows Vista and Active Directory are either trademarks or registered
trademarks of Microsoft Corporation in the United States and/or other countries. Red Hat and Red Hat Enterprise
Linux are registered trademarks of Red Hat, Inc. in the United States and/or other countries. Novell and SUSE are
registered trademarks of Novell Inc. in the United States and other countries. Oracle is a registered trademark of
Oracle Corporation and/or its affiliates. Citrix, Xen, XenServer and XenMotion are either registered trademarks or
trademarks of Citrix Systems, Inc. in the United States and/or other countries. VMware , Virtual SMP, vMotion,
vCenter and vSphere are registered trademarks or trademarks of VMware, Inc. in the United States or other
countries. IBM is a registered trademark of International Business Machines Corporation. Broadcom and
NetXtreme are registered trademarks of Broadcom Corporation. Qlogic is a registered trademark of QLogic
Corporation. Other trademarks and trade names may be used in this document to refer to either the entities claiming
the marks and/or names or their products and are the property of their respective owners. Dell disclaims proprietary
interest in the marks and names of others.
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Table of contents
Revisions ............................................................................................................................................................................................. 2
Executive summary .......................................................................................................................................................................... 8
1
1.2
1.3
1.4
2.2
2.3
2.4
3.2
3.3
Modifying the VMFS queue depth for virtual machines (DSNRO) ........................................................................ 15
3.4
3.5
6.2
6.3
6.4
6.5
Configuring the VMware iSCSI software initiator for a single path ...................................................................... 25
6.6
6.7
6.8
6.9
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
8.2
8.3
8.4
8.4.1 VMFS-3............................................................................................................................................................................ 41
8.4.2 VMFS-5 ............................................................................................................................................................................ 41
9
Compression .................................................................................................................................................................. 50
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Executive summary
This document will provide configuration examples, tips, recommended settings, and other storage
guidelines a user can follow while integrating VMware ESXi 5.x Server hosts with the Storage Center. This
document has been written to answer many frequently asked questions with regard to how VMware
interacts with the Storage Center's various features such as Dynamic Capacity (thin provisioning), Data
Progression (automated tiering), and Remote Instant Replay (replication).
Prerequisites
This document assumes the reader has had formal training or has advanced working knowledge of the
following:
Installation and configuration of VMware vSphere 5.x
Configuration and operation of the Dell Compellent Storage Center
Operating systems such as Windows or Linux
Dell Compellent advises customers to read the vSphere Storage Guide, which is publicly available on the
vSphere documentation pages to provide additional important information about configuring ESXi hosts to
use the SAN.
Intended audience
This document is highly technical and intended for storage and server administrators, as well as other
information technology professionals interested in learning more about how VMware vSphere 5.x
integrates with Storage Center.
Note: The information contained within this document is intended only to be general recommendations
and may not be applicable to all configurations. There are certain circumstances and environments
where the configuration may vary based upon individual or business needs.
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
1.1
1.2
Port zoning
If the Storage Center front-end ports are plugged into switch ports 0, 1, 2, & 3, and the first ESXi HBA port
is plugged into switch port 10, the resulting zone should contain switch ports 0, 1, 2, 3, & 10.
Repeat this for each of the HBAs in the ESXi host. If the environment has multiple fabrics, the additional
HBA ports in the host should have separate unique zones created in their respective fabrics.
Caution: Due to the supportability of port zoning, WWN zoning is preferred over port zoning.
1.3
WWN zoning
When zoning by WWN, the zone only needs to contain the host HBA port and the Storage Center frontend primary ports. In most cases, it is not necessary to include the Storage Center front-end reserve
ports because they are not used for volume mappings. For example, if the host has two HBAs connected
to two disjointed fabrics, the fibre channel zones would look similar to this:
Name: ESX1-HBA1
WWN: 2100001B32017114
WWN: 5000D31000036001
WWN: 5000D31000036009
Name: ESX1-HBA2
WWN: 210000E08B930AA6
WWN: 5000D31000036002
WWN: 5000D3100003600A
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
1.4
Virtual ports
If the Storage Center is configured to use Virtual Port Mode, all of the Front End virtual ports within each
Fault Domain should be included in the zone with each ESXi initiator.
Figure 1
10
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
2.1
2.2
The connection options field should be set to 1 for point to point only
The login retry count field should be set to 60 attempts
The port down retry count field should be set to 60 attempts
The link down timeout field should be set to 30 seconds
The queue depth (or Execution Throttle) field should be set to 255.
- This queue depth can be set to 255 because the ESXi VMkernel driver module and DSNRO can
more conveniently control the queue depth
2.3
2.4
11
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
(Default=dependent on application)
(Default=32)
(Default=32)
(Default=64)
(Default=Varies)
The following sections explain how the queue depth is set in each of the layers in the event it needs to be
changed.
Caution: The appropriate queue depth for a host may vary due to a number of factors, so it is
recommended to only increase or decrease the queue depth if necessary. See Appendix A for more info
on determining the proper queue depth.
3.1
3.2
12
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
In addition to setting the queue depth in the driver module, the disk timeouts must also be set within the
same command. These timeouts need to be set in order for the ESXi host to properly survive a Storage
Center controller failover.
Please refer to the latest VMware documentation for instructions on how to configure these settings:
VMware document: vSphere Troubleshooting
- Section Title: Adjust Queue Depth for QLogic and Emulex HBAs
Caution: Before executing these commands, please refer to the latest documentation from VMware for
any last minute additions or changes.
For each of these adapters, the method to set the driver queue depth and timeouts uses the following
general steps:
1.
Find the appropriate driver name for the module that is loaded:
a. For QLogic:
13
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
(i.e. -m=qla2xxx)
Note: In certain earlier versions of ESXi 5.x, the option to set the Login Timeout parameter is not
available. In order to enable the login timeout parameter, it may require applying the patch as described
in VMware KB article 2007680.
14
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
3.3
Figure 2
In ESXi 5.5 and later, the DSNRO setting has been changed from a global value to a per LUN value set via
the command line:
esxcli storage core device set -d <naa.dev> -O <value of 1-256>
This allows fine tuning of DSNRO on a per volume basis, however the drawback is that it must be set on
each datastore (and each host) if the desired queue depth is greater than 32.
15
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Figure 3
Note: The Disk.SchedNumReqOutstanding limit does not apply to LUNs mapped as Raw Device
Mappings (RDMs). Each RDM will have its own queue.
More information on the Disk.SchedNumReqOutstanding variable can be found in the following
documents:
VMware document: vSphere Troubleshooting
- Section: Change Maximum Outstanding Disk Requests in the vSphere Web Client
VMware KB Article: Setting the Maximum Outstanding Disk Requests for virtual machines
- Link: http://kb.vmware.com/kb/1268
3.4
16
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
3.5
Figure 4
17
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Figure 5
1.
Registry setting for the LSI Logic SAS vSCSI Adapter (Server 2008/R2)
Search for and download the following driver from the LSI Logic download page:
Filename: LSI_U320_W2003_IT_MID1011438.zip
Adapter: LSI20320-R
Driver: Windows Server 2003 (32-bit)
Version: WHQL 1.20.18 (Dated: 13-JUN-05)
2. Update the current LSI Logic PCI-X Ultra320 SCSI HBA driver to the newer WHQL driver version
1.20.18.
3. Using regedit, add the following keys: (Backup the registry first)
Windows Registry Editor Version 5.00
[HKLM\SYSTEM\CurrentControlSet\Services\symmpi\Parameters\Device]
"DriverParameter"="MaximumTargetQueueDepth=128;"
; The semicolon is required at the end of the queue depth value
"MaximumTargetQueueDepth"=dword:00000080
; 80 hex is equal to 128 in decimal
18
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Figure 6
Registry setting for the LSI Logic Parallel vSCSI Adapter (Server 2003)
19
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Using the registry editor, modify the following key: (Backup the registry first)
Figure 7
Note: This registry value is automatically set when installing VMware Tools version 3.0.2 and later. For
more information, please see the VMware KB Article 1014
Linux
For more information about setting disk timeouts in Linux, please refer to the following VMware
Knowledge Base articles:
Increasing the disk timeout values for a Linux virtual machine
- http://kb.vmware.com/kb/1009465
20
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Figure 8
BusLogic Parallel
This vSCSI controller is used for older operating systems. Due to this controllers queue depth limitations,
its use is not recommended unless it is the only option available for that particular operating system. This
is because when using certain versions of Windows, the OS issues only enough I/O to fill a queue depth of
one.
LSI Logic Parallel
Since this vSCSI adapter is supported by many operating system versions, and is a good overall choice, it is
recommended for virtual machines that do not support the LSI Logic SAS adapter.
LSI Logic SAS
This vSCSI controller is available for virtual machines with hardware version 7 & 8, and has similar
performance characteristics of the LSI Logic Parallel. This adapter is required for MSCS Clustering in
Windows Server 2008 because SCSI-3 reservations are needed. Some operating system vendors are
gradually withdrawing support for SCSI in favor of SAS, subsequently making the LSI Logic SAS controller a
good choice for future compatibility.
VMware Paravirtual
This vSCSI controller is a high-performance adapter that can result in greater throughput and lower CPU
utilization. Due to feature limitations when using this adapter, we recommend against using it unless the
virtual machine has very specific performance needs. More information about the limitations of this
adapter can be found in the vSphere Virtual Machine Administration Guide, in a section titled, About
VMware Paravirtual SCSI Controllers.
21
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.1
6.2
Figure 9
As an added benefit, when a new ESXi host is placed into the server cluster, all of the existing volume
mappings assigned to the cluster object will be applied to the new host. This means that if the cluster has
100 volumes mapped to it, presenting all of them to a newly created ESXi host is as simple as adding it to
the cluster object.
Similarly, if the host is removed from the server cluster, the cluster mappings will also be removed, so it is
important that those volumes are not being used by the host when they are removed. Only volumes that
are mapped to an individual host, such as the boot volume, will remain once a host is removed from the
server cluster.
Also in the mapping wizard, the system can auto select the LUN number, or a preferred LUN number can
be manually specified from the advanced settings screen as shown in the figure below.
22
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.3
Note: If the LUN number is not kept consistent between multiple hosts or multiple HBA's, VMFS
datastores may not be visible to all nodes, preventing use of vMotion, HA, DRS, or FT.
Keep in mind that when a volume uses multiple paths, the first ESXi initiator in each server will need to be
mapped to one front end port, while the second ESXi initiator will be mapped to the other front end port
in that same controller. For example:
"LUN10-vm-storage" Controller1/PrimaryPort1 FC-Switch-1 Mapped to ESX1/HBA1 as LUN 10
"LUN10-vm-storage" Controller1/PrimaryPort2 FC-Switch-2 Mapped to ESX1/HBA2 as LUN 10
23
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Likewise, if different volume is active on the second controller, it may be mapped such as:
"LUN20-vm-storage" Controller2/PrimaryPort1 FC-Switch-1 Mapped to ESX1/HBA1 as LUN 20
"LUN20-vm-storage" Controller2/PrimaryPort2 FC-Switch-2 Mapped to ESX1/HBA2 as LUN 20
This means that when configuring multipathing in ESXi, a single volume cannot be mapped to both
controllers at the same time, because a volume can only be active on one controller at a time.
6.4
24
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Table 1
6.5
Function
Description
Select LUN
This option is to manually specify the LUN. If this box is not checked,
the system will automatically assign the next available LUN.
Map to Controller
Configure Multipathing
This option designates how many of the Storage Center FE ports that
the system will allow the volume to be mapped through. For example
if each controller has 4 Front End ports, selecting unlimited will map
the volume through all 4, whereas selecting 2 will only use 2 of the 4
front end ports. The system will automatically select the 2 front end
ports based on which already have the fewest mappings.
Enable the "Software iSCSI Client" within the ESXi firewall (in the "Security Profile" of the ESXi host)
Add a "VMkernel port" to a virtual switch assigned to the physical NIC for iSCSI (See figures below)
From the Storage Adapters configuration screen, click Add
Select Add Software iSCSI Adapter, then click OK
From within the Storage Adapters, highlight the iSCSI Software Adapter (i.e. vmhba33), click
Properties
6. Under the Dynamic Discovery tab, add all of the Storage Center iSCSI IP addresses that are assigned
to the iSCSI cards in the Storage Center controller(s), or just the iSCSI Control Port IP address.
7. Rescan the iSCSI Initiator.
From Within the Storage Center GUI or Enterprise Manager:
8. Create a server object for the ESXi host using the IP Address previously specified for the VMkernel in
step 2 above
9. Map a volume to the ESXi host
From within the VMware vSphere Client:
10. Navigate to the Storage Adapters section, and rescan the iSCSI HBA for new LUN's.
25
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.6
Note: Using port binding is not recommended when the VMkernel ports are on different networks, as it
may cause very long rescan times and other storage management problems. Please see the following
VMware KB article for more information: Considerations for using software iSCSI port binding in
ESX/ESXi (2038869)
26
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Note: Some older versions of Storage Center allowed adding iSCSI initiators by either IP address or by
iSCSI Names. If prompted with that choice, using iSCSI Names is recommended, as shown in the figure
below.
27
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.6.1
If the network is comprised of multiple separate subnets for iSCSI, then use a separate vSwitch with single
iSCSI virtual port and single NIC. Once again, more iSCSI virtual ports can be used if there are more front
end ports in the controller.
I.e. Subnets have different IP addressing schemes Create the vSwitches as shown below
Note: Using port binding is not recommended when the VMkernel ports are on different networks, as it
may cause very long rescan times and other storage management problems. Please see the following
VMware KB article for more information: Considerations for using software iSCSI port binding in
ESX/ESXi (2038869)
28
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.7
Figure 18 VLANs assigned to vSwitches isolating in-guest iSCSI initiator traffic for customers
Since iSCSI traffic is not encrypted in its plain form, the best practice is to isolate that traffic for security
purposes. A common misconception is that CHAP encrypts iSCSI traffic, whereas it simply provides
authentication for the connection to prevent unauthorized access.
29
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
NOTE: The configuration of iSCSI VLANs is new functionality that can only be configured through the
Enterprise Manager Client, and is unavailable through the standard Storage Center System Manager Web
GUI.
6.8
6.9
30
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.9.1
Fixed policy
If the Fixed policy is used, it will give the greatest control over the flow of storage traffic. However, one
must be very careful to evenly distribute the load across all host HBAs, Front-End Ports, fabrics, and
Storage Center controllers.
When using the fixed policy, if a path fails, all of the LUNs using it as their preferred path will fail over to the
secondary path. When service resumes, the LUNs will resume I/O on their preferred path.
Fixed Example: (See figure below)
HBA1 loses connectivity; HBA2 takes over its connections.
HBA1 resumes connectivity; HBA2 will fail its connections back to HBA1.
6.9.2
Round robin
The round robin path selection policy uses automatic path selection and load balancing to rotate I/O
through all paths. It is important to note that round robin load balancing does not aggregate the storage
link bandwidth; it merely distributes the load for the volumes in bursts evenly and sequentially across paths
in an alternating fashion.
Using round robin will reduce the management headaches of manually balancing the storage load across
all storage paths as with a fixed policy; however there are certain situations where using round robin does
not make sense. For instance, it is generally not considered best practice to enable round robin between
an iSCSI path and fibre channel path, nor enabling it to balance the load between an 8 Gb FC and a 16 Gb
FC path. If round robin is enabled for one or more datastores/LUNs, care must be taken to ensure all the
paths included are identical in type, speed, and have the same queue depth settings.
Here is an example of what happens during a path failure using round robin.
31
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Caution: The round robin path selection policy may be unsupported for use with Microsoft Clustering
Services dependent on the version of ESXi, so it is advised to check current support status.
32
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.9.3
6.10
6.11
33
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.12
6.13
1. Make note of the volumes naa identifier. This will be referenced later.
2. From the datastore view, right click on the datastore and select Unmount.
34
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
6.14
35
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
7.1
36
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Once the ESXi host is up and is running correctly, the second path can then be added to the boot volume
by modifying the mapping. To do this, right click on the mapping and select, Modify Mapping.
Uncheck: Only map using specified server ports
Maximum number of paths allowed: Unlimited
Once the 2nd path has been added, the HBAs on the ESXi host can be rescanned.
37
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
8.1
Note: Within certain versions of ESXi 5.x, hosts may have a maximum addressable storage limits of 4 TB,
8 TB, or 60 TB due to the VMFS heap size. Please read the following VMware KB link for more info:
http://kb.vmware.com/kb/1004424
8.2
38
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
in some cases the appropriate number for high I/O load virtual machines may be less than 5, while the
number of virtual machines with low I/O disk requirements may be 25 or more.
Since the appropriate number of virtual machines that can be put onto each datastore is subjective and
dependent on the environment, a good recommendation is to start with 15 virtual machines, and
increase/decrease the number of virtual machines on each datastore as needed. Moving virtual machines
between datastores can even be done non-disruptively when licensed to use VMwares Storage vMotion
feature.
The most common indicator that the datastore has too many virtual machines placed on it is if the queue
depth of the datastore is regularly exceeding set limits therefore increasing disk latency. Remember that if
the driver module is set to a 256 queue depth, the maximum queue depth of each datastore is also 256.
This means that if there are 16 virtual machines on a datastore all heavily driving a 32 queue depth (16 * 32
= 512), they are essentially overdriving the disk queues by double, and the resulting high latency will most
likely degrade performance. (See Appendix A for more information on determining if the queue depth of a
datastore is being correctly utilized.)
A second less common indicator that a datastore has too many virtual machines placed on it would be the
frequent occurrence of SCSI Reservation Conflicts as seen when monitoring with esxtop. New to ESXi
5.x is a field in the Disk Device screen for Reserve Stats (RESVSTATS). That said, when monitoring it is
normal to see a few reservations (RESV/s) entries and even a few conflicts (CONS/s) from time to time, but
when noticing conflicts (CONS/s) happening very frequently on a particular volume, it may be time to
move some of the virtual machines to a different datastore. The VAAI Hardware Assisted Locking primitive
will help to alleviate the performance degradation caused by these SCSI-2 reservations, so if datastores are
impacted, is recommended to upgrade to a version of Storage Center firmware that supports this VAAI
primitive. (See the section below on VAAI for more information.)
Note: There are many resources available that discuss VMware infrastructure design and sizing, so this
should only be used as a general rule of thumb, and may vary based upon the needs of the environment.
8.3
39
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
whether or not to align VMFS partitions, it is recommended that testing is performed to determine the
impact that an aligned partition may have on particular applications because all workloads are different.
To manually align the VMFS block boundaries to the Storage Center page boundaries for performance
testing, the recommended offset when creating a new datastore is 8192 (or 4 MB).
Note: Using the Compellent vSphere Client Plug-in to create new datastores will automatically align
them to the recommended offset. Also note, that guest partition alignment is generally not necessary
when using Windows Server 2008 and above, due to a change in partition offset defaults from previous
versions.
Figure 26 This is an example of a fully aligned partition in the Storage Center, where one guest I/O will
only access necessary physical disk sectors
Figure 27 This is an example of an unaligned partition in a traditional SAN where performance can be
improved by alignment
40
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
8.4
8.4.1
VMFS-3
With VMFS-3, choosing a block size for a datastore determines the maximum size of a VMDK file that can
be placed on it. In other words, the block size should be chose based on the largest virtual disk planned to
be put on the datastore. Choose this datastore file system version if backwards compatibility with ESX 4.x
hosts is needed.
Table 2
Block Size
1 MB
256 GB
2 MB
512 GB
4 MB
1024 GB
8 MB
The default block size is 1 MB, so if virtual disks need to be sized greater than 256 GB, this value will need
to be increased. For example, if the largest virtual disk to be placed on a datastore is 200 GB, then a 1 MB
block size should be sufficient, and similarly, if there is a virtual machine that will require a 400 GB virtual
disk, then the 2 MB block size should be sufficient.
One should also consider future growth of the virtual machine disks, and the future upgrade to VMFS-5
when choosing the block size. If a virtual machine resides on a datastore formatted with a 1 MB block size,
and in the future it needs one of its virtual disks extended beyond 256 GB, the virtual machine would have
to be relocated to a different datastore with a larger block size. Also remember that if a VMFS-3 datastore
with a 2 MB block size is upgraded to VMFS-5, the block size remains at 2 MB which can cause VAAI to be
disabled and the host to use the default DataMover instead.
Note: Since certain VAAI offload operations require that the source and destination datastores have the
same VMFS block size, it is worth considering a standard block size for all datastores. Please consult the
vStorage APIs for Array Integration FAQ for more information.
8.4.2
VMFS-5
With VMFS-5 the only available block size is 1 MB allowing for up to a 64 TB datastore, and if running ESXi
5.5, up to a 62 TB VMDK. This format also allows the VAAI Dead Space Reclamation primitive (SCSI
UNMAP) to reclaim storage after a VMDK is deleted (See section below on VAAI). Keep in mind, this file
system version is not backwards compatible with ESX 4.x hosts, so there are special considerations when
using or migrating to this format.
41
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
42
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
9.1
9.1.1
9.1.2
43
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Create one paired datastore for the corresponding virtual machine page files
- This should contain virtual machine page files. Using Windows as an example, one would
create a 2GB - 16GB virtual disk (P:) on this volume to store the Windows paging file for each
virtual machine.
- This volume can be sized considerably smaller than the main datastore as it only needs
enough space to store page files.
Often the question is asked whether or not it is a good idea to place all of the operating system page files
on a single datastore. Generally speaking, this is not a good practice for a couple of reasons.
First, the page file datastore can also experience contention from queue depth utilization or disk I/O; so
too many VMDK files during a sudden memory swapping event could decrease performance even further.
For example, if a node in the ESXi HA cluster fails, and the effected virtual machines are consolidated on
the remaining hosts. The sudden reduction in overall memory could cause a sudden increase in paging
activity that could overload the datastore causing a storage performance decrease.
Second, it becomes a matter of that datastore becoming a single point of failure. Operating systems are
usually not tolerant of disk drives being unexpectedly removed. If an administrator were to accidentally
unmap the page file volume, the number of virtual machines within the failure domain would be isolated
to a subset of the virtual machines instead of all the virtual machines.
9.1.3
9.1.4
44
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Timesaver: To help organize the LUN layout for ESXi clusters, some administrators prefer to store their
layout in a spreadsheet. Not only does this help to design their LUN layout in advance, but it also helps
keep things straight as the clusters grow larger.
Note: There are many factors that may influence architecting storage with respect to the placement of
virtual machines. The method shown above is merely a suggestion, as business needs may dictate
different alternatives.
9.2
45
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Statistical Reporting
- Storage usage and performance can be monitored for an individual virtual machine
Backup/Restore of an entire virtual machine is simplified
- If a VM needs to be restored, an administrator can just unmap/remap a replay in its place
Disadvantages
There will be a maximum of 256 virtual machines in the ESXi cluster
- The HBA has a maximum limit of 256 LUNs that can be mapped to the ESXi host, and since we
can only use each LUN number once when mapping across multiple ESXi hosts, it would
essentially have a 256 virtual machine limit (assuming that no extra LUNs would be needed for
recoveries).
Increased administrative overhead
- Managing a LUN for each virtual machine and all the corresponding mappings may get
challenging
46
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
10
47
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
11
48
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Like previously mentioned at the beginning of this section, unless there is specific business needs that
require a particular virtual machine or application to have a specific RAID type, our recommendation is to
keep the configuration simple. In most cases, the Data Progression Recommended setting can be used
to automatically classify and migrate data based upon usage.
Note: A note about Data Progression Best Practices: A replay schedule for each volume should be
created that (at a minimum) takes one daily replay that doesnt expire for 25 hours or more. This will have
a dramatic effect on Data Progression behavior, which will increase the overall system performance.
49
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
11.1
Compression
With the introduction of Storage Center 6.5, Data Progression has the ability to compress data as part of its
daily processing cycle. This block level compression uses the Dell LZPS algorithm to reduce the size of
Replay data stored on Tier 3 drives in a manner that has a low impact on performance.
It is important to understand that in the first release of this feature, Data Progression will only compress
pages contained in Tier 3 classified as inaccessible. In the example figure above, when data progression
runs, the inactive pages that are color coded grey are the only pages eligible to be compressed and
immediately migrated to Tier 3. This reduces any negative performance impacts to the datastore because
the host actually does not have direct access to any of the blocks contained in the compressed pages.
The only time a compressed page could be accessed, is as part of a View volume created from Replay for
a recovery. In which case, if the page is accessed, the data will simply be decompressed on the fly.
Note: For those customers that are using Replays and View volumes for block level cloning of virtual
machine datastores, it is important to note that there may be read performance impacts if compression is
enabled on the cloned or master datastore volumes.
For more information about the compression feature, please read the Dell Compellent Storage Center
6.5.1 and Data Compression feature brief on Dell Tech Center.
50
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
11.2
51
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
12
12.1
12.1.1
12.1.2
52
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
12.1.3
Thin provisioned
(a.k.a. Thin)
The logical space required for the virtual disk is not allocated during creation, but it is allocated on
demand during first write issued to the block. Just like thick disks, this format will also zero out the block
before writing data inducing extra I/O and an additional amount of write latency.
12.1.4
12.2
12.3
53
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
12.4
12.5
54
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
example PowerShell script, which can be used to backup physical mode RDMs as part of
the pre-execution steps of the backup job.
The free space recovery agent will also work with volumes mapped directly to the virtual machine
via the Microsoft Software iSCSI initiator.
- Volumes mapped to the virtual machine through the Microsoft iSCSI initiator interact with the
SAN directly, and consequently, space recovery works as intended.
For more information on Windows free space recovery and compatible versions of Windows,
please consult the Dell Compellent Enterprise Manager User Guide.
55
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
13
13.1
13.1.1
Figure 33 Datastore2 and Datastore3 can be grown by 100GB, but Datastore1 cannot
To extend the space at the end of a Storage Center volume as shown above, it can be done from the
Storage Center GUI or Enterprise Manager. After the volume has been extended and the hosts HBA has
been rescanned, the properties of the datastore can be edited to grow it by clicking on the Increase
button, and then follow through the Increase Datastore Capacity wizard.
Be careful to select the volume that is Expandable, otherwise the unintended action will actually add a
VMFS extent to the datastore (see section below on VMFS extents).
Figure 34 Screenshot from the wizard after extending a 500GB datastore by 100 GB
Warning: If a VMFS-3 volume (or pRDM residing on a VMFS-3 datastore) is extended beyond the 2 TB
limits, that volume will become inaccessible by the ESXi host. If this happens, the most likely scenario
will result in recovering data from a replay or tape.
56
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
Note: As an alternative to extending a datastore volume when a virtual machine needs additional disk
space, consider creating a new datastore volume and migrating that virtual machine. This will help to
keep volume sizes manageable, as well as help to keep any single datastore from being overloaded due
to I/O contention.
Note: All of the above tasks can be automated by using the Dell Compellent vSphere Client Plug-in. This
can be downloaded from the Dell Compellent Knowledge Center.
13.1.2
13.2
Figure 35 Growing a virtual disk from the virtual machine properties screen
For Windows machines: After growing the virtual disk from the vSphere client, an administrator must log
into the virtual machine, rescan for new disks, and then use DISKPART or the disk management MMC
console to extend the drive.
Warning: Microsoft does not support extending the system partition (C: drive) of a machine in certain
versions of Windows.
57
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
13.3
58
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
14
14.1
14.1.1
14.1.2
59
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
data recovered will be as if the virtual machine had simply lost power. Most modern journaling
file systems such as NTFS or EXT3 are designed to recover from such states.
Replays taken via Dell Compellents Replay Manager Software
- Since virtual machines running transactional databases are more sensitive to crash consistent
data, Dell Compellent has developed its Replay Manger software to utilize Microsofts VSS
framework for taking replays of Microsoft Exchange and SQL databases. This software
package will ensure that the database is in a consistent state before executing the replay.
Replays taken via Dell Compellents scripting tools
- For applications that need a custom method for taking consistent replays of the data, Dell
Compellent has developed two scripting tools:
o Dell Compellent Command Utility (CompCU) This is a java based scripting tool that
allows scripting for many of the Storage Centers tasks (such as taking replays).
o Storage Center Command Set for Windows PowerShell This scripting tool will also
allow scripting for many of the same storage tasks using Microsofts PowerShell scripting
language.
- A good example of using one of these scripting utilities is writing a script to take a replay of an
Oracle database after it is put into hot backup mode.
Replays used for Storage Center Replication and VMware Site Recovery Manager
- Replicating replays to a disaster recovery site not only ensures an off-site backup, but in
addition when using Site Recovery Manager, provides an automated recovery of the virtual
infrastructure in the event a disaster is declared.
14.2
From the Storage Center GUI or Enterprise Manager, select the replay to recover from and then
choose: Local Recovery or Create Volume From Replay
2. Continue through the local recovery wizard to create the view volume, and map it to the ESXi host
designated to recover the data.
a. Be sure to map the recovery view volume using a LUN which is not already in use.
3. Rescan the HBAs from the Storage Adapter section to detect the new LUN
4. From the vSphere client, highlight an ESXi host, then select the configuration tab
a. Select Storage
b. Click Add Storage
c. Select Disk/LUN and then click Next
d. Select the LUN for the view volume that was just mapped to the host and then click Next.
e. Three options are presented:
i. Keep the Existing Signature This option should only be used if the original datastore
is not present on the host.
60
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
5.
ii. Assign a New Signature This option will regenerate the datastore signature so that it
can be accessed by the host. (Select this option if unsure of which option to use.)
iii. Format the disk This option will format the view volume, and create a new datastore
from it.
f. Finish through the wizard verifying all selections.
Once the datastore has been resignatured, the snap datastore will be accessible:
14.2.1
14.2.2
61
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
14.2.3
Note: Once the virtual machine has been recovered, it can be migrated back to the original datastore
using Storage vMotion, provided that the original datastore has enough space, or the original virtual
machine was deleted.
62
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
15
63
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
The benefit of an asynchronous replication is that the replays are transferred to the destination
volume, allowing for check-points at the source system as well as the destination system.
15.1
15.2
15.2.1
64
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
attempt to map the destination Live Volume as LUN 10. In some instances, the secondary site may
already have existing LUNs that conflict, thereby causing the Live Volume not operate as expected.
Due to the asynchronous nature of the replication and how it proxies data, it is recommended that Async
Live Volume only be used for Disaster Avoidance or planned migration. In other words, Live Volume is
most practically used for planned maintenance operations such as migrating workloads between racks or
datacenters. Because of the nature of the proxied I/O, any disruption to the link or primary Live Volume
will cause the secondary Live Volume datastore to become unavailable as well. If for any reason the
primary Live Volume goes down permanently, this means that admins will need to perform a recovery on
the secondary Live Volume from the last known good Replay. The Enterprise Manager Disaster recovery
wizard is designed to assist in the recovery in such situations.
15.2.2
65
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
15.3
15.4
Caution: It is extremely important that the destination volume, usually denoted by Repl of, never gets
directly mapped to an ESXi host while data is actively being replicated. Doing so will inevitably cause data
integrity issues in the destination volume, requiring the entire volume be re-replicated from scratch. The
safest recovery method is to always restore the virtual machine from a local recovery or view volume as
shown in previous sections. Please see the Copilot Services Technical Alert (CSTA) titled, Mapping
Replicated Volumes at a DR Site available on Dell Compellent Knowledge Center for more info.
66
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
16
16.1
67
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
The default setting for the congestion threshold is 30 milliseconds of latency or 90% of peak throughput.
16.2
68
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
When a virtual machine moves between datastores, its actual location at the time the replay
was taken may make the virtual machine harder to find during a recovery. For example, if a
virtual machine moves twice in one week while daily replays are being taken, the appropriate
volume that contains the good replay of the virtual machine may be difficult to locate.
- If the Storage Center version in use does not support VAAI, the move process could be slow
(without Full Copy) or leave storage allocated on its originating datastore (without Dead Space
Reclamation). See the section on VAAI for more information.
Storage DRS could produce incorrect recommendations for virtual machine placement when I/O
metric inclusion is enabled on a Storage Center system using Data Progression. This is because
when a datastore is inactive, the SIOC injector will perform random read tests to determine
latency statistics of the datastore. With Data Progression enabled, the blocks that SIOC reads to
determine datastore performance, could potentially reside on SSD, 15K, or even 7K drives. This
could ultimately skew the latency results and decrease the effectiveness of the SRDS
recommendations.
-
16.3
16.3.1
69
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
16.3.2
16.3.3
16.3.4
16.3.5
70
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
For more important information about VAAI primitives please see the following resources:
VMware document: vSphere Storage Guide
- Subsection: Array Thin Provisioning and VMFS Datastores
- Subsection: Storage Hardware Acceleration
VMware KB: Disabling VAAI Thin Provisioning Block Space Reclamation (UNMAP) in ESXi 5.0
- Link: http://kb.vmware.com/kb/2007427
71
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
17
17.1
Prerequisite reading
To fully understand the recommendations made within this chapter, it is highly suggested to read the
following documents before proceeding:
VMware NFS Best Practices
- http://www.vmware.com/files/pdf/VMware_NFS_BestPractices_WP_EN.pdf
FS8600 Networking Best Practices
- http://en.community.dell.com/techcenter/extras/m/white_papers/20437940.aspx
17.2
FS8600 architecture
FS8600 scaleout NAS consists of one to four FS8600 appliances configured as a FluidFS cluster. Each NAS
appliance is a rackmounted 2U chassis that contains two hotswappable NAS controllers in an active
active configuration. In each NAS appliance, the second NAS controller with which one NAS controller is
paired is called the peer controller. FS8600 scaleout NAS supports expansion that is, one can start with a
single NAS appliance and add NAS appliances to the FluidFS cluster as needed to increase performance.
72
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
17.2.1
17.2.2
17.2.3
17.3
17.3.1
73
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
17.3.2
17.3.3
Additional resources
VMware has knowledge base articles available to further explain these settings:
VMware KB: Increasing the default value that defines the maximum number of NFS mounts on an
ESXi/ESX host
- Link: http://kb.vmware.com/kb/2239
VMware KB: Definition of the advanced NFS options
- Link: http://kb.vmware.com/kb/1007909
17.4
17.4.1
74
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
75
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
17.4.2
Enabling deduplication
To enable FluidFS deduplication feature it needs to be enabled globally on FluidFS cluster. Once it is
enable globally, deduplication can be enabled on a per volume basis. The following screenshot
demonstrates how to enable deduplication globally and also per volume.
1)
To do this, right click on the FluidFS cluster and then click Edit Settings.
76
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
3) Now that deduplication has been enabled globally, navigate back to FluidFS volumes and choose the
volume that contains the VMware Export from the previous steps, right click on that volume and
choose Edit Data Reduction Setting.
4) In the following screen, click on the Enable check box and create the deduplication policy
17.4.3
77
From VMware vSphere client, select an ESXi host, click on the Configuration tab
Click on Storage
Click on Add Storage
In the add storage wizard choose Network File System and then click Next
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
78
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
9) Repeat the steps 1-8 above for each host within the cluster
79
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
18
Conclusion
This document addresses many of the best practices scenarios an administrator may encounter while
implementing or upgrading VMware vSphere with the Dell Compellent Storage Center.
18.1
More information
If more information is needed, please review the following web sites:
Dell Compellent
- General Web Site
o http://www.dellstorage.com/compellent/
- Compellent Customer Portal
o http://customer.compellent.com
- Knowledge Center
o http://kc.compellent.com
- Dell Tech Center
o http://en.community.dell.com/techcenter/storage/w/wiki/5018.compellent-technical-
content.aspx
-
Training
o http://www.dellstorage.com/storage-services/training-services/
VMware
- General Web Site
o http://www.vmware.com/
- VMware Education and Training
o http://www.vmware.com/education
- vSphere Online Documentation
o http://www.vmware.com/support/pubs/vsphere-esxi-vcenter-server-pubs.html
- VMware Communities
o http://communities.vmware.com
18.2
Getting help
Contacting Copilot Support
For customers in the United States:
Telephone: 866-EZ-STORE (866-397-8673)
- E-mail [email protected]
Additional Copilot Support contact information for other countries can be found at:
- Contacting Compellent Technical Support
80
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
81
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
ESXi
It is important to note that although the per LUN queue depth maximum is 256, the per adapter
maximum within ESXi is 4096. By increasing the per LUN queue depth from 64 to 128, it can take
fewer LUNs to saturate a ports queue. For example, 4096/64=64 LUNs, whereas 4096/128=32
LUNs.
The best way to determine the appropriate queue depth is by using the esxtop utility. This utility can be
executed from one of the following locations:
ESXi Shell via SSH
- Command: esxtop
vCLI 5.x or the vMA Virtual Appliance
- Command: resxtop (or resxtop.sh)
When opening the esxtop utility, the best place to monitor queue depth and performance is from the Disk
Device screen. Here is how to navigate to that screen:
1.
82
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
sense. However, if the LOAD is consistently less than 1.00 on a majority of the LUNs, and the performance
and latencies are acceptable, then there is usually no need to adjust the queue depth.
In the figure above, the device queue depth is set to 32. Please note that three of the four LUNs
consistently have a LOAD above 1.00. If the back end spindles are not maxed out, it may make sense to
increase the queue depth, as well as increase the DSNRO setting.
83
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020
84
Dell Compellent Storage Center Best Practices with VMware vSphere 5.x | Document number 680-041-020