B HXDP Admin Guide 5 5
B HXDP Admin Guide 5 5
B HXDP Admin Guide 5 5
5
First Published: 2023-08-22
Last Modified: 2023-12-21
Americas Headquarters
Cisco Systems, Inc.
170 West Tasman Drive
San Jose, CA 95134-1706
USA
http://www.cisco.com
Tel: 408 526-4000
800 553-NETS (6387)
Fax: 408 527-0883
THE SPECIFICATIONS AND INFORMATION REGARDING THE PRODUCTS IN THIS MANUAL ARE SUBJECT TO CHANGE WITHOUT NOTICE. ALL STATEMENTS,
INFORMATION, AND RECOMMENDATIONS IN THIS MANUAL ARE BELIEVED TO BE ACCURATE BUT ARE PRESENTED WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED. USERS MUST TAKE FULL RESPONSIBILITY FOR THEIR APPLICATION OF ANY PRODUCTS.
THE SOFTWARE LICENSE AND LIMITED WARRANTY FOR THE ACCOMPANYING PRODUCT ARE SET FORTH IN THE INFORMATION PACKET THAT SHIPPED WITH
THE PRODUCT AND ARE INCORPORATED HEREIN BY THIS REFERENCE. IF YOU ARE UNABLE TO LOCATE THE SOFTWARE LICENSE OR LIMITED WARRANTY,
CONTACT YOUR CISCO REPRESENTATIVE FOR A COPY.
The Cisco implementation of TCP header compression is an adaptation of a program developed by the University of California, Berkeley (UCB) as part of UCB's public domain version of
the UNIX operating system. All rights reserved. Copyright © 1981, Regents of the University of California.
NOTWITHSTANDING ANY OTHER WARRANTY HEREIN, ALL DOCUMENT FILES AND SOFTWARE OF THESE SUPPLIERS ARE PROVIDED “AS IS" WITH ALL FAULTS.
CISCO AND THE ABOVE-NAMED SUPPLIERS DISCLAIM ALL WARRANTIES, EXPRESSED OR IMPLIED, INCLUDING, WITHOUT LIMITATION, THOSE OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OR ARISING FROM A COURSE OF DEALING, USAGE, OR TRADE PRACTICE.
IN NO EVENT SHALL CISCO OR ITS SUPPLIERS BE LIABLE FOR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, OR INCIDENTAL DAMAGES, INCLUDING, WITHOUT
LIMITATION, LOST PROFITS OR LOSS OR DAMAGE TO DATA ARISING OUT OF THE USE OR INABILITY TO USE THIS MANUAL, EVEN IF CISCO OR ITS SUPPLIERS
HAVE BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
Any Internet Protocol (IP) addresses and phone numbers used in this document are not intended to be actual addresses and phone numbers. Any examples, command display output, network
topology diagrams, and other figures included in the document are shown for illustrative purposes only. Any use of actual IP addresses or phone numbers in illustrative content is unintentional
and coincidental.
All printed copies and duplicate soft copies of this document are considered uncontrolled. See the current online version for the latest version.
Cisco has more than 200 offices worldwide. Addresses and phone numbers are listed on the Cisco website at www.cisco.com/go/offices.
Cisco and the Cisco logo are trademarks or registered trademarks of Cisco and/or its affiliates in the U.S. and other countries. To view a list of Cisco trademarks, go to this URL:
https://www.cisco.com/c/en/us/about/legal/trademarks.html. Third-party trademarks mentioned are the property of their respective owners. The use of the word partner does not imply a
partnership relationship between Cisco and any other company. (1721R)
© 2023 Cisco Systems, Inc. All rights reserved.
CONTENTS
SED Encryption 85
Self-Encrypting Drives Overview 85
Verify if the HyperFlex Cluster Is Encryption Capable 85
Configuring Local Encryption Key 86
Modifying Local Encryption Key 86
Disabling Local Encryption Key 87
Secure Erase an Encrypted Disk 87
Remote Key Management 88
Configuring Remote Encryption Key 88
Generating Certificate Signing Requests 89
Configuring a Key Management Server Using CSRs (Certificate Signing Requests) 90
Generating Self-Signed Certificates 91
Configuring a key management server using SSCs (Self-Signed Certificates) 92
Restart Encryption 93
HyperFlex Software Encryption 94
Enabling HyperFlex Software Encryption Workflow 94
HyperFlex Software Encryption Guidelines and Limitations 94
Install HX Software Encryption Package: Clusters with 1 - 12 Nodes 95
Install HX Software Encryption Package: Clusters with 13+ Nodes 95
Backup Encryption Key of HyperFlex Software Encryption 96
Secure Disk Erase for HyperFlex Software Encryption 97
Managing Datastores 99
Adding Datastores 100
Editing Datastores 101
Unmounting Datastores 102
Deleting Datastores 102
Encryption Support for Datastores 103
Recovering from Partially Unmounted Datastores 104
Windows Site Recovery Manager (SRM 6.1 through SRM 8.3) 264
Installing a Site Recovery Manager for Windows 264
Installing a Storage Replication Adapter for Windows 265
Site Recovery Manager Virtual Appliance (SRM 8.4 and later) 265
Deploy the Site Recovery Manager Virtual Appliance 265
Installing a Photon OS Storage Replication Adapter 266
Setting up SRM Environment 266
Pairing Sites 266
Configuring Mapping 266
Refreshing Storage Replication Adapter 267
Adding an Array Manager Pair 268
Mapping Datastores on Cisco HyperFlex 268
Discovering Devices 269
Creating a Protection Group 269
Creating a Recovery Plan 269
Executing SRA Workflows 270
Testing a Recovery Plan 270
Testing the Recovery Plan Cleanup Task 271
Executing a Recovery 271
Executing a Reprotect Task 272
Troubleshooting SRA 272
Troubleshooting: Using Windows SRM Log Files 272
HyperFlex iSCSI Target Service Overview and Supported Use Cases 279
HyperFlex iSCSI Best Practices 280
Disks 316
Nodes 317
Network 319
iSCSI 321
Documentation Feedback
To provide feedback about Cisco technical documentation, use the feedback form available in the right pane
of every online document.
Bias-Free Language
The documentation set for this product strives to use bias-free language. For purposes of this documentation
set, bias-free is defined as language that does not imply discrimination based on age, disability, gender, racial
identity, ethnic identity, sexual orientation, socioeconomic status, and intersectionality. Exceptions may be
present in the documentation due to language that is hardcoded in the user interfaces of the product software,
language used based on standards documentation, or language that is used by a referenced third-party product.
HX Remote Plugin for Introduced the HTML 5.5(1a) Cisco HyperFlex Remote
VMware vCenter remote plugin Plugin for VMware
vCenter, on page 352
• Converged nodes—Converged nodes are the physical hardware on which the VM runs. They provide
computing and storage resources such as disk space, memory, processing, power, and network I/O.
When a converged node is added to the storage cluster, a storage controller VM is installed. The HX
Data Platform services are handled through the storage controller VM. Converged nodes add storage
resources to your storage cluster through their associated drives.
Run the Cluster Expansion workflow from the HX Data Platform Installer to add converged nodes to
your storage cluster. You can remove converged nodes using hxcli commands.
• Compute nodes—Compute nodes add compute resource but not storage capacity to the storage cluster.
They are used as a means to add compute resources, including CPU and memory. They do not need to
have any caching (SSD) or storage (HDD) drives. Compute nodes are optional in a HX storage cluster.
When a compute node is added to the storage cluster, an agent controller VM is installed. The HX Data
Platform services are handled through the agent controller VM.
Run the Cluster Expansion workflow from the HX Data Platform Installer to add compute nodes to your
storage cluster. You can remove compute nodes using hxcli commands.
• Drives—There are two types of drives that are required for any node in the storage cluster: Solid State
Drive (SSD) and Hard Disk Drive (HDD). HDD typically provides the physical storage units associated
with converged nodes. SSD typically supports management.
Adding HDD to existing converged nodes, also adds storage capacity to the storage cluster. When storage
is added to a HX node in the storage cluster, an equal amount of storage must be added to every node in
the storage cluster.
When disks are added or removed, the HX Data Platform rebalances the storage cluster to adjust for the
change in storage resources.
Adding or removing disks on your converged nodes is not performed through the HX Data Platform.
Before adding or removing disks, review the best practices. See the server hardware guides for specific
instructions to add or remove disks in nodes.
NVMe Caching SSD's slot information is unavailable from HX-Connect for all AF server PIDs except
for the All-NVMe server PIDs. Please refer to UCSM management console for NVMe SSD slot
information.
• Datastores—Storage capacity and datastore capacity. This is the combined consumable physical storage
available to the storage cluster through datastores, and managed by the HX Data Platform.
Datastores are logical containers that are used by the HX Data Platform to manage your storage use and
storage resources.
Datastores are where the host places virtual disk files and other VM files. Datastores hide the specifics
of physical storage devices and provide a uniform model for storing VM files.
Note Capacity addition in a cluster through the addition of disks or nodes can result in a rebalance. This background
activity can cause interference with regular User IO on the cluster and increase the latency. You must note
the time duration for the storage capacity at the time where performance impact can be tolerated. Also, this
operation may be performed in urgent situations that may warrant capacity addition.
In the HX Data Platform the concept of capacity is applied to both datastores and storage clusters. Values are
measured in base-2 (GiB/TiB), but for simplicity and consistency are labeled as GB or TB.
• Cleaner―A process run on all the storage cluster datastores. After it completes, all the storage cluster
datastores total capacity should be in a similar range to the total storage cluster capacity, excluding the
metadata. Datastore capacity listed typically will not match the HX storage cluster capacity. See the
Cisco HX Data Platform Command Line Interface Reference Guide for information on the cleaner
command.
• Cluster capacity―All the storage from all the disks on all the nodes in the storage cluster. This includes
uncleaned data and the metadata overhead for each disk.
The total/used/free capacity of cluster is based on overall storage capacity and how much storage is used.
• Condition―When the HX Storage Cluster enters a space event state, the Free Space Status fields are
displayed. The Condition field lists the space event state. The options are: Warning, Critical, and Alert.
• Available Datastore capacity―The amount of storage available for provisioning to datastores without
over-provisioning. Generally, this is similar to the cleaned storage cluster capacity, but it is not an exact
match. It does not include metadata or uncleaned data.
The provisioned/used/free capacity of each datastore is based on datastore (thin) provisioned capacity.
Because the datastore is thin provisioned, the provisioned capacity (specified by the administrator when
creating the datastore) can be well above the actual storage.
• Free Capacity, storage cluster―Same as available capacity. For the storage cluster, this is the difference
between the amount available to the storage cluster and the amount used in the storage cluster.
• Free capacity, datastore―Same as available capacity. For all the storage cluster datastores, this is the
difference between the amount provisioned to all the storage cluster datastores and the amount used on
all the storage cluster datastores.
The amount used on the whole storage cluster is not included in this datastore calculation. Because
datastores are frequently over provisioned, the free capacity can indicate a large availability on all the
storage cluster datastores, while the storage cluster capacity can indicate a much lower availability.
• Multiple users―Can have different datastores with different provisioned capacities. At any point in
time, users do not fully utilize their allocated datastore capacity. When allocating datastore capacity to
multiple users, it is up to the administrator to ensure that each user’s provisioned capacity is honored at
all time.
• Over-provisioning―Occurs when the amount of storage capacity allocated to all the datastores exceeds
the amount available to the storage cluster.
It is a common practice to initially over-provision. It allows administrators to allocate the capacity now
and backfill the actual storage later.
The value is the difference between the usable capacity and provisioned capacity.
It displays zero (0) value, unless more space has been allocated than the maximum physical amount
possible.
Review the over provisioned capacity and ensure that your system does not reach an out-of-space
condition.
• Provisioned―Amount of capacity allowed to be used by and allocated to the storage cluster datastores.
The provisioned amount is not set aside for the sole use of the storage cluster datastores. Multiple
datastores can be provisioned storage from the same storage capacity.
• Space Needed―When the HX Storage Cluster enters a space event state, the Free Space Status fields
are displayed. Space Needed indicates the amount of storage that needs to be made available to clear
the listed Condition.
• Used―Amount of storage capacity consumed by the listed storage cluster or datastore.
HX Data Platform internal meta-data uses 0.5% to 1% space. This might cause the HX Data Platform
Plug-in or HX Connect to display a Used Storage value even if you have no data in your datastore.
Storage Used shows how much datastore space is occupied by virtual machine files, including configuration
and log files, snapshots, and clones. When the virtual machine is running, the used storage space also
includes swap files.
• Usable Capacity―Amount of storage in the storage cluster available for use to store data.
Deduplication savings and compression savings are not simply added together. They are not independent
operations. They are correlated using the following elements where essentially the number of unique bytes
used for storage is reduced through deduplication. Then the deduplicated storage consumption is compressed
to make even more storage available to the storage cluster.
Deduplication and compression savings are useful when working with VM clones.
If the savings is showing 0%, this indicates the storage cluster is new. The total ingested data to the storage
cluster is insufficient to determine meaningful storage savings. Wait until sufficient data is written to the
storage cluster.
For example:
1. Initial values
2. Deduplication savings
= (1 - TUUS/TAS) * 100
= 50%
3. Compression Savings
= (1 - TUB/TUUS) * 100
= 75%
= 87.5%
Divide the result by 1024 to get a value in TiB. The replication factor value is 3 if the HX cluster is set to
RF=3, and the value is 2 if the HX cluster is set to RF=2. The 0.92 multiplier accounts for an 8% reservation
set aside on each disk by the HX Data Platform software for various internal filesystem functions.
Calculation Example: <capacity
disk size in GB> = 1200 for 1.2 TB disks <number of capacity
disks per node> = 15 for an HX240c-M6SX model server <number of HyperFlex nodes> = 8
replication factor = 3
Note This formula for calculating cluster capacity does not apply for Large Form Factor (LFF) clusters.
Error Messages
Error messages are issued if your data storage needs to consume high amounts of available capacity, the
performance and health of your storage cluster are affected. The error messages are displayed in vCenter
Alarms panels, HX Connect, and HX Data Platform Plug-in Alarms and Events pages.
Note The event and alarm details provided on vCenter and HX Connect are not always a 1:1 relationship. When
reviewing messages in HX Connect, it is a best practice to also review the events and tasks in vCenter.
The number of nodes in the storage cluster, combined with the Data Replication Factor and Access Policy
settings, determine the state of the storage cluster that results from node failures.
Note Before using the HX Data Platform HA feature, enable DRS and vMotion on the vSphere Web Client.
The following settings take effect when the storage cluster transitions into particular operational and resiliency
status states.
• Data Replication Factor —Sets the number of redundant data replicas.
• Cluster Access Policy—Sets the level of data protection and data loss.
Other transitional states might be displayed during cluster upgrades and cluster creation.
Color coding and icons are used to indicated various status states. Click icons to display additional information
such as reason messages that explain what is contributing to the current state.
Color coding and icons are used to indicate various status states. Click an icon to display additional information,
such as reason messages that explain what is contributing to the current state.
• Access Policy—Can be changed from the default setting after the storage cluster is created. The options
are strict for protecting against data loss, or lenient, to support longer storage cluster availability.
Table 1: Cluster State in 5+ Node Cluster with Number of Failed Nodes, HX Release 4.5(x) and later
Read/Write Read-Only
3 Lenient 2 --
3 Strict 1 2
2 Lenient 1 --
2 Strict -- 1
Table 2: Cluster State in 3 - 4 Node Clusters with Number of Failed Nodes HX Release 4.5(x) and later.
Read/Write Read-Only
3 Lenient or Strict 1 --
2 Lenient 1 --
2 Strict -- 1
Note HX storage clusters are capable of sustaining serial disk failures, (separate disk failures over time). The only
requirement is that there is sufficient storage capacity available for support self-healing. The worst-case
scenarios listed in this table only apply during the small window while HX is completing the automatic
self-healing and rebalancing.
3 Lenient 2 --
3 Strict 1 2
2 Lenient 1 --
2 Strict -- 1
Important Data Replication Factor cannot be changed after the storage cluster is configured.
Data Replication Factor is set when you configure the storage cluster. Data Replication Factor defines the
number of redundant replicas of your data across the storage cluster. The options are 2 or 3 redundant replicas
of your data.
• If you have hybrid servers (servers that contain both SSD and HDDs), then the default is 3.
• If you have all flash servers (servers that contain only SSDs), then you must explicitly select either 2 or
3 during HX Data Platform installation.
3 nodes 1 One node. The storage cluster does not automatically heal.
Replace the failed node to restore storage cluster health.
3 nodes 2 Two or more disks on two a. If one cache SSD fails, the storage cluster does not
nodes are blocklisted or automatically heal.
failed.
b. If one HDD fails or is removed, the disk is blocklisted
immediately. The storage cluster automatically begins
healing within a minute.
c. If more than one HDD fails, the system might not
automatically restore storage cluster health.
If the system is not restored, replace the faulty disk and
restore the system by rebalancing the cluster.
4 nodes 1 One node. If the node does not recover in two hours, the storage cluster
starts healing by rebalancing data on the remaining nodes.
To recover the failed node immediately and fully restore the
storage cluster:
a. Check that the node is powered on and restart it if
possible. You might need to replace the node.
b. Rebalance the cluster.
4 nodes 2 Two or more disks on two If two SSDs fail, the storage cluster does not automatically
nodes. heal.
If the disk does not recover in one minute, the storage cluster
starts healing by rebalancing data on the remaining nodes.
5+ nodes 2 Up to two nodes. If the node does not recover in two hours, the storage cluster
starts healing by rebalancing data on the remaining nodes.
To recover the failed node immediately and fully restore the
storage cluster:
a. Check that the node is powered on and restart it if
possible. You might need to replace the node.
b. Rebalance the cluster.
5+ nodes 2 Two nodes with two or The system automatically triggers a rebalance after a minute
more disk failures on each to restore storage cluster health.
node.
5+ nodes 2 One node and One or more If the disk does not recover in one minute, the storage
disks on a different node. cluster starts healing by rebalancing data on the remaining
nodes.
If the node does not recover in two hours, the storage cluster
starts healing by rebalancing data on the remaining nodes.
If a node in the storage cluster fails and a disk on a different
node also fails, the storage cluster starts healing the failed
disk (without touching the data on the failed node) in one
minute. If the failed node does not come back up after two
hours, the storage cluster starts healing the failed node as
well.
To recover the failed node immediately and fully restore the
storage cluster:
a. Check that the node is powered on and restart it if
possible. You might need to replace the node.
b. Rebalance the cluster.
For additional information about VMware snapshots, see the "Overview of virtual machine snapshots in
vSphere (KB 1015180)" on the VMware Customer Connect site.
• VMware vSphere Web Client and vSphere Client―Managing all the VMware ESXi servers in the vCenter
cluster.
• VMware ESXi ―Managing the individual ESXi host, providing host command line.
Admin Shell User defined during HX As specified Must match across all nodes
installation. during HX in storage cluster.
installation.
Predefined admin user. Support for SSH to the
Strong secure admin shell is limited
password to the user admin.
required.
Use the hxcli command
when changing the
password after installation.
vCenter admin [email protected] SSO enabled. Read only users do not have
default. access to HX Data Platform
As configured.
Plug-in.
SSO enabled.
As configured,
MYDOMAIN\name or
[email protected]
ESXi Server root SSO enabled. SSO enabled. Must match across all ESX
servers in storage cluster.
As configured. As configured.
Special characters―The following special characters are acceptable for user virtual machine or datastore
names:
• ampersand (&), apostrophe ('), asterisk (*), at sign (@), back slash (\), circumflex (^), colon (:), comma
(,), dollar sign ($), dot (.), double quotation ("), equal sign (=), exclamation (!), forward slash (/), hyphen
(-), left curly brace ({), left parentheses ((), left square bracket ([), less than sign (<), more than sign (>),
percent (%), pipe (|), plus sign (+), pound (#), question mark (?), right curly brace (}), right parentheses
()), right square bracket (]), semi-colon (;), tilde (~), underscore (_)
Username Requirements
Usernames can be specific to the HX Data Platform component and must meet UCS Manager username
requirements.
UCS Manager username requirements.
• Number of characters: between 6 and 32 characters
• Must be unique within Cisco UCS Manager.
• Must start with an alphabetic character.
• Must have alphabetic characters (upper or lower case).
• Can have numeric characters. Cannot be all numeric characters.
• Special characters: Limited to underscore (_), dash (-), and dot (.)
Note General rule about passwords: Do not include them in a command string. Allow the command to prompt for
the password.
• Minimum Length: 10
• Minimum 1 Uppercase
• Minimum 1 Lowercase
• Minimum 1 Digit
• Minimum 1 Special Character
• A maximum of 3 retry to set the new password
To change a controller VM password, always use the hxcli command. Do not use another change password
command, such as a Unix password command.
1. Log into the management controller VM.
2. Run the hxcli command.
hxcli security password set [-h] [--user USER]
The change is propagated to all the controller VMs in the HX cluster.
• Dictionary words:
Note HxConnect makes use of /auth call for login purpose and the limit applies there also.
Important • If you are a read-only user, you may not see all of the options described in the Help. To perform most
actions in HX Connect, you must have administrative privileges.
• Ensure that the time on the vCenter and the controller VMs are in sync or near sync. If there is too large
of a time skew between the vCenter time and the cluster time, AAA authentication will fail.
These users are created through vCenter. vCenter username format is: <name>@domain.local and specified in the
User Principal Name Format (UPN). For example, [email protected]. Do not add a prefix such as
"ad:" to the username.
• HX pre-defined users―To login using the HX Data Platform predefined users admin or root, enter a prefix local/.
For example: local/root or local/admin.
Actions performed with the local/ login only affect the local cluster.
vCenter recognizes the session with HX Connect, therefore system messages that originate with vCenter might
indicate the session user instead of local/root. For example, in Alarms, Acknowledged By might list
com.springpath.sysmgmt.domain-c7.
Click the eye icon to view or hide the password field text. Sometimes this icon is obscured by other field elements. Click
the eye icon area and the toggle function continues to work.
What to do next
• To refresh the HX Connect displayed content, click the refresh (circular) icon. If this does not refresh
the page, the clear the cache and reload the browser.
• To logout of HX Connect, and properly close the session, select User menu (top right) > Logout.
To execute HX Data Platform hxcli commands, log into the HX Data Platform Storage Controller VM
command line.
Important Do not include passwords in command strings. Commands are frequently passed to the logs as plain text.
Wait until the command prompts for the password. This applies to login commands as well as hxcli commands.
You may log into the HX Data Platform command line interface in the Storage Controller VM in the following
ways:
• From a command terminal
• From HX Connect Web CLI page
Only direct commands are supported through HX Connect.
• Direct commands―commands that complete in a single pass and do not require responses through
the command line. Example direct command: hxcli cluster info
• Indirect commands―multi-layered commands that require live response through the command line.
Example interactive command: hxcli cluster reregister
Step 2 From a browser, enter the DNS Name and /cli path.
a) Enter the path.
Example
# cs002-stctlvm-a.eng.storvisor.com/cli
This command applies the change to all the controller VMs in the storage cluster.
Note If you add new compute nodes and try to reset the cluster password using the hxcli security password
set command, the converged nodes get updated, but the compute nodes may still have the default password.
Note Before launching the Cisco HX Data Platform Installer, ensure that all the ESXi servers that are in the vCenter
cluster that you plan to include in the storage cluster are in maintenance mode.
Step 1 In a browser, enter the URL for the VM where HX Data Platform Installer is installed.
You must have this address from the earlier section on Deploying HX Data Platform Installer. For example
http://10.64.4.254
Attention Systems ship with a default password of Cisco123 that must be changed during installation. You cannot
continue installation unless you specify a new user supplied password.
Step 3 The HX Data Platform Installer Workflow page provides two options to navigate further.
• Create Cluster drop-down list—You can deploy a standard cluster, or a Stretched cluster.
• Cluster Expansion—You can provide the data to add converged nodes and compute nodes to an existing standard
storage cluster.
Step 3 Run the recover-password command. A prompt appears requesting Consent Token.
Note Contact TAC to help provide the Consent Token.
After using the recover-password command to change the password, passwords will no longer be synced on all nodes.
You will need to use hxcli security password set to change and sync the password again on all nodes.
Step 4 To sync the password on all nodes, run the hxcli security password set command from any node, and enter the new
password.
Example
admin:~$ hxcli security password set
Enter new password for user admin:
Re-enter new password for user admin:
admin:~$
Any command that would cause changes to the system software are blocked for the "diag" user. The default
list of blocked commands include:
• sudo
• apt-get
• li
• dpkg
• apt
• easy_install
• setfacl
• adduser
• deluser
• userdel
• groupadd
• groupdel
• addgroup
• delgroup
The Upgrade page provides access to HX Data Platform and Cisco UCS Manager firmware upgrade tasks.
Dashboard Page
Important If you are a read-only user, you may not see all of the options available in the Help. To perform most actions
in HyperFlex (HX) Connect, you must have administrative privileges.
Displays a status summary of your HX storage cluster. This is the first page that you see when you log into
Cisco HyperFlex Connect.
Resiliency Health section Provides the data health status and ability of the HX storage cluster to
tolerate failures.
Capacity section Displays a breakdown of the total storage versus how much storage is
used or free.
Also displays the storage optimization, compression-savings, and
deduplication percentages based on the data stored in the cluster.
Nodes section Displays the number of nodes in the HX storage cluster, and the division
of converged versus compute nodes. Hovering over a node icon displays
that node's name, IP address, node type, and an interactive display of
disks with access to capacity, usage, serial number, and disk type data.
Cluster Time field System date and time for the cluster.
Export menu Save a copy of the current page of table data. The table content is
downloaded to the local machine in the selected file type. If the listed
items are filtered, the filtered subset list is exported.
Click the down arrow to select an export file type. The file type options
are: cvs, xls, and doc.
To export content from other pages in the table, scroll to the bottom,
click through the page numbers, and apply the export.
Activity Page
Displays a list of recent activity on the HX storage cluster allowing you to monitor the progress of VM
operations, Cluster upgrade/expansion, enter/exit maintenance mode, and recovery jobs.
Expand All / Collapse All button Toggles the view of the job list to display top-level task information or
task details.
You can also expand and collapse individual tasks.
The following table specifies which Snapshot operations create an HX Task in the Activity Page.
Scheduled Snapshot task creation from HX Connect HX task added to the Activity page.
Snapshot creation from Schedule Snapshot HX task added to the Activity page.
Cluster License Status section Displays the Register Now link when you log into the HX storage cluster
for the first time or till the HX storage cluster license is registered:
Register Now link—To register a cluster license, click this link and
provide product instance registration token in the Smart Software
Licensing Product Registration screen. For more information on how
to get a product instance registration token, refer the Registering a
Cluster with Smart Licensing section in the Cisco HyperFlex Systems
Installation Guide for VMware ESXi.
Note To register a cluster license, you can also choose Register
Now from the Actions drop-down field.
HX storage cluster status field Provides functional status of the HX storage cluster.
• Online—Cluster is ready.
• Offline—Cluster is not ready.
• Read Only—Cluster is out of space.
• Unknown—Transitional state while the cluster is coming online.
vCenter link Secure URL to the VMware vSphere associated with this HX storage
cluster. Click the link to remotely access the vSphere Web Client .
HXDP Version field Installer package version installed on this HX storage cluster.
Data Replication Factor field Number of the redundant data replicas stored on this HX storage cluster.
Uptime field Length of time this HX storage cluster has been online.
DNS Server(s) IP address for the DNS server(s) for this HX storage cluster.
NTP Server(s) IP address for the NTP server(s) for this HX storage cluster.
Controller VM Access
Use Actions to access the controller VM using SSH as an administrator and perform actions such as Enable
Controller Access over SSH, Disable Controller Access over SSH or register your license.
Note Actions to enable or disable SSH can only be performed by domain users, and not local users. Domain users
are users in VC (ESXi).
Disable Controller Access over SSH Secure Shell (SSH) is disabled by default.
Node Data
Displays data about individual nodes in this HX storage cluster. To see this information in tabular format, go
to the Nodes page.
Type • Hyperconverged
• Compute
Hypervisor Address IP address for the management network to this HX storage cluster.
Disk Overview Graphic representation of the number of disks in use for each node, the
usage type and number of empty slots.
Note A disk outline with a red icon indicates a disk that is not
recognized and requires a Catalog Upgrade.
For nodes with disks, you can place your cursor over a disk to view an interactive display of information
including the following.
Disks
Locator LED Activates a physical light on the host to help locate a disk; options are
On and Off.
Used / Total Capacity (Persistent Amount of the disk used versus the total disk size.
Disks only)
Disk Drive Interface The disk drive interface type, for example SAS or SATA.
Nodes Page
Displays data about all of the nodes in this HX storage cluster in a table. Each column can be used to sort the
data.
Hypervisor Address column IP address for the management network of the Node referred in the Node
column.
Controller Address column IP address for the HX storage controller VM of the Node referred in the
Node column.
Controller Status column • Online—The connection between the VM and the disk is available.
• Offline—The connection between the VM and the disk is not
available.
• In Maintenance—the connection between the VM and the disk is
powered off from the host.
Version column HyperFlex Data Platform installer package version installed on this
node.
Disks Page
Displays data about all of the disks in this HX storage cluster in a 7-column table. Each column can be used
to sort the data.
Slot column Location of the SED drive. This identifies the drive for maintenance procedures.
(Optional) Secure erase This button in visible only if your HX storage cluster is encrypted using local-key
button encryption.
Select a disk to access the button.
Enter the encryption key in use on the cluster, click Secure erase, and then click
Yes, erase this disk to securely erase the local encryption key.
• /var/log/nginx/ssl-access.log
After you enable audit logging, these logs are exported to the remote syslog server. If the logs from the
controller VM are not pushed to the remote sylog server, or if the remote syslog server is not reachable, an
alarm is generated in the HX-Connect user interface. However, HX Connect does not monitor the disk space
available on the remote syslog server. The HX Connect user interface will not display an alarm if the disk on
the remote syslog server is full.
After you enable audit logging, you can choose to either temporarily disable audit logging, or you can choose
to delete the audit logging server configuration details.
Connection Type drop-down list Choose TLS or TCP as the connection type. The default and recommended
value is TLS. The TLS connection type is for encrypted transport over TLS.
The TCP connection type is for unencrypted transport over TCP.
Client Certificate Click Choose to search and locate a certificate file that must be stored on the
controller VM. This certificate creates a TLS connection between the controller
VM and the remote syslog server. A TLS connection ensures that the log files
are encrypted.
You must upload either a user-generated self-signed certificate or a CA-signed
certificate.
Private Key Click Choose to search and locate a generated private key file to be stored on
the controller VM. This key creates a TLS connection between the controller
VM and the remote syslog server.
Choosing a certificate and private key for the syslog server ensures that the log
files are encrypted. The certificate for the syslog server can either be a CA
certificate or a self-signed certificate.
Are you using a self-signed Check this check box if the syslog server uses a self-signed certificate.
certificate?
Click Choose to search and locate the self-signed certificate for the syslog
server.
########################
Following is a sample of the configuration file to establish a TCP connection with the remote syslog server:
#######################
## Audit Logging Configuration ###
source demo_tls_src {
tcp(ip(0.0.0.0) port(6515)
); };
########################
Restriction The following workflow should only be performed by Cisco TAC. If you have the need to manually rebalance
a cluster, contact TAC for assistance.
Note Forcing a manual rebalance can cause interference with regular User IO on the cluster and increase the latency.
Therefore, the HyperFlex system initiates a rebalance only when required in order to minimize performance
penalties.
This sample indicates that rebalance is enabled, and ready to perform a rebalance, but is not currently rebalancing
the storage cluster.
Important Rebalance typically occurs only when a single disk usage exceeds 50% or cluster aggregate disk usage is
greater than 50%.
You can check rebalance status through the HX Data Platform plug-in or through the storage controller VM
command line.
The following output indicates that rebalance is not currently running on the storage cluster.
rebalanceStatus:
percentComplete: 0
rebalanceState: cluster_rebalance_not_running
rebalanceEnabled: True
The Recent Tasks tab in the HX Data Platform plug-in displays a status message.
Note Do not delete storage controller VMs. Storage controller VM names have the prefix stCtlVM.
Step 1 To add a node, use the Expand Cluster feature of the HX Data Platform Installer.
Step 2 To delete unused VMs, complete the following:
a) Determine which guest VMs you can delete. You can consider factors such as disk space used by the VM or naming
conventions.
b) Go to vCenter > Virtual Machines to display the virtual machines in the inventory.
c) Double-click a VM that you want to delete.
d) Select the Summary > Answer Questions to display a dialog box.
e) Click the Cancel radio button and click OK.
f) Power off the VM.
g) Delete the VM.
Step 3 After the Out of Space condition is cleared, complete the following:
a) Go to vCenter > Virtual Machines to display the VM in the inventory.
b) Double-click a VM that you want to use.
c) Select the Summary > Answer Questions to display a dialog box.
d) Click the Retry radio button and click OK.
Step 2 On the new vCenter, create a new cluster using the same cluster name.
Step 3 Add ESX hosts to new vCenter in the newly created cluster.
What to do next
Proceed to Unregistering a Storage Cluster from a vCenter Cluster, on page 50.
Note • If multiple HX clusters are registered to the same vCenter, do not attempt this procedure until all HX
clusters have been fully migrated to different vCenter. Running this procedure is disruptive to any existing
HX clusters registered to the vCenter.
Step 1 Complete the steps in Removing HX Data Platform Files from the vSphere Client, on page 53.
Step 2 Complete the steps in Verifying HX Cluster is Unregistered from vCenter, on page 53.
What to do next
Proceed to Registering a Storage Cluster with a New vCenter Cluster, on page 54.
Note Newly deployed HX clusters starting with HyperFlex Release 4.0 no longer leverage the vSphere ESX Agent
Manager (EAM) for the HyperFlex Storage Controller VMs. HX clusters built prior to HX 4.0 will continue
to utilize EAM. If that cluster is migrated to a new vCenter, however, the EAM integration will not be
configured.
Step 2 To unregister the storage cluster extension: Log into the vCenter server MOB extension manager
First unregister the HyperFlex cluster.
a) In a browser, enter the path and command.
https://vcenter_server/mob/?moid=ExtensionManager
vcenter_server is the IP address of the vCenter where the storage cluster is currently registered.
b) Enter administrator login credentials.
Step 3 Locate the HX storage cluster extensions with the cluster IDs. Scroll through the Properties > extensionList to locate
the storage cluster extensions:
com.springpath.sysmgmt.cluster_domain_id and com.springpath.sysmgmt.uuid.cluster_domain_id.
Copy each of these strings into your clipboard. Exclude the double quotes (") on either end of string, if there are any.
vcenter_server is the IP address of the vCenter where the storage cluster is currently registered.
b) Enter administrator login credentials.
Step 6 Locate the stale HX storage cluster ESX agency extensions with the cluster IDs.
a) Scroll through the Properties > agency > Value.
b) Click an agency value.
c) In the Agency window, check the Properties > solutionID > Value extension. Verify has the correct
cluster_domain_id.
For example: com.springpath.sysgmt.domain-26
KB, Stopping, starting, or restarting VMware vCenter Server Appliance 6.0 services (2109887)article on the VMwear
Customer connect site.
Remove the HX Data Platform files from the vSphere Client. Select a method.
Linux vCenter
a) Log into the Linux vCenter server using ssh as a root user.
b) Change to the folder containing the HX Data Platform Plug-in folder.
For vCenter 6.0
# cd /etc/vmware/vsphere-client/vc-packages/vsphere-client-serenity/
Windows vCenter
a) Log into the Windows vCenter system command line using Remote Desktop Protocol (RDP).
b) Change to the folder containing the HX Data Platform Plug-in folder.
# cd "%PROGRAMDATA%\VMware\vSphere Web Client\vc-packages\vsphere-client-serenity
Example response:
If, after your storage cluster is re-registered, your compute only nodes fail to register with EAM, or are not present in the
EAM client, and not under the resource pool in vCenter, then run the command below to re-add the compute only nodes:
# stcli node add --node-ips <computeNodeIP> --controller-root-password <ctlvm-pwd> --esx-username
<esx-user> --esx-password <esx-pwd>
Step 4 (Optional) Once registration is successful, re-enable ESXi Lockdown mode if you disabled it prior to registering the
HyperFlex cluster to vCenter.
Renaming Clusters
After you create a HX Data Platform storage cluster, you can rename it without disrupting any processes.
Note These steps apply to renaming the HX Cluster, not the vCenter cluster.
Step 1 From the vSphere Web Client Navigator, select vCenter Inventory Lists > Cisco HyperFlex Systems > Cisco HX
Data Platform > cluster to rename.
Step 2 Open the Rename Cluster dialog box. Either right-click on the storage cluster or click the Actions drop-down list at the
top of the tab.
Step 3 Select Rename Cluster.
Step 4 Enter a new name for the storage cluster in the text field.
HX cluster names cannot exceed 50 characters.
Set the certMgmt mode in vCenter to Custom to add the ESXi hosts with third party certificate to vCenter.
Note By default, the certMgmt mode is vmsa. In the default vmsa mode, you can add only the ESX host with self
signed certificates. If you try to add an ESX with CA certificate to a vCenter, it will not allow you to add
the ESX host unless CA certificate is replaced with self-signed certificate.
Note The behavior of host addition in vCenter varies according to the certificate and certMgmt mode.
• When the host has self-signed certificate with the certMgmt mode set to the default value of vmsa in
vCenter:
• Only ESX host with self-signed certificate can be added.
• The addition of ESX with third party CA certificate is not allowed.
• If you try to add an ESX to a vCenter after replacing the self-signed certificate with a third party
CA certificate, the system will prompt you to replace third party CA certificate with self-signed
certificate. You can add the ESX host after replacing CA certificate with self-signed certificate.
• When the host has self-signed certificate with the certMgmt mode set to custom in vCenter:
• If you try to add an ESX to a vCenter after replacing the self-signed certificate with a third party
CA certificate, the system throws an error: ssl thumbprint mismatch and add host fails. In
this case, do the following to replace the third party CA certificate with the self-signed certificate:
1. Place the host in the maintenance mode (MM mode).
2. Replace the certified rui.crt and rui.key files with the backed up previous key and certificate.
3. Restart the hostd and vpxa service. The CA certificate comes up in the new node.
4. Right-click and connect to vCenter. The host removes the CA certificate and gets replaced with
self-signed certification in VMware.
• When the host has third party CA certificate with the certMgmt mode set to the default value of vmsa
in vCenter:
• ESX host with self-signed certificate can be added.
• The addition of ESX with third party CA certificate is not allowed.
• When the host has third party CA certificate with the certMgmt mode set to custom in vCenter:
• ESX host with self-signed certificate cannot be added.
• The self-signed certificate in ESX host needs to be replaced with a CA certificate of vCenter.
Step 1 Generate the host certificate (rui.crt) and key (rui.key) files and send the files to the certificate authority.
Note Ensure that a proper hostname or FQDN of the ESX host is provided while generating the rui.key and rui.crt
files.
Step 2 Replace the certified host certificate (rui.crt) and key (rui.key) files in the /etc/vmware/ssl directory on each ESXi host
after taking backup of the original host certificate (rui.crt) and key (rui.key) files.
Note Replace host certificate (rui.crt) and key (rui.key) files in a rolling fashion by placing only one host in
maintenance mode and then wait for the cluster to be healthy and then replace the certificates for the other
nodes.
a) Log into the ESXi host from an SSH client with administrator privileges.
b) Place the host in the maintenance mode (MM mode).
c) Take a backup of the previous key and certificate to the rui.bak file in the /etc/vmware/ssl/ directory.
d) Upload the new certified rui.crt and rui.key files to the /etc/vmware/ssl/ directory.
e) Restart the hostd and vpxa service, and check the running status using the following commands:
/etc/init.d/hostd restart
/etc/init.d/vpxa restart
/etc/init.d/hostd status
/etc/init.d/vpxa status
Note Before attempting to register the HyperFlex cluster to vCenter, you must disable ESXi Lockdown mode on
all ESXi hosts, and ensure SSH service is enabled and running. Once registration is successful, you may
re-enable Lockdown mode.
4. Restart the hostd and vpxa service using the following commands:
/etc/init.d/hostd restart
/etc/init.d/vpxa restart
Boost Mode
Boost Mode allows the Cisco HyperFlex cluster to deliver higher IOPs by increasing the storage controller
VM CPU resources by 4 vCPU. Enabling Boost Mode takes additional CPU resources from user VM for the
HX data platform, and should only be enabled in deployments where support has determined that the benefit
of additional CPUs, outweighs the impact to the sizing of your deployment. For more information about the
CPUs supported by Boost Mode, see the Cisco HyperFlex Spec Sheets.
• Cluster expansion requires you to apply Boost Mode to the new nodes.
• Boost Mode is supported in Cisco HX Release 4.0(2a) and later.
• Boost Mode should be enabled after support has determined that your deployment will benefit from the
additional CPUs.
Note CPU - number of physical cores must be equal to at least the new number of controller vCPUs. To verify the
number of physical cores in the vSphere Client; Click host > Configure > Hardware > Processors >
Processor cores per socket
Step 1 From the vCenter, right-click one controller VM and Shut Down Guest OS.
Step 2 Increase the number controller VM vCPUs to 16 for all-NVMe, or 12 for all flash C220, and all flash C240. In the vSphere
client, click Edit Settings for the VM and change the value of the CPU field in the first line.
Note Boost Mode number of controller VM vCPUs:
• All NVMe: 16
• All Flash C245: 12
• All Flash C240: 12
• All Flash C225: 12
• All Flash C220:12
Step 1 From the vCenter, right-click one controller VM and Shut Down Guest OS.
Step 2 Decrease the number controller VM vCPUs back to 12 for all-NVMe, or 8 for all flash C220, and all flash C240. In the
vSphere client, click Edit Settings for the VM and change the value of the CPU field in the first line.
Step 3 Click OK to apply the configuration changes.
Step 4 Power up the controller VM.
Step 5 Log into HX Connect and wait for the cluster to become healthy.
Step 6 Repeat the process for each host (or node) in the cluster.
servers. UI and API-based queries of each node’s secure boot status is supported so the cluster’s security
posture on-demand can be audited.
The following limitations apply to the UEFI boot mode:
• For HX Edge clusters, UEFI secure boot should only be enabled on HX Edge clusters running Cisco
IMC version 4.1(2a) and later. If secure boot is enabled on earlier Cisco IMC versions, secure boot will
need to be temporarily disabled during firmware updates.
• Support for Secure boot is available only for HyperFlex ESXi M5/M6 Server.
• Attestation of Secure Boot of ESXi hosts by vCenter is supported. This feature requires ESXi release
6.7 or later and TPM 2.0 modules on the converged or compute nodes. The TPM and TXT parameters,
which are required to enable usage of the TPM module, are automatically configured in the course of
enabling secure boot. No additional steps are required to use attestation.
• All the factory prepped M.2 RAID edge nodes run HXDP server firmware version 4.1(2a) or later. If a
customer downgrades in field or retrofits existing setup and tries to bring up a cluster with M.2RAID
nodes with HXDP server firmware version earlier than 4.1(2a), then the install may fail with the error
UEFI boot parameters cannot be configured for Legacy boot mode. The HXDP server firmware
must be upgraded to version 4.1(2a) or later and then re-try the install.
• If the Secure Boot is already enabled, Enable Secure Boot option is greyed out and no further action
needed.
• In the case that Enable Secure Boot workflow fails, then from the vCenter, confirm whether the host is
still in Maintenance Mode. If so, then exit Maintenance Mode before retrying the Enable Secure Boot
workflow.
Step 1 From the HX Connect UI, navigate to Upgrade > Select Upgrade Type
Step 2 From the Select Upgrade Type tab, select the Secure Boot mode check box.
Note After Secure Boot is enabled, it cannot be disabled.
Step 3 Enter your vCenter and UCSM credentials: Username and Admin password and click Upgrade.
After enabling Secure Boot on the cluster, any new converged or compute nodes subsequently added, automatically have
secure boot enabled. No manual intervention is required.
Step 4 To check the status of Secure Boot, navigate to System Information > Actions drop-down menu and select Check
Secure Boot Status.
Note If all nodes are enabled, the Secure Boot is enabled on all the nodes message is displayed.
Action
1. Disable boot security by unchecking the UEFI Secure Boot check box on the Compute > Configure
Boot Order page.
Catalog Update
Compatibility Catalog Update feature was introduced in Cisco HyperFlex Release 4.5(1a) for ESXi.
Catalog Update provides the ability to update the catalog version across a cluster during cluster creation,
expansion or hot adds of a newer model drives without needing to upgrade the HXDP version running on the
cluster.
• Clearly identify drives that are unsupported by the current catalog.
• Reduces the overhead when adding a new drive model to a cluster node by eliminating the need to upgrade
HXDP.
• Supported on HX Installer, HX Connect, and Intersight.
• Catalog is updated online and without impact to the running cluster.
Step 5 Upload the Catalog file saved locally. Drag and drop the file into the target or click the target to browse to the file location.
The upload operation is complete.
Step 6 Click Upgrade to complete the upgrade or Close to exit the Catalog Upgrade window.
The drive supportability check runs again after the catalog is upgraded. When all drives have compatible
catalogs a green success banner appears.
Step 5 Upload the Catalog file saved locally. Drag and drop the file into the target or click the target to browse to the file location.
The upload operation is complete.
Step 6 Click Upgrade to complete the upgrade or Close to exit the Catalog Upgrade window.
The drive supportability check runs again after the catalog is upgraded. When all drives have compatible
catalogs a green success banner appears.
To view the current catalog version after the Catalog Upgrade is complete, navigate to the upgrade page in
HX Connect for the running cluster.
Step 4 Upload the catalog file saved locally. Drag and drop the file into the target or click the target to browse to the file location.
The upload operation is complete.
Step 5 Click Upgrade to complete the upgrade or Close to exit the Catalog Upgrade window.
To view the current catalog version after the Catalog Upgrade is complete, return to the upgrade catalog
window by clicking the Settings Icon > Upgrade Catalog.
Note The cluster catalog is automatically upgraded when you upgrade HXDP version. No manual update of catalog
is needed when upgrading HXDP to a version that already contains an updated catalog.
Step 2 Check the HX Data Platform box on the Select Upgrade tab.
Note Combining a Catalog Upgrade with any other type of upgrade is not supported.
Step 3 Upload the locally saved Catalog file. Drag and drop the file into the target or click the target to browse to the file location.
The Catalog file upload operation is complete.
Step 4 Click Upgrade to complete the upgrade.
a) To monitor progress of the upgrade tasks on an HX storage cluster, click the Activity page in HX Connect.
Step 5 Click on the System Information page and verify that all disks have been claimed by HXDP and are in use.
Note Three node storage clusters. Contact Technical Assistance Center (TAC) for any task that requires removing
or shutting down a node in a three node cluster. With any 3 node storage cluster, if one node fails or is removed,
the cluster remains in an unhealthy state until a third node is added and joins the storage cluster.
Adding nodes. Nodes are added to the storage cluster through the Expand Cluster feature of the Cisco HX
Data Platform Installer. All new nodes must meet the same system requirements as when you installed the
Cisco HX Data Platform and created the initial storage cluster. For a complete list of requirements and steps
for using the Expand Cluster feature, see the appropriate Cisco HX Data Platform Install Guide.
Pre-Maintenance Tasks
Before you perform maintenance on the storage cluster, ensure the following.
• Identify the maintenance task to be performed.
• All maintenance operations such as remove/replace resources are done during maintenance windows
when the load on the system is low.
• The storage cluster is healthy and operational before the maintenance tasks.
• Identify disks using the HX Connect or HX Data Platform Plug-in Beacon options.
The HX Beacon option is not available for housekeeping 120GB SSDs. Physically check the server for
the location of the housekeeping SSD.
• Check the list of maintenance tasks that cannot be performed in parallel. See Serial vs. Parallel Operations,
on page 69 for more information on these tasks.. You can perform only some tasks serially to each other.
• Ensure that SSH is enabled on all the ESX hosts.
• Put the ESX host into HXDP Maintenance Mode prior to performing a maintenance task on the host.
The HXDP Maintenance Mode performs additional storage cluster specific steps compared to the vSphere
provided Host Maintenance Mode.
Example response that indicates the storage cluster is online and healthy:
locale: English (United States)
state: online
upgradeState: ok
healthState: healthy
state: online
state: online
Example response:
#of node failures tolerable to be > 0
Setting a Beacon
Beaconing is a method of turning on an LED to assist in locating and identifying a node (host) and a disk.
Nodes have the beacon LED in the front near the power button and in the back. Disks have the beacon LED
on the front face.
You set a node beacon through Cisco UCS Manager. You set a disk beacon through the Cisco HX Data
Platform Plug-in or HX Connect user interface.
Step 2 Turn on and off a disk beacon using the Cisco HX Data Platform Plug-in.
a) From the vSphere Web Client Navigator, select vCenter Inventory Lists > Cisco HyperFlex Systems > Cisco HX
Data Platform > cluster > Manage.
b) From Manage, select Cluster > cluster > host > Disks > disk.
c) Locate the physical location of the object and turn on the beacon.
From Actions drop-down list, select Beacon ON.
d) After you locate the disk, turn off the beacon.
From Actions drop-down list, select Beacon OFF
Remember Some VMs not supported by vMotion should be shut down, since it will hold the nodes from going into
maintenance mode.
1. Verify that the vMotion port group is configured with vmnic3 and vmnic7 in an active/standby configuration
across all of the ESXi hosts in the cluster.
2. Verify that a port group is configured for vMotion, and that the naming convention is EXACTLY the
same across all ESXi hosts in the cluster.
3. Verify that you have assigned a static IP to each vMotion port group, and that the static IPs for each
vMotion port group are in the same subnet.
4. Verify that the vMotion port group has the vMotion option checked in the properties, and that no other
port groups (such as management) have this option checked, on each ESXi host in the cluster.
5. Verify in the settings that the vMotion port group is set to 9000 MTU, (if you are using jumbo frames),
and the VLAN ID matches the network configuration for the vMotion subnet.
6. Verify you can ping from the vMotion port group on one ESXi host to the vMotion IP on the other host.
Type vmkping -I vmk2 -d -s 8972 <vMotion IP address of neighboring server>
resources to serve IO optimally. Each node is used to serve part of user IO as well as be responsible for
doing internal bookkeeping activities.
When a node leaves (entering Maintenance Mode), the in-flight IO needs to failover to other nodes in
the cluster. In addition to internal book-keeping resources and activities also need to failover. The time
required for this is proportional to data and activities which were being served by the node. This results
in additional latency for the in-flight user IO.
This is similar to the case where nodes come back from Maintenance Mode.
• See Entering Cisco HyperFlex Maintenance Mode and Exiting HXDP Maintenance Mode, on page 73
for steps.
Note After you complete any maintenance tasks, you must manually exit HXDP Maintenance Mode.
3. Log into the ESXi command line of this node as a user with root privileges.
4. Verify that the node has entered HXDP Maintenance Mode.
# esxcli system maintenanceMode get
You can monitor the progress of the Enter Maintenance Mode task in vSphere Web Client, under the
Monitor > Tasks tab.
If the operation fails, an error message displays. Try to fix the underlying problem and attempt to enter
maintenance mode again.
3. Log into the ESXi command line of this node as a user with root privileges.
4. Verify that the node has exited HXDP Maintenance Mode.
# esxcli system maintenanceMode get
You can monitor the progress of the Exit Maintenance Mode task in vSphere Web Client, under the Monitor >
Tasks tab.
If the operation fails, an error message displays. Try to fix the underlying problem and attempt to exit
maintenance mode again.
Note All IP addresses must be IPv4. HyperFlex does not support IPv6 addresses.
Name Description
Admin State field This can be one of the following:
• Enabled—Cisco UCS Manager runs the backup operation as soon as you
click OK.
• Disabled—Cisco UCS Manager does not run the backup operation when
you click OK. If you select this option, all fields in the dialog box remain
visible. However, you must manually run the backup from the Backup
Configuration dialog box.
Type field The information saved in the backup configuration file. This can be one of the
following:
• Full state—A binary file that includes a snapshot of the entire system. You
can use the file generated from this backup to restore the system during
disaster recovery. This file can restore or rebuild the configuration on the
original fabric interconnect, or recreate the configuration on a different fabric
interconnect. You cannot use this file for an import.
Note You can only use a full state backup file to restore a system
that is running the same version as the system from which the
backup file was exported.
• All configuration—An XML file that includes all system and logical
configuration settings. You can use the file generated from this backup to
import these configuration settings to the original fabric interconnect or to
a different fabric interconnect. You cannot use this file for a system restore.
This file does not include passwords for locally authenticated users.
• System configuration—An XML file that includes all system configuration
settings such as usernames, roles, and locales. You can use the file generated
from this backup to import these configuration settings to the original fabric
interconnect or to a different fabric interconnect. You cannot use this file
for a system restore.
• Logical configuration—An XML file that includes all logical configuration
settings such as service profiles, VLANs, VSANs, pools, and policies. You
can use the file generated from this backup to import these configuration
settings to the original fabric interconnect or to a different fabric interconnect.
You cannot use this file for a system restore.
Name Description
Preserve Identities check box This checkbox remains selected for All Configuration and System Configuration
type of backup operation, and provides the following functionality:
• All Configuration—The backup file preserves all identities derived from
pools, including vHBAs, WWPNs, WWNN, vNICs, MACs and UUIDs.
Also, the identities for Chassis, FEX, Rack Servers, and user labels for
Chassis, FEX, Rack Servers, IOMs and Blade Servers are preserved.
Note If this check box is not selected the identities will be reassigned
and user labels will be lost after a restore.
Location of the Backup File field Where the backup file should be saved. This can be one of the following:
• Remote File System—The backup XML file is saved to a remote server.
Cisco UCS Manager GUI displays the fields described below that allow you
to specify the protocol, host, filename, username, and password for the
remote system.
• Local File System—The backup XML file is saved locally.
Java-based Cisco UCS Manager GUI displays the Filename field with an
associated Browse button that let you specify the name and location for the
backup file.
Note Once you click OK, the location cannot be changed.
HTML-based Cisco UCS Manager GUI displays the Filename field. Enter
a name for the backup file in <filename>.xml format. The file is downloaded
and saved to a location depending on your browser settings.
Name Description
Protocol field The protocol to use when communicating with the remote server. This can be one
of the following:
• FTP
• TFTP
• SCP
• SFTP
• USB A—The USB drive inserted into fabric interconnect A.
This option is only available for certain system configurations.
• USB B—The USB drive inserted into fabric interconnect B.
This option is only available for certain system configurations.
Hostname field The hostname, IPv4 address of the location where the backup file is stored. This
can be a server, storage array, local drive, or any read/write media that the fabric
interconnect can access through the network.
Note If you use a hostname rather than an IPv4 address, you must
configure a DNS server. If the Cisco UCS domain is not registered
with Cisco UCS Central or DNS management is set to local,
configure a DNS server in Cisco UCS Manager . If the Cisco UCS
domain is registered with Cisco UCS Central and DNS management
is set to global, configure a DNS server in Cisco UCS Central.
Note All IP addresses must be IPv4. HyperFlex does not support IPv6
addresses.
Remote File field The full path to the backup configuration file. This field can contain the filename
as well as the path. If you omit the filename, the backup procedure assigns a name
to the file.
User field The username the system should use to log into the remote server. This field does
not apply if the protocol is TFTP.
Password field The password for the remote server username. This field does not apply if the
protocol is TFTP.
Cisco UCS Manager does not store this password. Therefore, you do not need to
enter this password unless you intend to enable and run the backup operation
immediately.
Step 9 (Optional) To view the progress of the backup operation, do the following:
a) If the operation does not display in the Properties area, click the operation in the Backup Operations table.
b) In the Properties area, click the down arrows on the FSM Details bar.
The FSM Details area expands and displays the operation status.
To shut down the Cisco HX storage cluster, perform the following steps:
Step 1 Gracefully shut down all workload VMs on all the Cisco HX datastores.
Alternatively, use vMotion to migrate the workload VMs to another cluster.
Note Do not shut down or move the storage controller VMs (stCtlVMs).
Note For clusters with a nested vCenter, performing an hxcli cluster shutdown may have certain limitations.
For more details, see Known Constraints with vCenter Deployment.
# hxcli cluster shutdown
b) Run the cluster information command. Confirm the storage cluster is offline.
# hxcli cluster info
In the command response text, check the cluster subsection and verify the healthstate is unknown.
This Cisco HX cluster shutdown procedure does not shut down the ESXi hosts.
If the maintenance or upgrade task does not require the physical components be powered off, exit these steps and proceed
to What to do next:
Step 3 To power off the HX storage cluster, complete Step 2 and Step 3, then complete the rest of the following steps.
Step 4 On each storage cluster ESX host, shutdown the controller VM (stCtlVM).
Choose a method:
Using vCenter Shut Down Guest OS
a) From vCenter client, locate the controller VM on each ESX host.
b) Right-click the controller VM and select Power > Shut Down Guest OS..
This method performs a graceful guest VM shutdown.
What to do next
1. Complete the task that required the storage cluster shutdown or power off. For example, an offline upgrade,
physically moving the storage cluster, or performing maintenance on nodes.
• For upgrade tasks, see the Cisco HyperFlex Systems Upgrade Guide.
• For hardware replacement tasks, see the server hardware guides.
Sometimes these tasks require that the host is shutdown. Follow the steps in the server hardware
guides for migrating VMs, entering HXDP Maintenance Mode, and powering down the servers, as
directed.
Note Most hardware maintenance tasks do not require the Cisco HX cluster is shutdown.
2. To restart the Cisco HX storage cluster, proceed to Power On and Start Up the Cisco HX Storage Cluster,
on page 80.
Step 5 If all the controller VMs are not automatically powered on, power on all the controller VMs (stCtlVM) using one of
the following methods:
Using vSphere Client
a) From the vSphere Client, view a storage controller host.
b) Right-click the stCtrlVM and select Power > Power On.
c) Repeat for each host.
Using ESXi host command line
a) Log into a host.
b) Identify the VMID of the stCtlVM.
# vim-cmd vmsvc/getallvms
b) If the command returns full storage cluster information, including build number, the storage cluster is ready to be
started. Proceed to restarting the storage cluster.
c) If the command does not return full storage cluster information, wait until all the services have started on the host.
Step 8 Start the storage cluster.
From the command line of any controller VM, run the command.
# hxcli cluster start
Depending upon the maintenance or upgrade task performed while the HX cluster was shutdown, the nodes might be
exited from HXDP Maintenance Mode or Host Maintenance Mode. Ignore any error messages about an unknown host
exception.
Step 9 Wait until the storage cluster is online and returns to a healthy state.
a) From any controller VM, run the command.
# hxcli cluster info
b) In the command response text, check the cluster subsection and verify the healthstate is online.
This could take up to 30 minutes, it could take less time depending upon the last known state.
Step 11 When the storage cluster is healthy and the datastores are remounted, power on the workload VMs.
Alternatively, use vMotion to migrate the workload VMs back to the storage cluster.
Note All IP address must be IPv4. IPv6 addresses are not supported.
Note You must have access to a Full State configuration file to perform a system restore.
You cannot perform a system restore with any other type of configuration or
backup file.
Step 4 If the system cannot access a DHCP server, enter the following information:
• IPv4 address for the management port on the fabric interconnect
• Subnet mask or prefix for the management port on the fabric interconnect
• IPv4 address for the default gateway assigned to the fabric interconnect
Step 5 Copy the web link from the prompt into a web browser and go to the Cisco UCS Manager GUI launch page.
Step 6 On the launch page, select Express Setup.
Step 7 On the Express Setup page, select Restore From Backup and click Submit.
Step 8 In the Protocol area of the Cisco UCS Manager Initial Setup page, select the protocol you want to use to upload the
full state backup file:
• SCP
• TFTP
• FTP
• SFTP
Name Description
Server IP The IPv4 address of the computer where the full state backup file is located. This
can be a server, storage array, local drive, or any read/write media that the fabric
interconnect can access through the network.
Backup File Path The file path where the full state backup file is located, including the folder names
and filename.
Note You can only use a full state backup file to restore a system that is
running the same version as the system from which the backup file
was exported.
User ID The username the system should use to log into the remote server. This field does
not apply if the protocol is TFTP.
Password The password for the remote server username. This field does not apply if the
protocol is TFTP.
Step 4 Log into vCenter and select the DirectPath I/O Configuration page.
From vCenter Client: Select the ESX host > Configuration tab > Hardware pane > Advanced Settings > Edit.
From vCenter Web Client: From the vCenter Inventory, select Resources > Hosts > ESX host > Manage > Settings >
Hardware > PCI Devices > Edit.
SED Encryption
Self-Encrypting Drives Overview
Self-Encrypting Drives (SEDs) have special hardware that encrypts incoming data and decrypts outgoing data
in real-time. The data on the disk is always stored in encrypted form. A media encryption key controls this
encryption and decryption. This key is never stored in the processor or memory.
A security key, also known as Key-Encryption Key or an authentication passphrase, is used to encrypt the
media encryption key. To enable SED, you must provide a security key. No key is required to fetch the data,
if the disk is not locked.
Cisco HyperFlex Systems enables you to configure security keys locally or remotely. When you configure
the key locally, you must remember the key. In case you forget the key, it cannot be retrieved, and the data
is lost if the drive power cycles. You can configure the key remotely by using a key management server (also
known as KMIP server). This method addresses the issues related to safe-keeping and retrieval of the keys in
the local management.
The encryption and decryption for SEDs is done through the hardware. Thus, it does not affect the overall
system performance. SEDs reduce the disk retirement and redeployment costs through instantaneous
cryptographic erasure. Cryptographic erasure is done by changing the media encryption key. When the media
encryption key of a disk is changed, the data on the disk cannot be decrypted, and is immediately rendered
unusable.
An SED based cluster can have encryption enabled and disabled at will. You are free to move between the
two states whenever you want. For more information, see the HX-Hardening Guide.
2. Select Global Inventory Lists > Cisco HyperFlex Systems > Cisco HX Data Platform > Cluster_Name
> Summary > .
3. If the HyperFlex cluster has SED drives and is encryption capable, Data At Rest Encryption-Capable
is listed at the top of the Summary tab.
Click Next.
Step 4 To secure the HyperFlex cluster using an encryption key generated and stored locally, select Local Key.
Click Next.
Click Next.
Step 4 Enter the Existing Encryption Key and the New Encryption Key for the cluster.
Note Enter exactly 32 alphanumeric characters.
Click Next.
Step 4 To disable the encryption key on the cluster, enter the encryption key in use for the cluster.
Step 5 Click Disable encryption.
Step 6 To confirm disabling the encryption key on the cluster, in the Disable encryption? dialog box, click Yes, disable
encryption.
Step 1 On the Cisco HyperFlex Connect Navigation Pane, choose System Information.
Step 2 From the Disks tab, select the disk from which you want to securely erase the local key.
Step 3 Click the Secure erase button.
Step 4 To securely erase the encrypted disk on the cluster, enter the encryption key in use on the cluster.
Step 5 Click Secure erase.
Step 6 In the Erase this disk? dialog box, click Yes, erase this disk to securely erase the encrypted disk.
Click Next.
Step 4 To secure the HyperFlex cluster using a remote security key generated by the key management (KMIP) server, select
Key Management Server.
You can configure a server with Self-Encrypting Drives in the cluster to use one of the following certificates.
• Use certificate authority signed certificates—Generate Certificate Signing Requests (CSRs) signed by an external
certificate authority.
• Use self-signed certificates—Generate self-signed certificates.
Click Next.
Step 5
What to do next
You can generate certificate signing requests or self-signed certificates.
Click Next.
Step 4 Select Key Management Server > Use certificate authority signed certificates.
Click Next.
Step 5 To generate the remote encryption key for configuring the key management (KMIP) server, complete the following
details.
Locality field The city or town in which the company requesting the certificate is
headquartered.
Enter up to 32 characters.
State field The state or province in which the company requesting the certificate is
headquartered.
Enter up to 32 characters.
Step 6 To generate Certificate Signing Requests (CSRs) for all the HyperFlex nodes and download them, click Generate
certificates.
Step 7 Download the certificates to get them signed by a certificate authority. Click Close.
What to do next
1. Upload the signed certificates.
2. Configure KMIP server (key management server).
Click Next.
Primary key management server Enter the primary Key Management Server IP address.
field
(Optional) Secondary key If you have a secondary key management server set up for redundancy, enter
management server field the details here.
Port number field Enter the port number you wish to use for the key management servers.
Public key field Enter the public root certificate of the certificate authority that you generated
during KMIP server configuration.
Step 9 Click Save to encrypt the cluster with remotely managed keys.
Click Next.
Step 5 To generate the remote encryption key for configuring the key management (KMIP) server, complete the following
details.
Locality field The city or town in which the company requesting the certificate is
headquartered.
Enter up to 32 characters.
State field The state or province in which the company requesting the certificate is
headquartered.
Enter up to 32 characters.
Step 6 To generate self-signed certificates for all the HyperFlex nodes and download them, click Generate certificates.
Step 7 Upload the signed certificates and configure KMIP server (key management server).
Step 3 From the Edit configuration drop-down list, select Manage certificates.
Step 4 Enter the following Cisco UCS Manager credentials, to set up a primary key management (KMIP) server and optionally
a secondary KMIP server.
Click Next.
Step 5 Enter the primary and secondary key management (KMIP) server credentials.
Primary key management server Enter the primary Key Management Server IP address.
field
(Optional) Secondary key If you have a secondary key management server set up for redundancy, enter
management server field the details here.
Port number field Enter the port number you wish to use for the key management servers.
Public key field Enter the public root certificate of the certificate authority that you generated
during KMIP server configuration.
Step 6 Click Save to encrypt the cluster with remotely managed keys.
Restart Encryption
Enter Cisco UCS Manager credentials to restart configuring the key management server or local key, for securely encrypting
the HyperFlex cluster.
2. Log into the management CIP to install the package Run the command priv
on all the controller VMs in the cluster. install-package.
Note If your cluster has VMware EVC enabled, make sure that the EVC baseline supports nodes with Advanced
Encryption Standards New Instructions (AES-NI). If your current EVC baseline does not support AES-NI,
change the EVC settings before enabling Software Encryption.
• AES-NI enablement is required to install HyperFlex Software Encryption packages on the HX Cluster.
• HyperFlex Software Encryption can not be enabled on existing Datastores.
• HyperFlex Software Encryption can only be enabled on newly created Datastores.
• Once HyperFlex Software Encryption is enabled for a cluster/datastore, it cannot be disabled for the
cluster or datastore.
• Once HyperFlex Software Encryption is enabled for a cluster, administrators can create either encrypted
or non-encrypted datastores.
Note The HyperFlex Software Encryption package is licensed by its own software PID, which is in addition to the
HyperFlex Data Platform and Intersight software licenses. For more information, refer to the Cisco HyperFlex
Systems Ordering and Licensing Guide.
Step 1 SFTP the encryption package to the HyperFlex cip node (the node holding the cluster management IP). Use the admin
account for the username/password and a file transfer application, such as winscp. This should upload the package
to the /tmp or /home/admin directory.
Step 2 To install the package on all available nodes of a cluster, SSH to the cip node and use the priv install-package --cluster
option.
Example:
priv install-package --cluster --path /tmp/storfs-se-core_<latest version>_x86_64.deb
Note Make sure that all nodes are up and not in maintenance mode when using the --cluster option to install the
encryption package.
What to do next
Go to Intersight HyperFlex Software Encryption to enable encryption on your cluster
Note The HyperFlex Software Encryption package is licensed by its own software PID, which is in addition to the
HyperFlex Data Platform and Intersight software licenses. For more information, refer to the Cisco HyperFlex
Systems Ordering and Licensing Guide.
Step 1 SFTP the encryption package to each HyperFlex node. Use the admin account for the username/password and a
file transfer application, such as winscp. This should upload the package to the /tmp or /home/admin directory.
Step 2 For clusters with more than 12 nodes, SSH to each node and use the priv install-package --local option.
Example:
priv install-package --local --path /home/admin/<package-filename>
Note Do not shut down the cluster before proceeding to the next step, enabling HyperFlex Software Encryption.
If you shut down the cluster and restart it, you will need to re-install the encryption package.
What to do next
Go to Intersight HyperFlex Software Encryption to enable encryption on your cluster.
Note It is recommended to backup DEK after HX Software Encryption is enabled and after every Rekey. A preivously
backed-up DEK cannot be restored after you perform a Rekey on your cluster.
To restore encrypted DEK configuration to the cluster when lost or corrupted from a previously saved back-up,
contact TAC.
Step 1 Run the hxcli encryption backup-keys -f <path to file name> command.
Note Filename path should start with /home/admin/.
Limitations:
• Boot Disk/Housekeeping disks are not allowed to be secure erased.
• Once a disk is secure erased, the disk can not be re-introduced in the same cluster.
• secure disk erase is not supported on SED drives.
• When secure disk erase is in-progress, you cannot perform erase on the same disk until it is completed.
Step 1 Run the secure_disk_erase command and specify the absolute path of the target disk.
Example:
-d DISK_PATH, --disk-path DISK_PATH
THIS UTILITY WILL IRRECOVERABLY ERASE DATA FROM DRIVE.PROCEED WITH CAUTION.
All data (including storfs) from the disk /dev/sdh will be destroyed, proceed [Y/N]:y
Successfully removed the disk from the system: '/dev/sdh'
Starting erase operation for disk '/dev/sdh'
SEAGATE ST1200MM0009 CN03 peripheral_type: disk [0x0]
<< supports protection information>>
Unit serial number: WFK25FY70000C917H4GQ
LU name: 5000c500a762ca2b
Step 4 After the erase process is completed, physically remove the erased drive from the node.
Managing Datastores
Datastores are logical containers used by the HX Data Platform to manage your storage usage and storage
resources. Datastores are where the host places virtual disk files and other VM files. Datastores hide the
specifics of physical storage devices and provide a uniform model for storing VM files.
You can add, refresh the list, edit name and size, delete, mount and unmount datastores from either the HX
Connect UI or the HX Data Platform Plug-in UI. You can only rename an unpaired datastore that is unmounted.
Renaming the HX Datastore from the vCenter administration interface is not supported, and should not be
done.
Before starting, review the following support notes:
Important • Do not rename an HX datastore from vCenter. The datastore names shown in HX Connect or Intersight
and in the ESXi host datastore (that appears in vCenter) must be identical including case sensitive. If
they are not identical, some operations such as expansion, mount/unmount of a datastore will be impacted.
• Enabling encryption on your cluster is only possible during the datastore creation procedure. Encryption
cannot be disabled for a datastore once enabled.
• HX Native Snapshots are not supported with multiple datastores.
• If using an M5/M6 node, you can use any left over space in the HyperFlex NFS or local Springpath
datastore for these purposes.
• When VMs have flat vmdk files, one with thin provisioned and one with thick provisioned, the total
combined storage usage of all flat VMDK files as reported by the vCenter/ESXi and HX Connect could
be higher than the Datastore usage itself reported by vCenter & HX Connect. This could be due to ESXi
and vCenter space reporting for each VM files ignoring the "uniqueBytes" attributes sent by underlying
NFS storage in Extended stats and attributes via VAAI APIs.
• For Vmware ESXi environments, ensure Storage I/O is disabled for all HyperFlex datastores in the
vCenter. This setting is on a per datastore setting, and enabling this can cause unexpected performance
impacts.
Adding Datastores
Datastores are logical containers, similar to file systems, that hide specifics of physical storage and provide
a uniform model for storing VM files. You can also use datastores to store ISO images and VM templates.
• From the vSphere Web Client Navigator, select vCenter Inventory Lists > Cisco HyperFlex Systems > Cisco
HX Data Platform > cluster > Manage > Datastores.
• From HX Connect, select Datastores.
Editing Datastores
A HX Data Platform datastore can be modified using the edit (pencil) option. Edit options are: 1. Change the
datastore name, or 2. Change the datastore storage allocation. That is, the size of the datastore.
Note Starting with HX Release 5.0(2a), decreasing the size of an existing datastore is not supported. When you
attempt to reduce the size of a datastore in 5.0(2a) or later release, the following error appears: Reducing
datastore size is not allowed to prevent data loss. If the datastore is new, you can delete and recreate it with
the correct size.
Unmounting Datastores
Prepare to unmount a datastore.
• No VM, template, snapshot, or CD/DVD image resides on the datastore. This is the most common error
while unmounting.
• Storage I/O Control is disabled for the datastore.
• The datastore is not used for vSphere HA heartbeat.
• The datastore is not used to host RDM metadata files. RDM is not supported.
• The datastore is not used as a scratch location.
Unmount a datastore.
Deleting Datastores
Prepare to delete the datastores.
Delete datastores.
If the new Datastore does not appear in the list, click the Refresh arrow and recheck the list.
Step 1 Depending upon the task you are attempting, complete the items in Prepare to mount a datastore, Prepare to unmount a
datastore, or Prepare to delete the datastores.
Step 2 Retry to mount, unmount, or delete the datastore through the HX Connect or HX Data Platform Plug-in UI or CLI again.
Step 3 If the datastore is not in the desire mount, unmount, or deleted state, complete the following.
a) Ensure VMs are not running on the datastore.
b) From ESX host, check to see if the HX Data Platform datastore is being used by VMware service, storageRM.
# ls -ltra /vmfs/volumes/stfs-ds1/ | grep -i iorm
Sample response
Sample response
storageRM is running
Sample response
Note When performing a hot-plug pull and replace on multiple drives from different vendors or of different types,
pause for a few moments (30 seconds) between each action. Pull, pause for about 30 seconds and replace a
drive, pause for 30 seconds. Then, pull, pause for 30 seconds and replace the next drive.
Sometimes, when a disk is removed it continues to be listed in cluster summary information. To refresh this,
restart the HX cluster.
Note Removing a functional drive from one HX cluster and installing it into another HX cluster is not supported.
Disk Requirements
The disk requirements vary between converged nodes and compute-only nodes. To increase the available
CPU and memory capacity, you can expand the existing cluster with compute-only nodes as needed. These
compute-only nodes provide no increase to storage performance or storage capacity.
Alternatively, adding converged nodes increase storage performance and storage capacity alongside CPU and
memory resources.
Servers with only Solid-State Disks (SSDs) are All-Flash servers. Servers with both SSDs and Hard Disk
Drives (HDDs) are hybrid servers.
The following applies to all the disks in a HyperFlex cluster:
• All the disks in the storage cluster must have the same amount of storage capacity. All the nodes in the
storage cluster must have the same number of disks.
• All SSDs must support TRIM and have TRIM enabled.
• All HDDs can be either SATA or SAS type. All SAS disks in the storage cluster must be in a pass-through
mode.
• Disk partitions must be removed from SSDs and HDDs. Disks with partitions are ignored and not added
to your HX storage cluster.
• Moving operational disks between servers within same cluster or moving them into expansion nodes
within the same active cluster is not supported.
• Optionally, you can remove or backup existing data on disks. All existing data on a provided disk is
overwritten.
Note New factory servers are shipped with appropriate disk partition settings. Do not
remove disk partitions from new factory servers.
In addition to the disks listed in the table below, all M5/M6 converged nodes have M.2 SATA SSD with ESXi
installed.
Note Do not mix storage disks type or storage size on a server or across the storage cluster. Mixing storage disk
types is not supported.
• When replacing cache or persistent disks, always use the same type and size as the original disk.
• Do not mix any of the persistent drives. Use all HDD or SSD and the same size drives in a server.
• Do not mix hybrid and All-Flash cache drive types. Use the hybrid cache device on hybrid servers and
All-Flash cache devices on All-Flash servers.
• Do not mix encrypted and non-encrypted drive types. Use SED hybrid or SED All-Flash drives. On SED
servers, both the cache and persistent drives must be SED type.
• All nodes must use same size and quantity of SSDs. Do not mix SSD types.
Please refer to the corresponding server model spec sheet for details of drives capacities and number of drives
supported on the different servers.
For information on compatible PIDs when performing an expansion of existing cluster, please refer to the
Cisco HyperFlex Drive Compatibility document.
Compute-Only Nodes
The following table lists the supported compute-only node configurations for compute-only functions. Storage
on compute-only nodes is not included in the cache or capacity of storage clusters.
Note When adding compute nodes to your HyperFlex cluster, the compute-only service profile template automatically
configures it for booting from an SD card. If you are using another form of boot media, update the local disk
configuration policy. See the Cisco UCS Manager Server Management Guide for server-related policies.
Important To ensure the encrypted data remains secure, the data on the drive must be securely erased prior to removing
the SED.
2. If you are using a local key for encryption, locate the key. You will be prompted to provide it.
3. To prevent data loss, ensure the data on the disk is not the last primary copy of the data.
If needed, add disks to the servers on the cluster. Initiate or wait until a rebalance completes.
4. Complete the steps below before removing any SED.
Step 5 Select the corresponding Slot row for the disk to be removed.
Step 6 Click Secure erase. This button is available only after a disk is selected.
Step 7 If you are using a local encryption key, enter the Encryption Key in the field and click Secure erase.
If you are using a remote encryption server, no action is needed.
Step 8 Confirm deleting the data on this disk, click Yes, erase this disk.
Warning This deletes all your data from the disk.
Step 9 Wait until the Status for the selected Disk Slot changes to Ok To Remove, then physically remove the disk as directed.
What to do next
Note Do not reuse a removed drive in a different server in this, or any other, HX Cluster . If you need to reuse the
removed drive, contact TAC.
1. After securely erasing the data on the SED, proceed to the disk replacing tasks appropriate to the disk
type: SSD or hybrid.
Check the Type column for the disk type.
• Solid State (SSDs)―See Replacing SSDs, on page 110 and the hardware guide for your server.
• Rotational (hybrid drives)―See Replacing or Adding Hard Disk Drives, on page 115 and the hardware
guide for your server.
• Status―Remains Ok To Remove.
• Encryption―Changes from Enabled to Unknown.
When the SED is replaced, the new SED is automatically consumed by the HX Cluster . If encryption is
not applied, the disk is listed the same as any other consumable disk. If encryption is applied, the security
key is applied to the new disk.
• Status―Transitions from Ignored > Claimed > Available.
• Encryption―Transitions from Disabled > Enabled after the encryption key is applied.
Replacing SSDs
The procedures for replacing an SSD vary depending upon the type of SSD. Identify the failed SSD and
perform the associated steps.
Note Mixing storage disks type or size on a server or across the storage cluster is not supported.
• Use all HDD, or all 3.8 TB SSD, or all 960 GB SSD
• Use the hybrid cache device on hybrid servers and all flash cache devices on all flash servers.
• When replacing cache or persistent disks, always use the same type and size as the original disk.
Step 2 If a failed SSD is a cache or persistent SSD, proceed based on the type of disk.
• For NVMe SSDs, see Replacing NVMe SSDs, on page 111.
• For all other SSDs, follow the instructions for removing and replacing a failed SSD in the host, per the server
hardware guide.
After the cache or persistent drive is replaced, the HX Data Platform identifies the SDD and updates the storage cluster.
When disks are added to a node, the disks are immediately available for HX consumption.
Step 3 To enable the Cisco UCS Manager to include new disks in the UCS Manager > Equipment > Server > Inventory >
Storage tab, re-acknowledge the server node. This applies to cache and persistent disks.
Note Re-acknowledging a server is disruptive. Place the server into HXDP Maintenance Mode before doing so.
Step 4 If you replaced an SSD, and see a message Disk successfully scheduled for repair, it means that the disk is present, but
is still not functioning properly. Check that the disk has been added correctly per the server hardware guide procedures.
Note Mixing storage disk types or size on a server or across the storage cluster is not supported.
When replacing NVMe disks, always use the same type and size as the original disk.
Note You can not use NVMe SSDs as the capacity or housekeeping drive(s) in All-Flash
nodes.
• For M5 servers: If you are replacing an NVMe cache drive with a non-NVMe drive (or vice versa, if you
are replacing a non-NVMe cache drive with an NVMe drive), you must replace the cable with a different
SAS cable (for example, UCSC-RNVME-240M5 = HXAF240c M5 Rear NVMe cable (1) or
UCSC-RSAS-C240M5 = C240 Rear UCSC-RAID-M5 SAS cbl(1)). This is required to ensure that the
drive is discovered properly.
Note For M6 servers: you cannot replace an NVMe cache drive with a non-NVMe
cache drive, due to the placement of the slots which are in the front.
Perform a physical check. NVMe cache SSDs and housekeeping SSDs do not respond to beacon requests.
If the failed SSD is not an NVMe SSD, see the Replacing SSD section of this guide.
After the cache or persistent drive is replaced, the HX Data Platform identifies the SDD and updates the storage cluster.
When disks are added to a node, the disks are immediately available for HX consumption.
Step 4 Reboot the ESXi host. This enables ESXi to discover the NVMe SSD.
Step 5 Exit ESXi host from HXDP Maintenance Mode.
Step 6 To enable the Cisco UCS Manager to include new disks in the UCS Manager > Equipment > Server > Inventory >
Storage tab, re-acknowledge the server node. This applies to cache and persistent disks.
Note Re-acknowledging a server is disruptive. Place the server into HXDP Maintenance Mode before doing so.
Step 7 If you replaced an SSD, and see a message Disk successfully scheduled for repair, it means that the disk is present, but
is still not functioning properly. Check that the disk has been added correctly per the server hardware guide procedures.
Note This procedure applies to HXAF220c M5, HX220c M5, HXAF240c M5, HX240c M5, servers only.
Identify the failed housekeeping SSD and perform the associated steps.
Step 2 Remove the SSD and replace with a new SSD of the same supported kind and size. Follow the steps in the server
hardware guide.
The server hardware guide describes the physical steps required to replace the SSD.
Note Before performing the hardware steps, enter the node into HXDP Maintenance Mode. After performing
the hardware steps, exit the node from HXDP Maintenance Mode.
Step 3 Using SSH, log into the storage controller VM of the cip node (any other working node) and run the following command
to create bootdev partitions.
priv createBootdevPartitions --target 10.20.24.69
Sample response
hxshell:~$ priv createBootdevPartitions --target 10.20.24.69
Enter the root password:
create Bootdev Partitions initiated on 10.20.24.69
Note The target should be the storage controller VM IP of the affected node.
Sample response
...........
/dev/sdb1 63G 324M 60G 1%
/var/stv /dev/sdb2 24G 173M 23G 1% /var/zookeeper
Step 6 Identify the HX Data Platform installer package version installed on the existing storage cluster.
# hxcli cluster version
The same version must be installed on all the storage cluster nodes. Run this command on the controller VM of any
node in the storage cluster, but not the node with the new SSD.
Step 7 SFTP the HX Data Platform installer packages into the storage controller VM of the affected node using the admin
account for user name/password and a file transfer application, such as winscp. This should upload the package to
the /tmp directory. Untar the package after copying it to the /tmp directory.
# tar -zxvf storfs-packages-<version>.tgz
Step 8 Using SSH, log into the storage controller VM of the cip node (any other working) node and run the following command:
priv housekeeping-preinstall --target 10.20.24.69
Sample response:
hxshell:~$ priv housekeeping-preinstall --target 10.20.24.69
Enter root password :
Copied secure files
Note This step copies the secure files of /etc/springpath/secure/* folder from another working controller
machine into the affected node.
Step 9 Run the following command on the storage controller VM of the cip node (any other working node) to install the HX
Data Platform installer packages.
Priv housekeeping-inst-packages –target 10.20.24.69
Sample response:
hxshell:~$ priv housekeeping-inst-packages –target 10.20.24.69
Enter root password :
Installed packages successfully
Step 10 Enter the following command on the storage controller VM of the cip node (any other working node) to perform the
post-install tasks..
priv housekeeping-postinstall --target 10.20.24.69
Sample response:
hxshell:~$ priv housekeeping-postinstall --target 10.20.24.69
Enter root password :
Successfully done post install tasks
Successfully installed SE core package on 10.20.24.69 (optional only when Software Encryption
is enabled on the cluster
Step 11 To confirm that the cip-monitor and stofs are in running status, run the priv service cip-monitor status and the
priv service storfs status commands.
Example:
hxshell:~$ priv service cip-monitor status
cip-monitor start/running, process 18251
Note Mixing storage disks type or size on a server or across the storage cluster is not supported.
• Use all HDD, or all 3.8 TB SSD, or all 960 GB SSD
• Use the hybrid cache device on hybrid servers and all flash cache devices on all flash servers.
• When replacing cache or persistent disks, always use the same type and size as the original disk.
Step 1 Refer to the hardware guide for your server and follow the directions for adding or replacing disks.
Step 2 Add HDDs of the same size to each node in the storage cluster.
Step 3 Add the HDDs to each node within a reasonable amount of time.
The storage starts being consumed by storage cluster immediately.
The vCenter Event log displays messages reflecting the changes to the nodes.
Note When disks are added to a node, the disks are immediately available for HX consumption although they will
not be seen in the UCSM server node inventory. This includes cache and persistent disks. To include the
disks in theEquipment > Manager > UCS > Equipment > Server > Inventory > Storage tab,
re-acknowledge the server node.
Note Re-acknowledging a server is disruptive. Place the server into HXDP Maintenance Mode before doing so.
Managing Nodes
Nodes are initially added to a storage cluster using the Create Cluster feature of the HX Data Platform Installer.
Nodes are added to an existing storage cluster using the Expand Cluster feature of the HX Data Platform
Installer. When nodes are added or removed from the storage cluster, the HX Data Platform adjusts the storage
cluster status accordingly.
• Tasks for node maintenance with a failed node.
• The ESXi or HX software needs to be reinstalled.
• A node component needs to be replaced.
• The node needs to be replaced.
• The node needs to be removed.
Note Though there are subtle differences, the terms server, host, and node are used interchangeably throughout
the HyperFlex documentation. Generally a server is a physical unit that runs software dedicated to a specific
purpose. A node is a server within a larger group, typically a software cluster or a rack of servers. Cisco
hardware documentation tends to use the term node. A host is a server that is running the virtualization and/or
HyperFlex storage software, as it is 'host' to virtual machines. VMware documentation tends to use the term
host.
Note A replication factor of three is highly recommended for all environments except HyperFlex Edge. A replication
factor of two has a lower level of availability and resiliency and should not be used in a production
environment. The risk of outage due to component or node failures should be mitigated by having active and
regular backups.
Step 2 Analyze the node failure and determine the action to take.
This frequently requires monitoring the node state through HX Connect, HX Data Platform Plug-in, vCenter, or ESXi;
checking the server beacons; and collecting and analyzing logs.
Note Removing the node must not reduce the number of available nodes below the minimum 3 nodes, as this
makes the storage cluster unhealthy. To remove a node in a 3 node cluster always requires TAC assistance.
You can remove a maximum of 2 nodes from an offline cluster.
The following tables lists the methods available to perform the associated node maintenance task.
3 1 or more TAC assisted only node Node does not need to be removed to
repair. perform repair. Includes replacing disks
on node.
4-8 1 Online or Offline node repair. Node does not need to be removed to
perform repair. Includes replacing disks
on node.
Remove Node
A non-reparable item on node fails. Disks on the removed node are not reused in the storage cluster.
5 or more 1 Online or offline replace node. Use Expand cluster to add new nodes.
All other nodes must be up and running.
Not reusing the disks.
5 or more 2 Offline replace 1 or 2 nodes. Use Expand cluster to add new nodes.
All other nodes must be up and running.
Not reusing the disks.
Replacing up to 2 nodes is supported.
Replacing 3 or more nodes requires TAC
assistance.
3 or more 1 or more TAC assisted only. TAC assisted node replacement required
to return cluster to minimum 3 nodes.
Note Reusing disks requires
assigning old node UUID to
new node. Disks UUIDs to
node UUID relationship is
fixed and cannot be
reassigned. This is a TAC
assisted task.
d) Confirm the DNS server can be queried from either the IP address or the host name.
# nslookup ip_address
# nslookup newhostname
Note In the above cases, the ESXi root password is secured as soon as installation is complete. In the event a
subsequent password change is required, the procedure outlined below may be used after installation to
manually change the root password.
As the ESXi comes up with the factory default password, you should change the password for security reasons.
To change the default ESXi root password post-installation, do the following.
Note If you have forgotten the ESXi root password, for password recovery please contact Cisco TAC.
Step 1 Log into the ESXi host service control using SSH.
Step 2 Acquire root privileges.
su -
Step 5 Enter the new password, and press Enter. Enter the password a second time for confirmation.
Note If the password entered the second time does not match, you must start over.
Step 3 Lookup the FQDN for each ESXi host in the storage cluster.
a) From the ESXi host command line.
# cat /etc/hosts
a) Check the ESXi hosts have the same DNS configuration as the controller VMs.
From vCenter, select each ESXi host then Configuration > DNS Servers.
Step 6 Locate and note the Datacenter Name and the Cluster Name.
From vCenter client or web client, scroll through to see the Datacenter Name and Cluster Name. Write them down.
They will be used in a later step.
The SSO URL is not required for HX version 1.8.1c or later. See Registering a Storage Cluster with a New vCenter
Cluster, on page 54 for additional information on reregistering a cluster.
Step 11 Enable VMware cluster HA and DRS using the post install script:
a) Log into the HX cluster IP as admin and run the command # hx_post_install .
b) Select Option 1 - "New/Existing Cluster" and input all login credentials
c) Type "y" if you want to enter a new license key
d) Type "y" to enable HA and DRS in the cluster
e) Select 'n' for all other options and exit the script.
Note When disks are removed, the disk UUIDs continue to be listed, even when not physically present. To reuse
disks on another node in the same cluster see TAC for assistance.
• Components that do not require the node be shutdown. These are hot-swappable.
• HDD data drives. Front bays
See Managing Disks for the storage cluster tasks and the hardware installation guides for the hardware
focused tasks. Both sets of tasks are required to replace this component.
• SSD cache drive. Front bay 1
See Managing Disks for the storage cluster tasks and the hardware installation guides for the hardware
focused tasks. Both sets of tasks are required to replace this component.
• Fan Modules
See the hardware installation guides to replace this component.
• Power Supplies
See the hardware installation guides to replace this component.
• Components that do required the node be put into maintenance mode and shutdown.
For all of the following components, see the hardware installation guides.
• Housekeeping SSD
Both the storage cluster tasks, and hardware focused tasks are required to replace this component.
• RTC Battery on motherboard
Note The motherboard itself is not a replaceable component. You must purchase a
battery from your local hardware store and replace it.
• DIMMS
• CPUs and Heatsinks
• Internal SD Card
• Internal USB Port
• Modular HBA Riser (HX 220c servers)
• Modular HBA Card
• PCIe Riser Assembly
• PCIe Card
• Trusted Platform Module
• mLOM Card
• RAID Controller
• Virtual Interface Card (VIC)
• Graphic Processing Unit (GPU)
Removing a Node
Removing a node is supported on the following cluster types:
Stretch No Yes
Depending upon the number of nodes in a cluster, you can remove a node when the cluster is either online or
you need to make the cluster offline. Before you do so, you must first ensure that you have completed the
required preparation steps.
The affecting context is based on the number of converged nodes. The number of compute nodes does not
affect the process to remove a node.
You can only remove 1 converged node at any time.
For clusters with 4 converged nodes, follow the offline node removal process. For clusters with 5 converged
nodes or more, follow the online node removal process.
Note If you remove a node when the cluster is offline, you cannot add the node back to the cluster.
Prior to removing a node or nodes for HyperFlex clusters with Logical Availability Zones (LAZ) configured,
LAZ must be disabled.
If LAZ is utilized in the HyperFlex cluster, then the number of remaining nodes must be in a balanced
configuration that supports LAZ per the LAZ Guidelines and Considerations prior to reenabling LAZ.
Example response that indicates the storage cluster is online and heathy:
locale: English (United States)
state: online
upgradeState: ok
healthState: healthy
state: online
state: online
Step 2 Ensure that SSH is enabled in ESX on all the nodes in the storage cluster.
Step 3 Ensure that the Distributed Resource Scheduler (DRS) is enabled.
DRS migrates only powered-on VMs. If your network has powered-off VMs, you must manually migrate them to a node
in the storage cluster that will not be removed.
Note If DRS is not available then manually move the Virtual Machines from the node.
Step 4 Make a note of zkEnsemble. It contains the data IP of controller VMs (CVM).
Example:
admin:~$ cat /etc/springpath/storfs.cfg | grep -i ense
crmZKEnsemble=10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.40:2181,
10.104.18.41:2181
If the node which was removed was ucs-308 whose CVM data IP is 10.104.18.40, then that CVMs data IP should no
longer appear after you run the above command after the node removal.
Step 5 Put the node to be removed into Maintenance mode. Choose a method: vSphere GUI, controller VM command line (CLI)
or HyperFlex Connect System Information panel:
GUI
a) Right-click each host, scroll down the list, and select Maintenance Mode > Enter Maintenance Mode.
The vSphere Maintenance Mode option is at the top of the host right-click menu. Be sure to scroll to the bottom of
the list to select Maintenance Mode.
b) In HX Connect, from the MANAGE > System Information panel's Node tab, select the node, and then click on
the button to Enter HXDP Maintenance Mode.
CLI
a) Log in to a controller VM as an admin user.
b) Run stcli cluster info and look for stNodes: section
stNodes:
----------------------------------------
type: node
id: 689324b2-b30c-c440-a08e-5b37c7e0eefe
name: ucs-305
----------------------------------------
type: node
id: 9314ac70-77aa-4345-8f35-7854f71a0d0c
name: ucs-306
----------------------------------------
type: node
id: 9e6ba2e3-4bb6-214c-8723-019fa483a308
name: ucs-307
----------------------------------------
type: node
id: 575ace48-1513-0b4f-bfe1-e6abd5ff6895
name: ucs-308
---------------------------------------
type: node
id: 875ebe8-1617-0d4c-afe1-c6aed4ff6732
name: ucs-309
Under the stNodes section, the id is listed for each node in the cluster. Find the node id or name you need to remove.
c) Move the ESX host into Maintenance mode.
# stcli node maintenanceMode (--id ID | --ip NAME) --mode enter
Step 6 Wait for 2 hours, monitor healing info in stcli cluster storage-summary. You should wait until you see "Storage cluster
is healthy." as shows in the following example:
Example:
admin:$ stcli cluster storage-summary | grep -i heali -A 8
healingInfo:
inProgress: False
resiliencyInfo:
messages:
----------------------------------------
Storage node 10.104.18.40 is unavailable.
----------------------------------------
Storage cluster is healthy.
----------------------------------------
Before the healing starts you will see following:
messages:
Space rebalanc ing in progress, 83 % completed.
inProgress: True
percentComplete: 83
estimatedCompletionTimeInSeconds: 211
resiliencyInfo:
messages:
What to do next
Proceed to Removing a Node. Choose the Online or Offline method based on the number of nodes in your
storage cluster.
Note You can remove multiple nodes in a series, as long it is done one at a time and when the cluster is healthy
between each successive node removal. You must also have followed the steps required to prepare to remove
a node. For more information, see Preparing to Remove a Node, on page 127.
Note Prior to removing a node or nodes for HyperFlex clusters with Logical Availability Zones (LAZ) configured,
LAZ must be disabled.
If LAZ is utilized in the HyperFlex cluster, then the number of remaining nodes must be in a balanced
configuration that supports LAZ per the LAZ Guidelines and Considerations prior to reenabling LAZ.
Note Do not remove the controller VM or other HX Data Platform components before you complete the steps in
this task.
Step 1 Run the stcli cluster info command and look for stNodes: section to find the node which needs to be removed.. This
information is also available when you put the node in maintenance mode.
Example:
----------------------------------------
stNodes:
type: node
id: 689324b2-b30c-c440-a08e-5b37c7e0eefe
name: ucs305
type: node
id: 9314ac70-77aa-4345-8f35-7854f71a0d0c
name: ucs306
type: node
id: 9e6ba2e3-4bb6-214c-8723-019fa483a308
name: ucs307
type: node
id: 575ace48-1513-0b4f-bfe1-e6abd5ff6895
name: ucs308
type: node
id: 875ebe8-1617-0d4c-af
name: ucs 309
The stcli node remove command to remove nodes from the 5-node cluster are:
• stcli node remove –-ip-1 ucs 308 or
• stcli node remove –-id-1 575ace48-1513-0b4f-bfe1-e6abd5ff6895
After the stcli node remove command completes successfully, the system rebalances the storage cluster until the storage
cluster state is Healthy. Do not perform any failure tests during this time. The storage cluster remains healthy.
Because the node is no longer in the storage cluster, you do not need to exit HXDP Maintenance Mode.
Note It is highly recommended that you work with TAC when removing a converged node in a storage cluster.
Do not reuse a removed converged node or its disks in the original cluster.
Note If you want to reuse a removed node in another storage cluster, contact Technical Assistance Center (TAC).
Additional steps are required to prepare the node for another storage cluster.
Step 2 Confirm that the node is removed from the storage cluster.
a) Check the storage cluster information.
# stcli cluster storage-summary
b) Check the ActiveNodes entry in the response to verify the cluster has one less node.
c) Check that the node which was removed is not part of Ensemble. For example:
Example:
admin:~$ cat /etc/springpath/storfs.cfg | grep -i ense
crmZKEnsemble=10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.40:2181,
10.104.18.41:2181
For example, if the node which was removed was ucs-308 whose CVM data IP is 10.104.18.40, then that CVMs data
IP should no longer appear after you run the above command after the node removal as seen above.
.
If there are more than 5 nodes and the removed node was part of ensemble, then the new node ip appears in the
crmZKEnsemble. For example, if the cluster initially has 7 nodes (10.104.18.37 to 10.104.18.43), and crmZKEnsemble
has 10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.40:2181, 10.104.18.41:2181, then after
removal of 10.104.18.40, crmZKEnsemble has either:
10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.42:2181, 10.104.18.41:2181, or:
10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.43:2181, 10.104.18.41:2181
Step 3 Verify that disks from the removed node no longer appear by running the hxcli disk list command:
admin:~$ hxcli disk list --no-loader
+-----------+-------------------+------+----------+---------+------------+-------------+
| NODE NAME | HYPERVSIOR STATUS | SLOT | CAPACITY | STATUS | TYPE | USAGE |
+-----------+-------------------+------+----------+---------+------------+-------------+
| ucs305 | ONLINE | 1 | 111.8 GB | Claimed | Solidstate | System |
For example, if you removed ucs-308, then its disks no longer appear.
Step 4 Remove the host from the vCenter Hosts and Cluster view.
a) Log in to vSphere Web Client Navigator. Navigate to Host in the vSphere Inventory.
b) Right-click on the host and select All vCenter Actions > Remove from Inventory. Click Yes.
Step 5 Confirm that all the node-associated datastores are removed. For example, run the following command in ESXi:
[root@ucs308:~] esxcfg-nas -l
ds4 is 169.254.226.1:ds4 from 6894152532647392862-8789011882433394379 mounted available
ds3 is 169.254.226.1:ds3 from 6894152532647392862-8789011882433394379 mounted available
ds2 is 169.254.226.1:ds2 from 6894152532647392862-8789011882433394379 mounted available
ds5 is 169.254.226.1:ds5 from 6894152532647392862-8789011882433394379 mounted available
ds1 is 169.254.226.1:ds1 from 6894152532647392862-8789011882433394379 mounted available
Note If any node-associated datastores are listed, then unmount and delete those datastores manually.
Note It is highly recommended that you work with TAC when removing a converged node in a storage cluster.
Note Do not remove the controller VM or other HX Data Platform components before you complete the steps in
this task.
Step 1 Follow the process to prepare for removing a node. For more information, see Preparing to Remove a Node, on page 127.
Step 2 (For 4-node clusters only) Prepare to shutdown, then shutdown the storage cluster.
a) Gracefully shutdown all resident VMs on all the HX datastores.
Optionally, vMotion the VMs.
b) Gracefully shutdown all VMs on non-HX datastores on HX storage cluster nodes, and unmount.
c) From any controller VM command line, issue the hxcli cluster shutdown command.
# hxcli cluster shutdown
Step 3 Run the stcli cluster info command and look for stNodes: section to find the node which needs to be removed.. This
information is also available when you put the node in maintenance mode.
Example:
----------------------------------------
type: node
id: 569c03dc-9af3-c646-8ac5-34b1f7e04b5c
name: example1
----------------------------------------
type: node
id: 0e0701a2-2452-8242-b6d4-bce8d29f8f17
name: example2
----------------------------------------
type: node
id: a2b43640-cf94-b042-a091-341358fdd3f4
name: example3
----------------------------------------
type: node
id: d2d43691-daf5-50c4-d096-941358fede374
name: example5
Step 4 Remove the desired node using the stcli node remove command.
For example:
To remove 1 node
• stcli node remove –ip-1 example5 or
• stcli node remove –id-1 d2d43691-daf5-50c4-d096-941358fede374
Response:
Successfully removed node: EntityRef(type=3, id='', name='10.10.2.4')
This command unmounts all datastores, removes from the cluster ensemble, resets the EAM for this node, stops all
services (stores, cluster management IP), and removes all firewall rules.
This command does not remove the node from vCenter. The node remains in vCenter. This command also does not
remove the installed HX Data Platform elements, such as the controller VM.
Due to the node no longer being in the storage cluster, you do not need to exit HXDP Maintenance Mode.
Note If you want to reuse a removed node in another storage cluster, contact Technical Assistance Center (TAC).
Additional steps are required to prepare the node for another storage cluster.
Step 6 Confirm that the node is removed from the storage cluster once the cluster is up.
a) Check the storage cluster information.
# stcli cluster storage-summary
b) Check the ActiveNodes entry in the response to verify the cluster has one less node.
c) Check that the node which was removed is not part of Ensemble.
Example:
admin:~$ cat /etc/springpath/storfs.cfg | grep -i ense
crmZKEnsemble=10.104.18.37:2181,10.104.18.38:2181,10.104.18.39:2181, 10.104.18.40:2181,
10.104.18.41:2181
Step 7 Verify that disks from the removed node no longer appear by running the hxcli disk list command:
admin:~$ hxcli disk list --no-loader
+-----------+-------------------+------+----------+---------+------------+-------------+
| NODE NAME | HYPERVSIOR STATUS | SLOT | CAPACITY | STATUS | TYPE | USAGE |
+-----------+-------------------+------+----------+---------+------------+-------------+
| ucs305 | ONLINE | 1 | 111.8 GB | Claimed | Solidstate | System |
| ucs305 | ONLINE | 2 | 894.3 GB | Claimed | Solidstate | Caching |
| ucs305 | ONLINE | 3 | 1.1 TB | Ignored | Rotational | Persistence |
| ucs305 | ONLINE | 4 | 1.1 TB | Ignored | Rotational | Persistence |
| ucs305 | ONLINE | 5 | 1.1 TB | Ignored | Rotational | Persistence |
| ucs305 | ONLINE | 6 | 1.1 TB | Ignored | Rotational | Persistence |
| ucs305 | ONLINE | 7 | 1.1 TB | Ignored | Rotational | Persistence |
| ucs305 | ONLINE | 8 | 0 B | Unknown | | |
| ucs306 | ONLINE | 1 | 111.8 GB | Claimed | Solidstate | System |
| ucs306 | ONLINE | 2 | 894.3 GB | Claimed | Solidstate | Caching |
| ucs306 | ONLINE | 3 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs306 | ONLINE | 4 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs306 | ONLINE | 5 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs306 | ONLINE | 6 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs306 | ONLINE | 7 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs306 | ONLINE | 8 | 0 B | Unknown | | |
| ucs307 | ONLINE | 1 | 111.8 GB | Claimed | Solidstate | System |
| ucs307 | ONLINE | 2 | 894.3 GB | Claimed | Solidstate | Caching |
| ucs307 | ONLINE | 3 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs307 | ONLINE | 4 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs307 | ONLINE | 5 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs307 | ONLINE | 6 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs307 | ONLINE | 7 | 1.1 TB | Claimed | Rotational | Persistence |
| ucs307 | ONLINE | 8 | 0 B | Unknown | | |
+-----------+-------------------+------+----------+---------+------------+-------------+
For example, if you removed ucs-308, then its disks no longer appear.
Step 8 Remove the host from the vCenter Hosts and Cluster view.
a) Log in to vSphere Web Client Navigator. Navigate to Host in the vSphere Inventory.
b) Right-click on the host and select All vCenter Actions > Remove from Inventory. Click Yes.
Step 9 Confirm that all the node-associated datastores are removed. For example, run the following command in ESXi:
[root@ucs308:~] esxcfg-nas -l
ds4 is 169.254.226.1:ds4 from 6894152532647392862-8789011882433394379 mounted available
ds3 is 169.254.226.1:ds3 from 6894152532647392862-8789011882433394379 mounted available
ds2 is 169.254.226.1:ds2 from 6894152532647392862-8789011882433394379 mounted available
ds5 is 169.254.226.1:ds5 from 6894152532647392862-8789011882433394379 mounted available
ds1 is 169.254.226.1:ds1 from 6894152532647392862-8789011882433394379 mounted available
Note If any node-associated datastores are listed, then unmount and delete those datastores manually.
Step 1 Migrate all the VMs from a compute node that needs to be removed.
Step 2 Unmount the datastore from the compute node.
Step 3 Check if the cluster is in the healthy state, by running the following command:
stcli cluster info --summary
Or
stcli node remove --ip-1
Step 6 Remove any DVS from the ESXi host in vCenter, if there is a DVS.
Step 7 Remove the ESXi host from vCenter.
Step 8 Check if the cluster is in the healthy state, by running the following command:
stcli cluster info --summary
Step 9 If compute node virtnode entry still exists in stcli cluster info output, perform the following:
Restart the stMgr management service using priv service stMgr restart on the SCVMs.
Step 10 Clear stale entries in the compute node by logging out of Cisco HX Connect and then logging into Cisco HX Connect.
Step 11 Disable and re-enable the High Availability (HA) and Distributed Resource Scheduler (DRS) services to reconfigure
the services after node removal.
Step 1 Remove the ESXi host from the vCenter Cluster inventory.
Step 2 Re-install ESXi with the same version that matches the rest of the HX Cluster.
Step 3 Delete only the UCS Service Profile from UCS Manager that was previously associated with this removed node.
Step 4 Use the HX Installer of the same version and run an Expansion Workflow.
Note Be sure to check the Clear Disk Partitions box during the expansion workflow.
Note If you have LAZ configured (enabled by default for clusters of size 8 or more), please review Logical
Availability Zones, on page 154 prior to moving ahead with expansion.
• If you have replication configured, put replication in pause mode before performing upgrade, expansion
or cluster maintenance. After the upgrade, expansion or cluster maintenance is completed, then resume
replication. Perform the pause and resume on any cluster that has replication configured to or from this
local cluster.
• If you are using RESTful APIs to perform cluster expansion, sometimes the task may take longer than
expected.
• ESXi installation is supported on M.2 SATA SSD for M5/M6 converged nodes. For compute-only nodes,
ESXi installation is supported for SAN boot, front SSD/HDD, or single M.2 SSD (using UCS-MSTOR-M2
controller). Installing ESXi on USB Flash is not supported for compute-only nodes.
HW RAID M.2 (UCS-M2-HWRAID and HX-M2-HWRAID) is a supported boot configuration starting
with HX Data Platform release 4.5(1a) and later.
• You must click on the discovered cluster to proceed with expanding a standard ESX cluster. Not doing
so results in errors.
• Use only Admin credentials for the Controller VM during expansion workflow. Using any other credentials
other than Admin may cause the expansion to fail.
• In the event you see an error about unsupported drives or catalog upgrade, see the Compatibility Catalog.
• Starting with HX Release 5.0(1b) and later, you can expand ESXi based 10/25 GbE HyperFlex Edge
clusters with 2 nodes via Intersight.
Please refer to the Intersight documentation for all requirements: Cluster Expansion Requirements.
• Starting with HX Release 5.0(2b) you can not add new nodes with 375G WL cache drives to an existing
cluster with nodes that have 1.6TB cache drives.
• Moving operational disks between servers within same cluster or moving them into expansion nodes
within the same active cluster is not supported.
Note If the Hypervisor configuration fails with the SOL logging failure message, access the installer CLI through
SSH with root and default password and configure the ESXi hypervisor. Then, run the advanced installer and
check the HX Storage Software and Expand Cluster check boxes to proceed with the ESXi installation
process.
• Upgrade the HX cluster and UCS Manager to the appropriate recommended release for your deployment.
For more information, see the Cisco HyperFlex Recommended Software Release and Requirements
Guide.
• Download and deploy the matching HX Data Platform Installer (release should be same as cluster) release
to run the expansion workflow.
Caution Failure to enable EVC at the time of the warning will require a complete shutdown
of the storage cluster and all associated VMs at a later point in time. Do not skip
this warning.
• Perform the EVC mode configuration in vCenter and then retry the validation.
• Cluster expansion will then validate a second time and then continue with the expansion.
Note If the storage cluster is in an out of space condition, when you add a new node,
the system automatically rebalances the storage cluster. This is in addition to the
rebalancing that is performed every 24 hours.
• Ensure that the node you add is of the same model (HX220 or HX240) type (Hybrid, All Flash or NVME),
and disk configuration (SED or non-SED). In addition, ensure that the number of capacity disks matches
the existing cluster nodes.
• To add a node that has a different CPU family from what is already in use in the HyperFlex cluster,
enable EVC. For more details, see the Setting up Clusters with Mixed CPUs section in the Cisco HyperFlex
Systems Installation Guide for VMware ESXi.
• Ensure that the software version on the node matches the Cisco HX Data Platform release, the ESXi
version, and the vCenter version. To identify the software version, go to the Storage Cluster Summary
tab in vCenter and check the HX Data Platform release in the top section. Upgrade if necessary.
Note If you upgraded the cluster, you must download and install a new installer VM,
that matches the current release of HXDP running on the cluster.
• Ensure that the new node has at least one valid DNS and NTP server configured.
• If you are using SSO or Auto Support, ensure that the node is configured for SSO and SMTP services.
• Allow ICMP for ping between the HX Data Platform Installer and the existing cluster management IP
address.
Note If you are using RESTful APIs to perform cluster expansion, the task may take longer than expected.
c) Read the EULA, check the I accept the terms and conditions checkbox, and click Login.
Step 2 On the Workflow page, select Cluster Expansion.
Step 3 On the Credentials page, complete the following fields.
To perform cluster creation, you can import a JSON configuration file with the required configuration data. The following
two steps are optional if importing a JSON file, otherwise you can input data into the required fields manually.
Note For a first-time installation, contact your Cisco representative to procure the factory preinstallation JSON
file.
a. Click Select a file and choose your JSON file to load the configuration. Select Use Configuration.
b. An Overwrite Imported Values dialog box displays if your imported values for Cisco UCS Manager
are different. Select Use Discovered Values.
Field Description
Field Description
vCenter Credentials
Hypervisor Credentials
Step 4 Click Continue. A Cluster Expand Configuration page is displayed. Select the HX Cluster that you want to expand.
If the HX cluster to be expanded is not found, or if loading the cluster takes time, enter the IP of the Cluster Management
Address in the Management IP Address field.
Step 5 The Server Selection page displays a list of unassociated HX servers under the Unassociated tab, and the list of
discovered servers under the Associated tab. Select the servers under the Unassociated tab to include in the HyperFlex
cluster.
If HX servers do not appear in this list, check Cisco UCS Manager and ensure that they have been discovered.
For each server you can use the Actions drop-down list to set the following:
• Launch KVM Console—Choose this option to launch the KVM Console directly from the HX Data Platform
Installer.
• Disassociate Server—Choose this option to remove a service profile from that server.
Note If there are no unassociated servers, the following error message is displayed:
No unassociated servers found. Please login to UCS Manager and ensure server ports are
enabled.
The Configure Server Ports button allows you to discover any new HX nodes. Typically, the server ports are configured
in Cisco UCS Manager before you start the configuration.
Step 7 Click Continue. The Hypervisor Configuration page appears. Complete the following fields:
Attention You can skip the completion of the fields described in this step in case of a reinstall, and if ESXi networking
has been completed.
Field Description
Subnet Mask Set the subnet mask to the appropriate level to limit and control IP addresses.
For example, 255.255.0.0.
Hypervisor Settings
Ensure to select Make IP Addresses and Hostnames Sequential, to make the IP addresses sequential.
Note You can rearrange the servers using drag and drop.
Static IP Address Input static IP addresses and hostnames for all ESXi hosts.
Field Description
Step 8 Click Continue. The IP Addresses page appears. You can add more compute or converged servers, by clicking Add
Compute Server or Add Converged Server.
Ensure to select Make IP Addresses Sequential, to make the IP addresses sequential. For the IP addresses, specify if
the network should belong to Data Network or Management Network.
For each HX node, complete the following fields for Hypervisor Management and Data IP addresses.
Field Description
Management Hypervisor Enter the static IP address that handles the Hypervisor management network
connection between the ESXi host and the storage cluster.
Management Storage Controller Enter the static IP address that handles the HX Data Platform storage controller
VM management network connection between the storage controller VM and
the storage cluster.
Data Hypervisor Enter the static IP address that handles the Hypervisor data network connection
between the ESXi host and the storage cluster.
Data Storage Controller Enter the static IP address that handles the HX Data Platform storage controller
VM data network connection between the storage controller VM and the
storage cluster.
When you enter IP addresses in the first row for Hypervisor (Management), Storage Controller VM (Management),
Hypervisor (Data), and Storage Controller VM (Data) columns, the HX Data Platform Installer applies an incremental
auto-fill to the node information for the rest of the nodes. The minimum number of nodes in the storage cluster is
three. If you have more nodes, use the Add button to provide the address information.
Note Compute-only nodes can be added only after the storage cluster is created.
Advanced Configuration
Jumbo frames Check to set the MTU size for the storage data network on the host vSwitches
and vNICs, and each storage controller VM.
Enable Jumbo Frames checkbox
The default value is 9000.
Note To set your MTU size to a value other than 9000, contact Cisco
TAC.
Disk Partitions Check to remove all existing data and partitions from all nodes added to the
storage cluster. You must backup any data that should be retained.
Clean up Disk Partitions checkbox
Important Do not select this option for factory prepared systems. The disk
partitions on factory prepared systems are properly configured.
For manually prepared servers, select this option to delete existing
data and partitions.
Step 9 Click Start. A Progress page displays the progress of various configuration tasks.
Note If the vCenter cluster has EVC enabled, the deploy process fails with a message: The host needs
to be manually added to vCenter. To successfully perform the deploy action, do the following:
• Log into the ESXi host to be added in vSphere Client.
• Power off the controller VM.
• Add the host to the vCenter cluster in vSphere Client.
• In the HX Data Platform Installer, click Retry Deploy.
Step 10 When cluster expansion is complete, click Launch HyperFlex Connect to start managing your storage cluster.
Note When you add a node to an existing storage cluster, the cluster continues to have the same HA resiliency
as the original storage cluster until auto-rebalancing takes place at the scheduled time.
Rebalancing is typically scheduled during a 24-hour period, either 2 hours after a node fails or if the storage
cluster is out of space.
Step 11 Create the required VM Network port groups and vMotion vmkernel interfaces using HyperFlex hx_post_install
script or manually to match the other nodes in the cluster.
a) SSH to HyperFlex cluster management IP.
b) Log in as the admin user.
c) Run the hx_post_install command.
d) Follow the on-screen instructions, starting with vMotion and VM network creation. The other configuration steps
are optional.
Step 12 After the new nodes are added to the storage cluster the High Availability (HA) services are reset so that HA can
recognize the added nodes.
a) Log into vCenter.
b) In the vSphere Web Client, navigate to the Host: Home > vCenter > Inventory Lists > Hosts and Clusters >
vCenter > Server > Datacenter > Cluster > Host
c) Select the new node.
d) Right-click and select Reconfigure for vSphere HA.
• Ensure that the new node uses the same configuration as the other nodes in the storage cluster. This
includes VLAN IDs and switch types (whether vSwitches), VLAN tagging with External Switch VLAN
Tagging (EST), VLAN tagging with Virtual Switch Tagging (VST), or Virtual Distributed Switch.
• Enable EVC if the new node to be added has a different CPU family than what is already used in the HX
cluster. For more details, see the Setting up Clusters with Mixed CPUs section in the Cisco HyperFlex
Systems Installation Guide for VMware ESXi.
• Ensure that the software release on the node matches the Cisco HX Data Platform release, the ESXi
release and the vCenter release. To identify the software release, go to the Storage Cluster Summary
tab in vCenter and check the HX Data Platform version in the top section. Upgrade if necessary.
• Ensure that the new node has at least one valid DNS and NTP server configured.
• If you are using SSO or Auto Support, ensure that the node is configured for SSO and SMTP services.
• Compute-only nodes are deployed with automatic detection and configuration of disk and boot policies
based on the boot hardware.
Starting with HX Data Platform release 4.5(1a) and later, compute-only nodes are deployed with automatic
detection and configuration of disk and boot policies based on the inventoried boot hardware. Users
cannot directly select the UCSM policies. Instead, the boot device is automatically determined based on
the first acceptable boot media discovered in the server. The tables below show the priority order for
M5/M6 generation servers. Reading from top to bottom, the first entry that is a match based on the
inventoried hardware are selected automatically during cluster expansion. For example, when expanding
with a B200 compute node with a single M.2 boot SSD, the second rule in the table below is a match
and used for SPT association.
If the server is booted using a mechanism not listed (such a SAN boot), the catch-all policy of anyld is
selected and administrators may subsequently modify the UCSM policies and profiles as needed to boot
the server.
Priority for M6
Priority for M5
Priority for M5
3 compute-nodes-m5-sd FlexFlash 2
4 compute-nodes-m5-ldr1 MegaRAID 2
5 compute-nodes-m5-sd FlexFlash 1
Step 3 Locate the server to ensure that the server has been added to the same FI domain as the storage cluster and is an approved
compute-only model. Check the latest Release Notes for Cisco HX Data Platform for a full list of compatible Compute-only
nodes.
Step 1 Verify that the HX Data Platform installer is installed on a node that can communicate with all the nodes in the storage
cluster and compute nodes that are being added to the storage cluster.
Step 2 If the HX Data Platform installer is not installed, see Deploy the HX Data Platform Installer.
Once install beings, you should monitor compute-only node service profile association in UCS Manager. Wait until the
server is fully associated before continuing on to install ESXi.
Step 1 Download the HX Custom Image for ESXi from the Cisco.com download site for Cisco HyperFlex. See Download
Software.
Select a networked location that can be accessed through Cisco UCS Manager.
Step 5 From the KVM console session, select Virtual Media > Map CD/DVD and mount the HX Custom Image for ESXi
image. If you do not see the Map CD/DVD option, first activate virtual devices.
a) Select Virtual Media > Activate Virtual Devices.
This opens in a pop-up window.
b) Click Accept the session > Apply.
Step 6 From the Map CD/DVD option, map to the location of the HX-Vmware.iso file.
a) Select the HX-Vmware.iso file.
b) Select Map Device.
There is a check mark indicating that the file is on a mapped location, once the process is complete. The mapped
file's full name includes the ESXi build ID.
Note If you are using RESTful APIs to perform cluster expansion, sometimes the task may take longer than expected.
Note After you add a compute-only node to an existing cluster, you must manually configure the vmk2 interface
for vmotion.
c) Read the EULA, check the I accept the terms and conditions checkbox, and click Login.
Step 2 On the Workflow page, select Cluster Expansion.
Field Description
vCenter Credentials
Hypervisor Credentials
Step 4 Click Continue. A Cluster Expand Configuration page is displayed. Select the HX Cluster that you want to expand.
If the HX cluster to be expanded is not found, or if loading the cluster takes time, enter the IP of the Cluster Management
Address in the Management IP Address field.
Step 5 (M6 Servers Only) Click Continue. A Server Selection page is displayed. On the Server Selection page, the
Associated tab lists all the HX servers that are already connected. Do not select them, on the Unassociated tab, select
the servers you wish to add to the cluster.
Step 6 Click Continue. The Hypervisor Configuration page appears. Complete the following fields:
Attention You can skip the completion of the fields described in this step in case of a reinstall, and if ESXi networking
has been completed.
Field Description
Subnet Mask Set the subnet mask to the appropriate level to limit and control IP addresses.
For example, 255.255.0.0.
Hypervisor Settings
Ensure to select Make IP Addresses and Hostnames Sequential, to make the IP addresses sequential.
Note You can rearrange the servers using drag and drop.
Static IP Address Input static IP addresses and hostnames for all ESXi hosts.
Step 7 Click Continue. An IP Addresses page is displayed. Click Add Compute-only Node to add a new node.
If you are adding more than one compute-only node, select Make IP Addresses Sequential.
Field Information
Management Hypervisor Enter the static IP address that handles the Hypervisor
management network connection between the ESXi host
and storage cluster.
Data Hypervisor Enter the static IP address that handles the Hypervisor data
network connection between the ESXi host and the storage
cluster.
Controller VM Enter the default Admin username and password that were
applied to controller VMs when they were installed on the
existing HX Cluster.
Note The name of the controller VM cannot be
changed. Use the existing cluster password.
Step 8 Click Start. A Progress page displays the progress of various configuration tasks.
Note By default no user intervention is required if you are booting from FlexFlash (SD Card). However, if you
are setting up your compute-only node to boot from a local disk, complete the following steps in Cisco
UCS Manager :
Note If the vCenter cluster has EVC enabled, the deploy process fails, The host needs to be
manually added to vCenter. To successfully perform the deploy action, do the following:
Step 1 Edit Configuration - Returns you to the Cluster Configuration page. You fix the issues listed in the validation page.
Step 2 Start Over - Allows you to reverse the settings you applied by clearing progress table entries and you are returned to
the Cluster Configuration page to restart a new deployment. See Technical Assistance Center (TAC).
Step 3 Continue - Adds the node to the storage cluster in spite of the failure generating errors. See Technical Assistance Center
(TAC).
Note Select the Continue button only if you understand the failures and are willing to accept the possibility of
unpredictable behavior.
For more information about cleaning up a node for the purposes of redeploying HyperFlex, see the HyperFlex Customer
Cleanup Guides for FI and Edge.
zones:
----------------------------------------
pNodes:
----------------------------------------
state: ready
name: 10.10.18.61
----------------------------------------
state: ready
name: 10.10.18.59
----------------------------------------
zoneId: 0000000057eebaab:0000000000000003
numNodes: 2
----------------------------------------
pNodes:
----------------------------------------
state: ready
name: 10.10.18.64
----------------------------------------
state: ready
name: 10.10.18.65
----------------------------------------
zoneId: 0000000057eebaab:0000000000000001
numNodes: 2
----------------------------------------
pNodes:
----------------------------------------
state: ready
name: 10.10.18.60
----------------------------------------
state: ready
name: 10.10.18.63
----------------------------------------
zoneId: 0000000057eebaab:0000000000000004
numNodes: 2
----------------------------------------
pNodes:
----------------------------------------
state: ready
name: 10.10.18.58
----------------------------------------
state: ready
name: 10.10.18.62
----------------------------------------
zoneId: 0000000057eebaab:0000000000000002
numNodes: 2
----------------------------------------
isClusterZoneCompliant: True
zoneType: logical
isZoneEnabled: True
numZones: 4
AboutCluster Time : 08/22/2019 2:31:39 PM PDT
Command Description
stcli cluster get-zone Gets the zone details. This option is used to check if the zone is
enabled.
Command Description
stcli cluster set-zone --zone 1 (Recommended) Enables and creates zones (default number of
zones)
stcli rebalance start
Important You must execute the rebalance start command
after you enable and create zones.
stcli cluster set-zone --zone 1 --numzones Enables zones and creates a specific number of zones.
<integer-value>
Important The number of zones can only be 3, 4, or 5.
stcli rebalance start
Important You must execute the rebalance start command
after you enable and create zones.
Sample response
d) To power off a VM. Run the command specifying the VM to power off.
# vim-cmd vmsvc/power.shutdown 1
These options will perform a relinquish action for a graceful shutdown versus a hard shutdown which is not desired.
Step 7 Locate the vSphere HA - VM Monitoring option and select the following:
• Override checkbox
• Disabled from the drop-down list
data offloads and supported for powered on VMs. VAAI also called hardware acceleration or hardware
offload APIs, are a set of APIs to enable communication between VMware vSphere ESXi hosts and
storage devices. Use HX Data Platform Ready Clones to clone VMs in seconds instead of minutes.
• Batch customization of guest VMs - Use the HX Data Platform Customization Specification to instantly
configure parameters such as IP address, host name, VM name for multiple guest VMs cloned from a
host VM.
• Automation of several steps to a one-click process - The HX Data Platform Ready Clones feature
automates the task to create each guest VM.
• VDI deployment support - Ready Clones are supported for desktop VMs on VDI deployments which
are using VMware native technology.
• Datastore access - Ready Clone work on partially mounted/accessible datastores as long as the VM
being cloned is on an accessible mountpoint.
Note For sentinel based HX snapshot, sentinel snapshots are not automatically deleted
after ready clone operation. See the HX Native Snapshots Overview, on page 14
for implications of using sentinel based HX snapshots.
Ready Clones fail for any VM that is not on a HX Data Platform datastore. This applies to Ready Clones
on a VM level, VM folder level, or resource pool level.
• VMs can have only native snapshots. Ready Clones cannot be created from VMs with snapshots that
have redo logs, (non-native snapshots).
• Use only the single vNIC customization template for Ready Clones.
• Beginning in Cisco HX Release 3.0, SSH does not need to be enabled in ESX on all the nodes in the
storage cluster.
Note If you click Ready Clones to clone a VM when the OVA deployment of that VM is in progress, you will get
an error message. You can clone a VM only after the successful VM deployment.
VM Name Prefix Enter a prefix for the guest virtual machine name.
This prefix is added to the name of each Ready Clone created.
Note The VM Name Prefix which is used to name a Ready Clone, must
contain only letters, numbers, and the hyphen (-) character. The
name must start with a letter and cannot contain only digits or
hyphen.
Starting clone number Enter a clone number for the starting clone.
Each Ready Clone must have a unique name, numbering is used to ensure a
unique element in the name.
Increment clone numbers by Enter a value using which the clone number in the guest virtual machine name
must be increased, or leave the default value 1 as is. The system appends a
number to the names of the virtual machine Ready Clones (such as clone1,
clone2, and clone3). By default, the number starts from 1. You can change this
value to any number.
Use same name for Guest Name Select this check box to use the vCenter VM inventory name as the guest host
virtual machine name.
If you uncheck this box, a text box is enabled. Enter the name you want to use
for the guest host virtual machine name.
Preview After required fields are completed, HX Data Platform lists the proposed Ready
Clones names. As you change the content in the required fields, the Clone Name
and Guest Name fields update.
Power on VMs after cloning Select this check box to turn the guest virtual machines on after the cloning
process completes.
Note Use HX Data Platform Ready Clones to create multiple clones of a VM in one click!
For example, you can create ten different clones with different static IP addresses from a Windows VM.
Step 1 From the vSphere Web Client Navigator, select Global Inventory Lists > Virtual Machines. This displays the list of
VMs in vCenter.
Step 2 Select the VM to clone, and open the Actions menu. Either right-click the VM or click the Actions menu in the VM
information portlet.
If needed, view the list of clusters and associated VMs to verify the VM is a storage cluster VM.
Step 3 Select Cisco HX Data Platform > Ready Clones to display the Ready Clones dialog box.
Step 4 Specify the following information in the Ready Clones dialog box:
Control Description
Number of clones Type the number of clones that you want to create. You can create a batch of 1 through 256
clones at a given time.
Customization Click the drop-down list and select a Customization Specification for the clone from the
Specification drop-down list (which includes the customization specifications available in vCenter).
The system filters the customization specifications for the selected host VM. For example,
if the selected host VM uses Windows OS for guest VMs, the drop-down list displays
Windows OS customization specifications.
Starting clone number Type a clone number for the starting clone.
Use same name for 'Guest Select this check box to use the vCenter VM inventory name as the guest host VM name. If
Name' you uncheck this box, a text box is displayed. Enter the name you want to use for the guest
host VM name.
The system displays the guest VM names in the Guest Name column in the dialog box.
There is a similar option in the Customization Specification itself. This HX Data Platform
Ready Clone batch customization process overrides the option that you specify in the
Customization Specification option.
Control Description
• If the Customization Specification contains a NIC or network adapter that uses a static
gateway and static subnet and the guest name resolves to a static IP address, then the
system assigns the network adapter the static IP address associated with the guest name.
It also sets the storage cluster name or host name to the guest name specified.
• If the Customization Specification contains a NIC or network adapter that obtains the
IP address using DHCP, then the systems sets only the storage cluster name or host
name to the guest name specified.
Increment clone number Type a value using which the clone number in the guest VM name must be increased, or
by leave the default value 1 as is. The system appends a number to the names of the VM clones
(such as clone1, clone2, and clone3). By default, the number starts from 1. You can change
this value to any number.
Power on VMs after Select this check box to turn the guest VMs on after the cloning process completes.
cloning
The customization specification you created is listed in the Customization Specification Manager. You can
use it to customize virtual machine guest operating systems.
Note The default administrator password is not preserved for Windows Server 2008 after customization. During
customization, the Windows Sysprep utility deletes and recreates the administrator account on Windows
Server 2008. You must reset the administrator password when the virtual machine boots the first time after
customization.
The customization specification you created is listed in the Customization Specification Manager. You can
use it to customize virtual machine guest operating systems.
Step 1 Obtain the valid DNS names and ensure that they resolve to valid IP addresses.
For example, to provision a batch of 100 Windows VMs where the guest name is userwinvm1 to userwinvm100, check
that userwinvm1 through userwinvm100 are valid IP addresses.
For additional information about VMware snapshots, see the "Overview of virtual machine snapshots in
vSphere (KB 1015180)" on the VMware Customer Connect site.
• Datastore Access - Snapshots work on partially mounted or accessible datastores as long as the VM
being snapshotted is on an accessible mount-point.
Attention Beginning with HX Release 4.5(2a) and VMware ESXi 7.0 U2 Sentinel snapshots are not applicable.
• HX native snapshots - When creating the first HX native snapshot an HX SENTINEL snapshot is created
prior to the HX native snapshot. The SENTINEL snapshot is a baseline snapshot which ensures that
subsequent snapshots are HX native snapshots. . When a SENTINEL snapshot is present, creating
additional snapshots using vSphere results in the creation of HX native snapshots.
Note There should be no VMware snapshots (non-native) present when creating native
snapshots.
• HX Snapshots compatibility with VMware VAIO - Creating HX Snapshots with VMWare VAIO
configured is not supported. Attempting to create HX snapshots will power the VM off. HX Snapshots
cannot coexist with Virtual Machines enabled with vSphere APIs for IO Filtering (VAIO). The VAIO
framework might be used by backup solutions to enable Continuous Data Protection (CDP) for Virtual
Machines. To use the backup solutions with CDP, delete any existing HX snapshots before enabling the
CDP functionality.
To determine if your product uses VMware VAIO framework, review the VMware site list of qualified
vendors in the VMware Compatibility Guide.
• Maximum Number of Stored Snapshots - VMware has a limitation of 31 snapshots per VM. The
limitation total is equal to the sum of all VMware created snapshots, the HX SENTINEL snapshot, and
HX native snapshots.
Beginning with HX Release 4.0 Snapshot operations beyond the number set in the snapshot.maxSnapshots
property of the VM fail, with the following error message: Snapshot operation cannot be performed.
• Scheduled Snapshots - Do not have overlapping snapshots scheduled on VMs and their resource pools.
Performance
• VMware vSphere Storage APIs Array Integration (VAAI) - To attain the best HX snapshot
performance and functionality, upgrade to the ESXi 7.0 U2 or later.
VMs
Release Consideration
Beginning with HX Release 4.5(2a) Consolidation time is no longer proportional to the I/O load on the
using VMware ESXi 7.0 U2 and Virtual Machine.
later
Sentinel snapshots are no longer created.
• VM Name - The VM name must be unique per vCenter for taking an HX native snapshot.
• Ready storage cluster - To enable the creation of an HX native snapshot, the storage cluster must be
healthy, have sufficient space, and be online. The datastore in which the VM resides must be accessible.
The VM must be valid and not in a transient state, such as vMotioning.
vCenter
• vMotion - vMotion is supported on VMs with HX native snapshots.
• Storage vMotion - Storage vMotion is not supported on VMs with HX native snapshots. If the VM
needs to be moved to a different datastore, delete the snapshots before running storage vMotion.
Naming
• Duplicate names - VMs or Resource Pools with duplicate names within the HX Data Platform vCenter
are not supported and cause HX native snapshots to fail. This includes parents and children within nested
resource pools and resource pools within different vCenter clusters.
• Characters in Names - Special characters are not supported. Using special characters in names results
in names appearing different than specified.
• Maximum Snapshot Name Length -80 characters.
Note When creating a new virtual machine disk on the HyperFlex datastore and you
want to enable the creation of thick-provisioned disks, there is no option to create
a thick-provisioned disk. This is a known issue in VMware. For more information,
see Creating VMDK with NFS-backed storage does not allow thick-provisioning
with vendor plugin.
Note ESXi cannot distinguish between thick provision lazy zeroed and thick provision
eager zeroed virtual disks on NFS datastores. When you use NFS datastores, the
vSphere client allows you to create virtual disks in Thick Provision Lazy Zeroed
(zeroedthick) or Thick Provision Eager Zeroed (eagerzeroedthick) format.
However, when you check the disk type on the Virtual Machine Properties dialog
box, the Disk Provisioning section always shows Thick Provision Eager Zeroed
as the disk format (no matter which format you selected during the disk creation).
• Virtual Disk Types - VMware supports a variety of virtual disk backing types. The most common is
the FlatVer2 format. HX native snapshots are supported for this format.
There are other virtual disk formats such as Raw Device Mapping (RDM), SeSparse, VmfsSparse (Redlog
format). VMs containing virtual disks of these formats are not supported with HX native snapshots.
Login Access
• SSH - Ensure that SSH is enabled in ESX on all the nodes in the storage cluster.
Limitations
• If vCenter running on a VM in the storage cluster, do not take an HX native snapshot of the vCenter VM.
For additional informaion see the VMware VirtualCenter Server service fails due to a quiesced snapshot
operation on the vCenter Server database virtual machine (2003674) article on the VMware site.
Important Always use the HX native snapshots feature to create the first snapshot of a VM. This ensures that all
subsequent snapshots are in native format.
• Do not use the VMware snapshot feature to create the first snapshot of a VM. VMware snapshots use
redo log technology that result in degraded performance of the original VM. This performance degrades
further with each additional snapshot.
• If there are redo log snapshots that should be deleted, edit the /etc/vmware/config file and set
snapshot.asyncConsolidate="TRUE" on the ESXi host(s) where the redo log snapshots reside.
• HX native snapshots do not impact VM performance after the initial HX SENTINEL and HX native
snapshot are created.
• Add all the VMDKs to the VM prior to creating the first HX native snapshot.
When VMDKs are added to a VM, additional HX SENTINEL snapshots are taken. Each additional HX
SENTINEL consumes additional space.
For example, when adding 2 new VMDKs to an existing VM that has HX native snapshots, at the next
scheduled HX native snapshot, 1 new HX SENTINEL is created. If it is necessary to add one or more
additional VMDKs, review any existing HX native snapshot schedule retention plans and make sure that
the total number of retained HX native snapshots plus any HX SENTINEL snapshots will not attempt
to exceed a total value of 31.
The HX storage controller VM time is used to set the schedule. The vSphere UTC time is used to create the
HX native snapshots. The logs and timestamps vary depending upon the method used to view them.
When a schedule is created using the HX vSphere client (HTML5) plug-in, the scheduled times are converted
to UTC from the HX storage controller VM time zone. When you view the schedule through the vSphere
client (HTML5) Scheduled Tasks it displays the tasks in browser time zone.
When converted to the same timezone, they translate to the same time. For example: 5:30pm PST, 8:30PM
EST, 1:30AM UTC are all the same time.
To have vSphere Scheduled Tasks tab display the same time for a scheduled snapshot that you create in the
HX vSphere client (HTML5) plug-in, set the storage controller VM to UTC.
To have scheduled snapshots run based on local time zone settings, set that time zone for the storage cluster.
By default, the storage controller VM uses the UTC time zone set during HX Data Platform installation.
If the vSphere and the storage controller VM are not using the same time zone, the vSphere scheduled tasks
tab might display a different time than the schedule in the HX vSphere client (HTML5) plug-in schedule
snapshot dialog.
When you configure an hourly snapshot, the snapshot schedule runs between a specific start time and end
time. The vSphere Task window might display a status that a scheduled snapshot was completed outside the
hourly end time based on the timezone
Identify and set the time zone used by the storage controller VM
1. From the storage controller VM command line, view the configured time zone.
$ hxcli services timezone show
Related Topics
Schedule Snapshot, on page 349
Step 1 From the vSphere client (HTML5) Navigator display the list of VMs in vCenter at the VM level. Display the VM list
with one of the following methods, Hosts and Clusters, VMs and Templates , Storage, Networking, or Global
Inventory Lists
Example:
Global Inventory Lists > VMs
Step 2 Select a storage cluster VM and open the Actions menu. Either right-click the VM or click the Actions menu in the VM
information portlet.
Note Ensure there are no non-HX Data Platform datastores on the storage cluster resource pool or the snapshot
will fail.
Step 3 Select Cisco HX Data Platform > Take Snapshot to display the Snapshot dialog box.
Step 4 Complete the dialog box
Name Type the Snapshot name. Maximum snapshot name length: 80 characters.
Snapshot option check box Use the check boxes to select Snapshot the virtual machine's memory or
Quiesce guest file system (Needs VMware Tools installed)
Related Topics
Snapshot Now, on page 346
Note If you re-register the vCenter cluster, your HX native snapshot schedules are lost. If this happens, reconfigure
HX native snapshot schedules.
When scheduling an HX native snapshots consider your back up requirements. For critical data, retain more
frequent HX native snapshots. If there is a disaster, it is possible to restore recent HX native snapshots or
create a custom real-time HX native snapshot. For less critical data, considering creating less frequent HX
native snapshots.
HX native snapshot scheduling helps control backup costs. For each VM in a storage cluster, you can schedule
hourly, daily, or weekly snapshots. The maximum frequency for any specific VM is once per hour. Hourly
settings are available in 15 minute increments.
For example, HX native snapshots are taken each day, given the following settings.
VM 1 hourly snapshots to run at hour:15 minutes, between 10 PM and 1 AM.
VM 2 hourly snapshots to run at hour:30 minutes, between 8 PM and 12 AM.
VM 3 and 4 hourly snapshots to run at hour:45, between 6 AM and 8 AM.
VM 5 daily snapshot to run at 6:00 AM
Each day these HX native snapshots are taken. Notice that the last HX native snapshot is before the ending
hour:00.
6:00 AM — VM 5
6:45 AM — VM 3, VM 4
7:45 AM — VM 3, VM 4
8:30 PM — VM2
9:30 PM — VM2
10:15 PM — VM1
10:30 PM — VM2
11:15 PM — VM1
11:30 PM — VM2
12:15 AM — VM1
To schedule an HX native snapshot every hour over 24 hours:
Related Topics
Schedule Snapshot, on page 349
Step 2 Select a storage cluster VM or resource pool and open the Actions menu.
Either right-click the object or click the Actions menu.
Step 3 From the Actions menu, select Cisco HX Data Platform > Schedule Snapshot to display the Schedule Snapshot dialog
box.
Step 4 Select the snapshot frequency.
Click the boxes for hourly, daily, and/or weekly frequency and set the starting days, times, and duration.
Step 1 From the Schedule Snapshot dialog box, select the Enable Hourly Snapshot, Enable Daily Snapshot, or Enable Weekly
Snapshot check box.
Step 2 Click the Start at drop-down list to select a start time. Select hour, minutes in 15 minute increments, and AM or PM.
Step 3 For an hourly snapshot schedule, click the Until drop-down list to select an end time. Select hour, minutes in 15 minute
increments, and AM or PM. Set the minute to the same value as the Start at time.
The HX Data Platform plug-in creates a snapshot of the VM every hour between the start and end times.
Step 4 Select the corresponding check box to specify Days of the week on which you want to take the snapshots.
Step 5 Under Retention, either type a number or use the arrow button to specify the maximum number of copies to retain for
each schedule.
Related Topics
Schedule Snapshot, on page 349
Step 2 Select a storage cluster VM or resource pool and open the Actions menu.
Either right-click the object or click the Actions menu.
Step 3 From the Actions menu, select Cisco HX Data Platform > Schedule Snapshot to display the Schedule HX Native
Snapshot dialog box.
Step 4 Uncheck the scheduled options that are no longer required.
Step 5 Click OK to accept the changes, this includes deleting previously configured schedules, and exit the dialog.
Step 6 Confirm the schedule is deleted.
Select a storage cluster VM or resource pool. Click the HX vCenter tabs, Manage > Scheduled Tasks. The previous HX
native snapshot schedule should not be listed.
Step 1 From the vSphere client (HTML5), select the VM level, VM folder level, or resource pool level. For example, vCenter
Inventory Lists > Virtual Machines to display the list of VMs in vCenter.
Step 2 Select a storage cluster VM and open the Actions menu. Either right-click the VM or click the Actions menu in the VM
information portlet.
Step 3 Select Snapshots > Manage Snapshots to open the vSphere Snapshot Manager.
Step 4 Select a snapshot to revert to from the hierarchy of snapshots for the selected VM.
Step 5 Click Revert to > Yes > Close.
The reverted VM is included in the list of VMs and powered off. In selected cases, a VM reverted from a VM snapshot
is already powered on. See the following table for more details.
Powered on (includes memory) Reverts to the HX VM snapshot, and the VM is powered on and running.
Powered on (does not include memory) Reverts to the HX VM snapshot, and the VM is powered off.
Powered off (does not include memory) Reverts to the HX VM snapshot, and the VM is powered off.
Step 6 If the reverted VM is powered off, then select the VM and power it on.
Step 1 From the vSphere client (HTML5), select VMs and Templates > vcenter_server > Snapshots > datacenter > vm.
Step 2 Right-click the vm and select Snapshots > Manage Snapshots.
Step 3 Select an HX native snapshot and click Delete.
Note Delete the HX SENTINEL snapshot by using Delete All option only. Do not delete the HX SENTINEL
snapshot individually.
Note Only the most recently replicated DP snapshot is retained on the destination cluster. Retaining additional DP
snapshots is not supported.
Each VM is individually protected by assigning it protection attributes, including the replication interval
(schedule). The shorter the replication interval, the fresher the replicated snapshot data is likely to be. DP
snapshot intervals can range from once every 5 minutes to once every 24 hours.
A protection group is a group of VMs that have a common DP snapshot schedule , quiescence parameter
value, and a common start time.
Setting up DP snapshots requires two existing clusters running a currently supported HX Data Platform
Software Release. Both clusters must be on the same HX Data Platform version. Use HyperFlex Connect to
complete the setup.
First, set up a local replication network for each cluster. Use HX Connect to provide a set of IP addresses to
be used by local cluster nodes to replicate to the remote cluster. HX Connect creates VLANs through UCS
Manager, for dedicated local replication network use.
Note When this option is chosen in HX Connect, UCSM is configured only when both UCS Manager and fabric
interconnect are associated with the HyperFlex cluster. When UCSM and FI are not present, you must enter
the VLAN ID, and not select UCSM configuration in HX Connect.
The two clusters, and their corresponding existing relevant datastores must be explicitly paired. The pairing
setup can be completed using HX Connect from one of the two clusters. This requires administrative credentials
of the other cluster.
Virtual machines can be protected (or have their existing protection attributes modified) by using HX Connect
at the cluster where they are currently active.
HX Connect can monitor the status of both incoming and outgoing replication activity on a cluster.
After a disaster, a protected VM can be recovered and run on the cluster that serves as the DP snapshot recovery
site for that VM.
Note Documentation for the N:1 DR for HyperFlex feature is located in the Intersight Help Center. The URL is
https://www.intersight.com/help/saas/resources/replication_for_cisco_hyperFlex_clusters.
The following is a list of requirements and considerations when configuring virtual machine replication and
performing disaster recovery of virtual machines:
Networking Requirements
The replication network should be reliable and have a sustained minimum symmetric bandwidth that is the
same as the bandwidth configured in the HyperFlex replication network. Do not share the network with any
other applications or traffic on an uplink or downlink. Other requirements are as follows:
Requirement Description
Requirement Description
Adaptive Bandwidth Control Replication network variability may cause network bandwidth to vary and may
include the introduction of network errors. Adaptive Bandwidth Control for
replication will dynamically adjust the replication speed to scale down if errors
are detected and scale up to the configured replication bandwidth limit when
the errors are cleared.
Note Adaptive Bandwidth Control is only in effect when the replication
network bandwidth limit is enabled and configured as a non-zero
number. Adaptive Bandwidth Control is disabled when the
replication network bandwidth limit is not enabled (default).
Enabling the replication bandwidth limit requires entering a
bandwidth value in the range of 10 to 100,000 Mbit/s. It is
recommended that the Administrator always configure a
replication network bandwidth limit on both clusters, rather than
use the default setting.
Measuring Available • Install the iperf utility on both user VMs by running the command:
Replication Network
apt get install iperf
Bandwidth
• Run the iperf server on the user VM deployed on site B by running the
command:
iperf -s
Example Output:
------------------------------------------------------------
Server listening on TCP port 5001
TCP window size: 85.3 KByte (default)
------------------------------------------------------------
Requirement Description
Measuring Available • Run the following iperf command on the user VM on site A
Replication Network
iperf -c <server ip> -i <interval in secs> -t <time in seconds>
Bandwidth (continued)
Example Output:
Client connecting to a.b.c.d TCP port 5001
TCP window size: 85.0 KByte (default)
------------------------------------------------------------
local w.x.y.z port 47642 connected with a.b.c.d port 5001
0.0-10.0 sec 44.8 MBytes 37.6 Mbits/sec
10.0-20.0 sec 222 MBytes 187 Mbits/sec
20.0-30.0 sec 312 MBytes 261 Mbits/sec
30.0-40.0 sec 311 MBytes 261 Mbits/sec
40.0-50.0 sec 312 MBytes 262 Mbits/sec
50.0-60.0 sec 311 MBytes 261 Mbits/sec
60.0-70.0 sec 312 MBytes 262 Mbits/sec
70.0-80.0 sec 312 MBytes 262 Mbits/sec
80.0-90.0 sec 311 MBytes 261 Mbits/sec
90.0-100.0 sec 312 MBytes 262 Mbits/sec
100.0-110.0 sec 311 MBytes 261 Mbits/sec
Note Conduct testing in both directions, from the first paired cluster
to the second paired cluster, and then from the second paired
cluster to the first paired cluster. If this is a shared link with other
applications, perform testing at the time the replication schedules
are planned to run. When this link is shared, the available
bandwidth for replication could be impacted and may result in
congestion on the replication network which may result in packet
drops. The HyperFlex replication engine monitors packet drops
and throttles the replication traffic if required.
Maximum Latency The maximum replication network latency supported is 75 ms between two
paired clusters. There are conditions where it is possible that some replication
jobs will encounter error conditions and fail. For example, this may occur when
multiple replication jobs execute simultaneously with low network bandwidth
and high latency. If this situation occurs, increase the replication network
bandwidth, or reduce job concurrency by staggering the number of concurrent
replication jobs. If this situation persists, VM unprotect operations may take
longer than expected.
Measuring Replication Network Latency
You can measure the average replication network latency by running a ping
command on any of the storage controller VMs on site A and site B.
From site A, execute ping command as performed in the following example:
ping -I eth2 "Repl IP of any ctlvm on site B" -c 100
Example Output:
100 packets transmitted, 100 received, 0% packet loss, time 101243ms
rtt min/avg/max/mdev = 0.112/0.152/0.275/0.031 ms
Requirement Description
Network Ports The comprehensive list of ports required for HyperFlex component
communication is located in appendix A of the HX Data Platform Security
Hardening Guide. The port/protocl requirements (as of Version 4.5.2a rev 3
dated September 2021) for HyperFlex replication are: ICMP, TCP: 9338, 9339,
3049, 4049, 4059, 9098, 8889, and 9350.
Testing Network Ports
Internal to the HyperFlex cluster, firewall entries are made on the source and
destination storage controller VMs during the site pairing operation to allow
the HX data platform access to the systems bi-directionally. You need to allow
this traffic on WAN routers for each HX node replication IP address and
management CIP address.
When you configure the local replication network on a HyperFlex cluster, you
can manually perform a Test Local Replication Network action to test
connectivity across the replication IP addresses of each storage controller VM
on the local cluster. This test includes port connectivity and firewall checks.
When the two clusters have been paired, you can manually perform a Test
Remote Replication Network action to test connectivity between each local
storage controller VM and each remote storage controller VM. Port connectivity
and firewall checks are performed.
You can also use the Linux “netcat” utility as an additional option to check
port connectivity.
Network Loss Reliable transmission of data enables replication between two paired clusters
to function optimally. Packet loss in data transmission between two paired
clusters may degrade replication performance.
Diagnosing Dropped Packets
There are two cases where packet loss may occur - network congestion and
transient network device errors.
If dropped packets occur on a replication network due to network congestion
the HyperFlex cluster replication engine automatically throttles back replication
bandwidth. Throttling replication network bandwidth reduces network
congestion and results in the reduction of dropped packets. In extreme cases,
replication bandwidth throttling may result in replication jobs taking longer to
complete than anticipated.
Dropped packets that occur on a replication network due to transient network
device errors may cause replication failures that occur randomly or at specific
times of the day.
Dropped packets are not reported in the HX Connect user interface.
Occurrences of packet drop are logged in the HyperFlex storage controller
logs. Users that experience noticeable replication job elongation or other failures
can contact support for further assistance.
Cluster Requirements
When configuring virtual machine replication and performing disaster recovery of virtual machines, please
ensure that the following cluster requirements are met:
difference-based replication job. An additional full copy worth of space needs to be budgeted on
both of the paired clusters.
• A protected VM undergoes storage vMotion and is migrated to a different datastore. If the destination
datastore is mapped to a datastore on the paired cluster, the next scheduled protection job will
perform a full copy replication job instead of a difference-based replication job. An additional full
copy worth of space needs to be budgeted on both of the paired clusters.
• A protected VM has a DP snapshot that was taken in conjunction with a replication job. Subsequent
to this, an initial HX native snapshot is created that also creates an HX Sentinel snapshot. The next
scheduled protection job will perform a full copy replication job instead of a difference-based
replication job. An additional full copy worth of space needs to be budgeted on both of the paired
clusters.
• When a protected VM DP snapshot is taken during an HX native snapshot workflow that created
intermediate delta disks, the next scheduled protection job will perform a full copy replication job
instead of a difference-based replication job. An additional full copy worth of space needs to be
budgeted on both of the paired clusters.
• When a new VMDK is added to an already protected VM, that specific VMDK will be full-copied
once.
Not having sufficient storage space on the remote cluster can cause the remote cluster to reach capacity
usage maximums. If you note any Out of Space errors, refer to Handling Out of Space Errors for more
information. Pause all replication schedules until space available on the cluster has been properly adjusted.
Always ensure that cluster capacity consumption is below the space utilization warning threshold.
Supported Configurations
Supported configurations for native replication (NRDR 1:1) are: 2N/3N/4N Edge, FI, and DC-no-FI based
clusters to 2N/3N/4N Edge, FI, and DC-no-FI based clusters, including stretched clusters, all managed through
HX Connect.
HyperFlex hardware acceleration cards (PID: HX-PCIE-OFFLOAD-1) are supported with native replication
beginning with HX 4.5(1a). You must enable HX Hardware Acceleration on both of the paired HyperFlex
clusters.
Rebooting Nodes
Do not reboot any nodes in an HX cluster during a restore, replication, or recovery operation. Note that node
reboot operation may occur as part of an upgrade process. You should pause the replication scheduler prior
to an upgrade, and then resume it after the upgrade has completed.
Component Description
HX Data Platform Version Ensure that the HyperFlex clusters that are going to be paired for
replication are running the same HX data platform software version.
Note that the use of different HX data platform versions is only
supported during HX data platform upgrades. In this scenario, one of
the paired HyperFlex clusters may be running a different version of HX
data platform software for the period of time until both of the paired
clusters have been upgraded. Ensure that you upgrade both of the paired
clusters to the same HX data platform version within the shortest
possible time frame based on site specific constraints. Also note that a
maximum of one major HX data platform release version difference is
permitted when upgrading paired clusters. Additionally, the changing
of any replication configuration parameter is not supported when the
paired clusters are not both running the same HX data platform version
during an upgrade.
Node Status Ensure that all HyperFlex cluster nodes are online and fully operational
prior to the creation of the local replication networks and performing
the site pairing process.
Component Description
Node Failure In the highly unlikely and rare event of a node failure, there may be an
impact to replication. As an example, replication jobs in progress will
stop if the node which has the replication CIP address enters an
inoperative state. At the point in time when the replication CIP address
is claimed by another node in the cluster, the replication job will
automatically resume. Similarly, if a recovery job was running on the
node with replication CIP address and the node failed, the job would
fail. The replication CIP address would subsequently be claimed by
another node in the cluster. Retry the operation upon noting the failure.
vCenter Recommendations Cisco HXDP Release 5.0(2a) and earlier-Ensure that each of the two
paired HyperFlex clusters is managed by a unique vCenter instance.
Also ensure that vCenter is deployed in a different fault domain for
availability during disaster recovery scenarios. In cases where a single
vCenter server is used for both of the paired HyperFlex clusters, certain
workflows cannot be performed entirely within the HX Connect user
interface.
Consideration Description
Thin Provisioning Protected VMs are recovered with thin provisioned disks irrespective
of how disks were specified in the originally protected VM.
VM Device Limitations Do not protect VMs with connected ISO images or floppies as
individually protected VMs, or within a protection group. You can set
any configured CD or DVD drive to “Client Device” with the
“Connected” state disabled. There is no need to delete the device from
the VM configuration. If there is a need to temporarily mount an ISO
image, you can unprotect the VM and then protect it again once you
have set the CD or DVD drive to “Client Device” and then disconnected.
Templates Templates are not supported with disaster recovery. Do not attempt to
protect a template.
Recovery of Virtual Machines with When recovering a protected VM that has VMware redolog snapshots,
Snapshots the VM is recovered and all previous snapshots redolog snapshots are
preserved.
Data Protection Snapshots Replicated DP snapshots are stored on the mapped datastore on the
paired cluster. You should not perform a manual deletion of DP
snapshots as this is not supported. Deleting snapshot directories or
individual files will compromise HX data protection and disaster
recovery.
Note To avoid deleting DP snapshots manually, it is important
to remember that VMware does not restrict operations on
datastores by the administrative user. In any VMware
environment, datastores are accessed by an administrative
user via the vCenter browser or by logging into the ESXi
host. Because of this, the snapshot directory and contents
are browsable and accessible to administrators.
Consideration Description
VMware SRM – Purposeful VM If a VM is deleted from VMware vCenter and the VM is located in an
Deletion “Other DRO” datastore pair, the SRM recovery plan for this datastore
pair will fail during recovery. To avoid this scenario, first unprotect the
VM using the following command on one of the HyperFlex storage
controller VMs: stcli dp vm delete --vmid <VM_ID>
HyperFlex Software Encryption Software encryption must be enabled on clusters in both paired
datastores to be able to protect VMs on encrypted datastores.
You must install an appropriate SRA on the Site Recovery Manager Server hosts at both the protected and
recovery sites. If you use more than one type of storage array, you must install the SRA for each type of array
on both of the Site Recovery Manager Server hosts.
Before installing an SRA, ensure that SRM and JDK 8 or above version are installed on Windows machines
at the protected and recovery sites.
To install an SRA, do the following:
1. Download SRA from the VMware site.
In the https://my.vmware.com/web/vmware/downloads page, locate VMware Site Recovery Manager and
click Download Product. Click Drivers & Tools, expand Storage Replication Adapters, and click Go
to Downloads.
2. Copy the Windows installer of SRA to SRM Windows machines at both the protected and recovery sites.
3. Double-click the installer.
4. Click Next on the Welcome page of the installer.
5. Accept the EULA and click Next.
6. Click Finish.
After SRA installation, refer the following guide as per the SRM release version and do the SRM environment
setup:
• SRM 8.1 configuration—https://docs.vmware.com/en/Site-Recovery-Manager/8.1/srm-admin-8-1.pdf
• SRM 6.5 configuration—https://docs.vmware.com/en/Site-Recovery-Manager/6.5/srm-admin-6-5.pdf
• SRM 6.0 configuration—https://docs.vmware.com/en/Site-Recovery-Manager/6.0/srm-admin-6-0.pdf
After configuration, SRM works with SRA to discover arrays and replicated and exported datastores, and to
fail over or test failover datastores.
SRA enables SRM to execute the following workflows:
• Discovery of replicated storage
• Non-disruptive failover test recovery using a writable copy of replicated data
• Emergency or planned failover recovery
• Reverse replication after failover as part of failback
• Restore replication after failover as part of a production test
Local cluster―The cluster you are currently logged into through HX Connect, in a VM replication cluster
pair. From the local cluster, you can configure replication protection for locally resident VMs. The VMs are
then replicated to the paired remote cluster.
Migration―A routine system maintenance and management task where a recent replication DP snapshot
copy of the VM becomes the working VM. The replication pair of source and target cluster do not change.
Primary cluster―An alternative name for the source cluster in VM disaster recovery.
Protected virtual machine― A VM that has replication configured. The protected VMs reside in a datastore
on the local cluster of a replication pair. Protected VMs have a replication schedule configured either
individually or by inclusion in a protection group.
Protection group―A means to apply the same replication configuration to a group of VMs.
Recovery process―The manual process to recover protected VMs in the event the source cluster fails or a
disaster occurs.
Recovery test―A maintenance task that ensures the recovery process will be successful in the event of a
disaster.
Remote cluster―One of a VM replication cluster pair. The remote cluster receives the replication snapshots
from the Protected VMs in the local cluster.
Replication pair―Two clusters that together provide a remote cluster location for storing the replicated DP
snapshots of local cluster VMs.
Clusters in a replication pair can be both a remote and local cluster. Both clusters in a replication pair can
have resident VMs. Each cluster is local to its resident VMs. Each cluster is remote to the VMs that reside on
the paired local cluster.
DP snapshot―Part of the replication protection mechanism. A type of snapshot taken of a protected VM,
which is replicated from the local cluster to the remote cluster.
Secondary cluster―An alternative name for the target cluster in VM disaster recovery.
Source cluster―One of a VM replication cluster pair. The source cluster is where the protected VMs reside.
Target cluster―One of a VM replication cluster pair. The target cluster receives the replicated DP snapshots
from the VMs of the source cluster. The target cluster is used to recover the VMs in the event of a disaster
on the source cluster.
• An ongoing testing strategy for each SLA which serves to prove that the solution meets the business
requirements it was designed for.
Note that backups and backup copies must be stored external to the HyperFlex cluster being protected. For
example, backups performed to protect VMs on a HyperFlex cluster should not be saved to a backup repository
or a disk library that is hosted on the same HyperFlex cluster.
The built-in HyperFlex data protection capabilities are generalized into the following categories:
• Data Replication Factor—Refers to the number of redundant copies of data within a HyperFlex cluster.
A data replication factor of 2 or 3 can be configured during data platform installation and cannot be
changed. The data replication factor benefit is that it contributes to the number of cluster tolerated failures.
See the section titled, HX Data Platform Cluster Tolerated Failures, on page 10 for additional information
about the data replication factor.
Note Data Replication Factor alone may not fulfill requirements for recovery in the
highly unlikely event of a cluster failure, or an extended site outage. Also, the
data replication factor does not facilitate point-in-time recovery, retention of
multiple recovery points, or creation of point-in-time copies of data external to
the cluster.
Note HX native snapshots alone may not fulfill requirements for recovery in the unlikely
event of a cluster failure, or an extended site outage. Also, HX native snapshots
do not facilitate the ability to create point-in-time copies of data external to the
cluster. More importantly, unintentional deletion of a VM also deletes any HX
native snapshots associated with the deleted VM.
• Asynchronous Replication—Also known as The HX Data Platform disaster recovery feature, it enables
protection of virtual machines by replicating virtual machine DP snapshots between a pair of network
connected HyperFlex clusters. Protected virtual machines running on one cluster replicate to the other
cluster in the pair, and vice versa. The two paired clusters typically are located at a distance from each
other, with each cluster serving as the disaster recovery site for virtual machines running on the other
cluster.
Note Asynchronous Replication alone may not fulfill requirements for recovery when
multiple point-in-time copies need to be retained on the remote cluster. Only the
most recent snapshot replica for a given VM is retained on the remote cluster.
Also, asynchronous replication does not facilitate the ability to create point-in-time
copies of data external to either cluster.
It is recommended to first understand the unique business requirements of the environment and then deploy
a comprehensive data protection and disaster recovery solution to meet or exceed those requirements.
Protection attributes can be created and assigned to protection groups. To assign those protection attributes
to VMs, they can be added to a protection group.
For example, there are three different SLAs: gold, silver, and bronze. Set up a protection group for each SLA,
with replication intervals such as 5 or 15 minutes for gold, 4 hours for silver, and 24 hours for bronze. Most
VMs can be protected by simply adding them to one of the three already created protection groups.
To protect VMs, you can choose from the following methods:
Note When you select multiple VMs, you must add them to a protection group.
• Independently―Select one VM and configure protection. Set the replication schedule and the VMware
quiesce option for the specific VM. Changes to the replication settings will only affect the independently
protected VM. The VM is not included in a protection group.
• Existing protection group―Select one or more VMs and add them to an existing protection group. The
schedule and VMware quiesce option settings are applied to all of the VMs in the protection group. If
the protection group settings are changed, the changes are applied to all of the VMs in the protection
group.
• New protection group―Select two or more VMs and choose to create a new protection group. Define
the protection group name, schedule, and VMware quiesce option settings. These settings are applied to
all the VMs in the protection group. If the protection group settings are changed, the changes are applied
to all of the VMs in the protection group.
• Assign a replication schedule to the VMs to set the frequency (interval) for creating DP snapshots on the
source cluster and replicate them to the target cluster. Replication schedules can be assigned to individual
VMs and to protection groups.
Replication Workflow
1. Install HX Data Platform, create two clusters.
2. Create at least one datastore on each cluster.
3. Log into HX Connect.
4. Before creating the replication network, verify the IP addresses, subnet mask, VLAN, gateway, and IP
range to be used for the replication network. After the replication network is created, validate connectivity
within the cluster over the replication network.
5. The default value of MTU is 1500. If the HyperFlex cluster uses OTV or other tunneling mechanisms,
ensure choosing an MTU value which will work for inter-site or inter-cluster connectivity. Starting with
Cisco HyperFlex Release 5.0(2a) the MTU field is editable.
6. Configure cluster replication network on each cluster. The replication network information is unique
to each cluster.
Specify the subnet, gateway, range of IP addresses, bandwidth limit for dedicated use by the replication
network. HX Data Platform configures a VLAN through UCS Manager for both clusters.
7. An intra-cluster network test is performed to validate connectivity between the nodes in the cluster,
after the replication network is configured. If the intra-cluster network test fails, the replication network
configuration is rolled back. Reconfigure the replication network after resolving any issues.
8. Before creating the replication pair, ensure that you have updated the corporate network to support this
pairing.
9. Create a replication pair from one cluster to the other, connecting the two clusters. After the replication
pair is created, a test of the inter-cluster pair network is performed to validate bidirectional connectivity
between the clusters. Set the datastore mapping(s) from both clusters.
10. Optionally, you can create protection groups.
• Set the schedule. Each protection group must have one schedule.
• Create multiple protection groups if you want to have various replication intervals (schedules) for
different VMs. A VM can only belong to one protection group.
11. Select VMs to protect, as individual virtual machines or VMs assigned to protection groups.
12. Set protection, do the following:
a. Select one or more VMs. Click Protect.
b. From the Protect VM wizard, the options are:
• Protect a single VM with an existing protection group.
• Protect a single VM independently.
Set the schedule.
• Protect multiple VMs with an existing protection group.
Step 3 In the Configure Replication Network dialog box, under the VLAN Configuration tab, enter the network information.
Create a new VLAN radio button Click this radio button to create a new VLAN.
Note If you are configuring replication network on edge cluster, do not
use the Create VLAN option. Use the existing VLAN option and
follow the same procedure.
VLAN Name field This field is automatically populated with a default VLAN name when the
Create a new VLAN radio button is selected. The VLAN ID is concatenated
to the name.
For Stretched Cluster, provide Cisco UCS Manager credentials for primary and secondary FIs (site A and site B).For
normal cluster, provide Cisco UCS Manager credential for single FI.
UCS Manager host IP or FQDN field Enter Cisco UCS Manager FQDN or IP address.
For example, 10.193.211.120.
Gateway field Enter the gateway IP address for use by the replication network. The gateway
is separate from the HX Data Platform Management traffic network and Data
traffic network.
For example, 1.2.3.4.
Note The gateway IP address must be accessible even if the disaster
recovery is being setup for a flat network.
Add IP Range button Click to add the range of IP addresses entered in IP Range From and To fields.
Note When you use an existing VLAN for replication network, the replication network configuration fails. You
must add the self-created replication VLAN to the management vNIC templates in Cisco UCS Manager.
What to do next
• Be sure to configure the replication network on both HX Storage clusters for the replication pair.
• After the replication network is created on the cluster, each converged node on the cluster would be
configured with an IP address on the eth2 interface.
• Check for duplicate IP assignment using 'arping'.
For example:arping -D -b -c2 -I ethX $replicationIP` (replace ethX and $replicationIP
with actual values).
Step 3 From the Actions drop-down list, select Edit Replication Network.
Step 4 In the Edit Network Configuration dialog box, you can edit the range of IPs to use and set the replication bandwidth
limit for replication traffic. The replication network subnet and gateway are displayed for reference only and cannot be
edited.
Gateway field The gateway that is configured for the replication network. This is value cannot
be edited.
IP Range field Enter a range of IP addresses for use by the replication network.
• The minimum number of IP addresses required is the number of nodes in
the HX Storage Cluster plus one more.
For example, if the HX Storage Cluster has 4 nodes, the IP Range must be
at least 5 IP addresses.
• The from value must be lower than the to value.
For example, From 10.10.10.20 To 10.10.10.30.
• You can add IP addresses at any time.
• If you plan to add nodes to your cluster, include sufficient number of IP
addresses to accommodate any additional nodes.
Add IP Range field Click to add the range of IP addresses that are entered in IP Range From and
To fields.
Set replication bandwidth limit check Enter the maximum network bandwidth that the replication network is allowed
box (Optional) to consume for outbound traffic.
Valid Range: 10 to 10,000. The default is unlimited, which sets the maximum
network bandwidth to the total available replication network bandwidth.
The replication bandwidth is used to copy DP snapshots from the local HX
Storage Cluster to the paired remote HX Storage Cluster.
The replication network is now updated. Any additional replication network IP addresses are available for
new nodes should they be added to the storage cluster. Replication traffic adjusts to any changes made to the
bandwidth limit.
Note When pairing HX clusters, if you get the error: Check your cluster status or logs for possible
solutions appears, check if the pairing is successful by running the following command:
Step 1 From HX Connect, log into either the local or remote HX cluster as a user with administrator privileges and do one of
the following:
a) Select Replication > Pair Cluster if you are doing cluster pairing for the first time.
b) Select Replication > Create Replication Pair.
The Create Replication Pair option is enabled only when you delete an existing replication pair after unprotecting
all the VMs and removing all the dependencies.
Step 2 Enter a Name for the replication pair and click Next.
Enter a name for the replication pairing between two HX Storage clusters. This name is set for both the local and remote
cluster. The name cannot be changed.
User Name and Password fields Enter vCenter single sign-on or cluster specific administrator credentials for the
remote HX cluster.
HX Data Platform verifies the remote HX cluster and assigns the replication pair name.
Once the Test Cluster Pair job is successful, you can proceed to the next step. On the Activity page, you can view the
progress of the Test Cluster Pair job.
Note Virtual machines to be protected must reside on one of the datastores in the replication pair.
Step 5 To protect VMs using the HX Data Platform disaster recovery feature, click Native Protection and do the following:
a) The Local Datastore column displays a list of the configured datastores on the local HX Storage cluster. Map one
local datastore to one remote datastore.
b) From the Remote Datastore pull-down menu, choose a datastore that needs to be paired with the local datastore.
c) Click Map Datastores.
If you chose to cancel the datastore mapping by clicking Cancel, you can map the datastores later using Map
Datastores that appears under DATASTORE MAPPED in the Replication dashboard.
Note • The virtual machines to be protected must be on the datastores you select. Moving virtual machines
from the configured datastores for the replication pair, also removes protection from the virtual machines.
• Moving virtual machine to another paired datastore is supported. If the VMs are moved to unpaired
datastore, the replication operation fails.
Note Once a local datastore is mapped to a remote datastore, the corresponding local datastore will not appear
under Other DRO Protection.
Step 6 To protect VMs using SRM through disaster recovery orchestrator (DRO), click Other DRO Protection and do the
following:
a) The Local Datastore column displays a list of the unpaired configured datastores on the local HX cluster. Map one
local datastore to one remote datastore.
b) From the Remote Datastore pull-down menu, choose a datastore that need to be paired with the local datastore.
c) From the Direction pull-down menu, choose Incoming or Outgoing as the direction of VM movement for the mapped
datastores.
d) From the Protection Schedule pull-down menu, choose the schedule for protecting all the VMs in the datastore.
e) Click Map Datastores.
If you chose to cancel the datastore mapping by clicking Cancel, you can map the datastores later using Map
Datastores that appears under DATASTORE MAPPED in the Replication dashboard.
Note If a new VM is added to the protected datastore, the newly added VM is also get protected by Cisco HyperFlex.
Note The replication pairs that are edited under Other DRO Protection, are exposed to SRM.
What to do next
To check the protection status of virtual machines, do one of the following:
• Click Virtual Machines in HX Connect. This displays the list of the virtual machines on the local cluster
along with the protection status. If the VM is protected by SRM, the status is displayed as Protected (by
other DRO).
Note In the Virtual Machine page, the status of the VMs protected by SRM is displayed
as unprotected until the completion of first auto protect cycle. Until then, the
user is not recommended to manually protect those VMs.
Field Description
MTU Test Value Default MTU value is 1500. MTU can be set in the range 1024 to 1500.
Note • Starting with HXDP versions 5.0(2a) and later, you can edit
the MTU value after configuring the cluster. With older
versions of HXDP, the existing replication network
configuration will need to be removed. The replication
network can then be configured with the correct MTU value.
Step 4 To protect VMs using the HX Data Platform disaster recovery feature, click Native Protection and do the following:
a) The Local Datastore column displays a list of the configured datastores on the local HX Storage Cluster. Map one
local datastore to one remote datastore.
b) From the Remote Datastore pull-down menu, choose a datastore that needs to be paired with the local datastore.
c) Click Map Datastores.
To change the local datastore selection:
a. From the Remote Datastore pull-down menu, choose Do not map this datastore to remove the mapping from the
current local datastore.
b. From the Remote Datastore pull-down menu, choose a datastore to be paired with the local datastore.
Note Once a local datastore is mapped to a remote datastore, the corresponding local datastore will not appear
under Other DRO Protection.
Step 5 To protect VMs using SRM through disaster recovery orchestrator (DRO), click Other DRO Protection and do the
following:
a) The Local Datastore column displays a list of the unpaired configured datastores on the local HX cluster. Map one
local datastore to one remote datastore.
b) From the Remote Datastore pull-down menu, choose a datastore that need to be paired with the local datastore.
c) From the Direction pull-down menu, choose Incoming or Outgoing as the direction of VM movement for the mapped
datastores.
d) From the Protection Schedule pull-down menu, choose the schedule for protecting all the VMs in the datastore.
e) Click Map Datastores.
Note New VMs added to the protected datastore are also protected.
Note The replication pairs that are edited under Other DRO Protection, are exposed to SRM.
What to do next
To check the protection status of virtual machines, do one of the following:
• Click Virtual Machines in HX Connect. This displays the list of the virtual machines on the local cluster
along with the protection status. If the VM is protected by SRM, the status is displayed as Protected (by
other DRO).
Note In the Virtual Machine page, the status of the VMs protected by SRM is displayed
as unprotected until the completion of first auto protect cycle. Until then, the
user is not recommended to manually protect those VMs.
Run the stcli dp peer delete on one of the clusters in a pair to ensure that the pairing relation is removed from both
clusters in the pair.
When successful, both the clusters are available for fresh configuration of data protection.
c) Ensure all the possible remote datastores are set to Do not map this datastore.
d) Click Finish.
Step 3 Select Replication > Replication Pairs > Delete.
Step 4 Enter administrator credentials for the remote cluster and click Delete.
Password field Enter the administrator password for the remote HX Storage Cluster.
Protection groups are created on the HX cluster that the administrative user is logged on to. Protection groups
provide protection to the VMs that are members of a given protection group. If protection groups have protected
virtual machines that replicate to the remote cluster, they are listed in HX Connect.
Note The administration of protection groups can only be performed from the local cluster where it was created.
Protect virtual machines in this Select how often the virtual machines are to be replicated to the paired cluster.
group every field
The pull-down menu options are: 5 minutes, 15 minutes, 30 minutes, 1 hour,
90 minutes, 2 hours, 4 hours, 8 hours, 12 hours, and 24 hours. The default value
is 1 hour.
Start protecting the virtual machines Select this radio button if you want the first replication to start immediately
immediately radio button after you add the first virtual machine to this protection group.
Cluster time zone and Current time on cluster are references to help you to
choose the appropriate replication start time. Start time is based on the local
cluster clock. For example:
10 hours, 3 minutes from now with Current time on cluster, 1:56:15PM, means
that the first replication occurs at 11:59:00PM.
The hours, minutes from now states when the first replication will occur. This
is updated when you change the time field setting.
Use VMware Tools to quiesce the Select this check box to take quiesced DP snapshots. Leaving the check box in
virtual machine check box an unchecked state will take crash consistent DP snapshots.
This only applies to virtual machines with VMware Tools installed.
Step 5 Click the Replication > Protection Groups to view or edit the new protection group.
If the number of VMs is zero (0), add virtual machines to the new protection group to apply the replication schedule
configured in the protection group.
Quiescence Overview
Quiescence is a process of bringing the on-disk data of a physical or virtual computer into a state suitable for
backups. This process might include operations such as flushing dirty buffers from the operating system's
in-memory cache to disk or other higher-level application-specific tasks.
HX Data Protection (DP) snapshots can be created with the guest file system quiesced. The quiesce option
is available when using Cisco HyperFlex Connect, the HyperFlex command line user interface (UI), and HX
REST APIs. VMware tools should be installed in the guest VM when creating HX DP snapshots using the
quiesce option. For information on VMware, go to the VMware Website for the following:
• VMware Compatibility Guide.
HXDP Software Release 5.0(2a) and earlier supports the following guest states:
• guestToolsCurrent
• guestToolsUnmanaged
When the quiesce data protection snapshot fails, the DataProtectionVmError occurs which prompts an HX
event and an HX alarm.
Use VMware Tools to quiesce the Select the check box to take quiesced DP snapshots. The checkbox is unchecked
virtual machine check box by default; leaving the check box unchecked, takes crash consistent DP
snapshots.
This only applies to virtual machines with VMware Tools installed.
Step 4 Click Save Changes to save the interval and VMware Tools quiescence settings for the protection group. View the
Protection Groups tab to verify the interval frequency.
Step 1 Log into HX Connect with administrator privileges and select Virtual Machines.
This lists the virtual machines on the local HX cluster.
Step 2 Select one (1) or more unprotected VMs from the list.
Click in the virtual machine row to select it. As you click a virtual machine row, the corresponding virtual machine check
box is selected.
Add to an existing protection group Select an existing protection group from the pull-down list.
radio button
The interval and schedule settings of the protection group are applied to the
selected VM or VMs.
Create a new protection group radio Enter a name for the new protection group for this local cluster.
button
Protection groups are unique to each cluster. The name is referenced on
theremote cluster, but not editable on the remote cluster. You can create multiple
protection groups on each cluster.
Step 5 Select a protection group from the pull-down list and click Next.
Be sure the protection group you choose has the schedule interval desired.
The Protect Virtual Machines wizard, Summary page is displayed.
Step 6 Confirm the information in the Summary page and click Add to Protection Group.
The selected VM or VMs are added to the protection group. View theReplication or Virtual Machines pages to confirm
that the VM oir VMs have been added to the protection group.
Step 1 Log into HX Connect with administrator privileges and select Virtual Machines.
This lists the virtual machines on the local HX cluster.
Step 4 Click the radio button, Create a new protection group, add a name for the protection group, and click Next.
The Protection Schedule Wizard Page wizard page is displayed.
Step 5 Complete the schedule and VMware quiesce option, as needed, and click Next.
Protect virtual machines in this Select how often the virtual machines are to be replicated to the paired cluster.
group every field The default value is every 1 hour.
Start protecting the virtual machines Select this radio button if you want the first replication to start immediately
immediately radio button after you add the first virtual machine to this protection group.
Start protecting the virtual machines Select this radio button if you want to set a specific time for the first replication
at radio button to start. To start replication requires:
• At least one virtual machine is added to the protection group.
• The scheduled start time is reached.
Cluster time zone and Current time on cluster are references to help you to
choose the appropriate replication start time. Start time is based on the local
cluster clock. For example:
10 hours, 3 minutes from now with Current time on cluster, 1:56:15PM, means
that the first replication occurs at 11:59:00PM.
Use VMware Tools to quiesce the Click the check box to take quiesced DP snapshots. An unchecked check box
virtual machine check box takes crash consistent DP snapshots. The check box is unchecked by default.
This only applies to virtual machines with VMware Tools installed.
Step 6 Confirm the information in the Summary page and click Add to Protection Group.
Review the summary content to confirm the settings to apply to the selected virtual machines.
• Name of the protection group
• Number of virtual machines to protect
• Names of virtual machines
• Storage provisioned for each virtual machine
• Storage used (consumed) by each virtual machine
The selected VM or VMs are added to the protection group. View the Replication or Virtual Machines pages to verify
that the VM(s) have been added to the protection group.
• Independently―Select one (1) VM and configure protection. Set the replication schedule and the
VMware Tools quiesce option for the specific VM.
Changes to the replication settings only affect the independently protected VM. The VM is not a member
of a protection group.
• Existing protection group―Select one or more VMs and add them to an existing protection group. The
schedule and VMware Tools quiesce option settings are applied to all the VMs in the protection group.
When the protection group settings are changed, the changes are applied to all VMs in the protection
group.
Step 1 Log into HX Connect with administrator privileges and select Virtual Machines.
Step 2 Select one unprotected virtual machine from the list. Click in the virtual machine row to select it.
Click in the virtual machine row to select it. As you click a virtual machine row, the corresponding virtual machine check
box is selected.
Protect this virtual machine Enables the interval, schedule options, and VMware Tools quiescence option
independently radio button for defining protection for this VM.
Protect this virtual machine every Select from the pull-down list how often the virtual machines are to be replicated
field to the paired cluster.
The list values are: 5 minutes, 15 minutes, 30 minutes, 1 hour, 90 minutes, 2
hours, 4 hours, 8 hours, 12 hours, and 24 hours.
Start protecting the virtual machines Select this radio button if you want the first replication to start immediately
immediately radio button after you add the first virtual machine to this protection group.
Cluster time zone and Current time on cluster are references to help you to
choose the appropriate replication start time. Start time is based on the local
cluster clock. For example:
10 hours, 3 minutes from now with Current time on cluster, 1:56:15PM, means
that the first replication occurs at 11:59:00PM.
VMware Tools to quiesce the virtual To take quiesced DP snapshots, check the check box. The unchecked check box
machine check box takes crash consistent DP snapshots. The check box is unchecked by default.
This only applies to virtual machines with VMware Tools installed.
Note There is no need to unprotect VMs to pause replication for HX cluster activities. See Pausing Replication, on
page 231.
Field Description
Virtual Machine Power State radio button By default, the Off option is selected. The recovered VM is powered
on as per the selected option.
Field Description
Test Virtual Machine Name Prefix field Enter a prefix that you want to add to the virtual machine after test
recovery. The prefix helps to identify the type and context of the
resources.
Notification Setting radio button Choose Normal Mode to get a confirmation prompt with summary of
configuration at the time of recovery, test recovery, or migration. Choose
Silent Mode to not get a confirmation prompt. On choosing silent mode,
a confirmation window appears with the description of default behavior
of silent mode. If you agree to the default behavior of silent mode, click
OK.
Resource Pool area Map the resources in the protected site to the resources in the remote
site for the recovery, test recovery, and migrate configuration.
Check the Same as Recovery Configuration check box to use the
resource mapping of recovery configuration for the test recovery
configuration.
Click Add Rule to add one more resource pool mapping. Click the
delete icon to remove a rule. To edit a rule, delete the rule and add the
updated rule as a new rule.
Folder area Map the folders in the protected site to the folders in the remote site for
the recovery, test recovery, and migrate configuration.
Check the Same as Recovery Configuration check box to use the
folder mapping of recovery configuration for the test recovery
configuration.
Click Add Rule to add one more folder mapping. Click the delete icon
to remove a rule. To edit a rule, delete the rule and add the updated rule
as a new rule.
Network area Map the network in the protected site to the network in the remote site
for the recovery, test recovery, and migrate configuration.
Check the Same as Recovery Configuration check box to use the
network mapping of recovery configuration for the test recovery
configuration.
Click Add Rule to add one more network mapping. Click the delete
icon to remove a rule. To edit a rule, delete the rule and add the updated
rule as a new rule.
Note The RECOVERY SETTINGS field shows the last validated result. Once a rule is created for recovery,
there is no automatic periodic validation. However, a validation job can be executed to check the validity of
the rules existing in the recovery settings
The validation job summary in the Activity page directs the user to check the Recovery Settings page to
view the validation result.
After configuring the recovery settings, validation of the recovery settings can be performed by choosing
Validate Recovery Settings from the Actions drop-down list. The successful initiation of the recovery setting
validation message is displayed. The RECOVERY SETTINGS field displays the time stamp of last validation.
To monitor the progress of validation, click the Activity tab. In the normal notification setting mode, during
recovery, test recovery, or migration of a virtual machine, the configured recovery settings are displayed.
It is possible to view the recovery configurations and edit them as required by checking the Modify recovery
configuration for current operation check box. But, the recovery settings changes are applicable only for
the current operation and the changes will not be updated to the global recovery settings.
Note • Testing recovery does not disrupt the running clusters. The intent is to verify, in the event of an actual
disaster, that the VMs are recoverable.
• Using the HX Connect user interface, to test VM recovery, you can run a maximum of 10 tasks in a
sequence without waiting for the previously submitted task to complete.
Important Only one copy of the test recovered VM can be made at any point. If you need to have another test recovered
VM, please delete the previously created test recovered VM.
Resource Pool drop-down list Select a location for the test VM to be stored.
Folders drop-down list Select a location for the test VM to be stored, for example:
• Discovered Virtual Machine
• HX Test Recovery
Power On/Off radio button By default, the Power ON option is selected. The recovered VM is powered
on or left off after it is created as per the selected option.
VM Name field Enter a new name for the created test VM.
Test Networks radio button Select which HX Storage Cluster network to use for transferring the data from
the replication snapshot.
Network options for example:
• Storage Controller Data Network
• Storage Controller Management Network
• Storage Controller Replication Network
• VM Network
Map Networks radio button Select to create a map between the source and the target cluster networks.
• Source Network―Network name at the source side on which the VM is
connected.
• Target Network―Select target network from the drop-down list, where
the VM has to be connected.
Attention • You may configure the folder, network, or resource pool parameters to be used during recovery, test
recovery and migrate operations. If the global recovery setting is not configured, you will need to explicitly
map individual VMs at the time of recovery.
• Recover VM is not supported between different vSphere versions. If the Target is at a lower version
vSphere environment and does not support the hardware version of a protected VM on the primary, VM
test recovery and recovery may fail. Cisco recommends to test recover each protected VM to validate
the support on the target site.
Upgrade the target environment to enable recovery of protected VMs.
• Cancelling a recovery process (rolling back) is not supported. Attempting to cancel a recovery process
changes all VMs in an unprotected ‘ready to recovery’ state.
• When running recovery on VMs, you may specify explicit network mapping when recovering the VMs
to avoid unintentional network connections to recovered VMs.
You can skip specifying network mapping in the following cases:
• If the source VMs use vSphere Standard Switches and if all ESXi hosts on the recovery side have
standard switch networks with the same name.
• If the source VMs use vSphere Distributed Switch (vDS) port groups and if the recovery site has
identically named vDS port groups.
• If you want to specify network mapping, ensure that both the name and the type of the VM network
matches between the source and the target.
• When running recovery on virtual machines that are individually protected, or that are in different
protection groups, the maximum number of concurrent recovery operations on a cluster is 20.
Step 3 To recover the VM and build a new VM on the local cluster, click the Recover VM button.
Note When you configure recovery settings, the following fields are auto-populated.
Resource Pool drop-down list Select a location for the new VM to be stored.
Power On/Off radio button By default Power ON option is selected. The recovered VM is powered on or
left off after it is created as per the selected option.
Map Networks Select to create a map between the source and target cluster networks.
• Source Network―Network name at the source side on which the VM is
connected.
• Target Network―Select target network from the drop-down list, where
the VM has to be connected.
The recovered VMs are displayed in the Standalone Protected VMs subpane.
Step 2 Recover the remaining virtual machines from the Standalone Protected VMs subpane, which were a part of the protection
group. See Recovering Virtual Machines, on page 224 for more details.
Planned Migration
Performing a planned migration pauses the replication schedule, replicates the most recent copy, recovers on
the target, switches the ownership from the source to the target, and resumes replication on the target that is
now the new source.
To perform a planned migration, take the following steps:
Step 1 Log into HX connect of the target cluster. The target cluster is where the replicated DP snapshots were copied to.
Step 2 On the target cluster, select Replication > Remote VMs Tab > protected_vm.
Step 3 Click Migrate.
Note All the fields that are listed here are optional.
Resource Pool drop-down list Select a location for the new VM to be stored.
Power On/Off radio button By default Power ON option is selected. The recovered VM is powered on or
left off after it is created as per the selected option.
Map Networks Select to create a map between the source and target cluster networks.
• Source Network―Network name at the source side on which the VM is
connected.
• Target Network―Select target network from the drop-down list, where
the VM has to be connected.
Low Bandwidth and Temporary Packet Loss - In the event VM migration operation fails with an error
message that contains "THRIFT_EAGAIN (timed out)", retry the VM Migration. The timeout error is due to
temporary network congestion caused by bandwidth saturation or underlying network packet loss.
Step 1 Using the Web CLI, run the following command to prepare for failover on the source:
# stcli dp vm prepareFailover --vmid <VMID>
Note You can use the stcli dp vm list --brief command to determine the VMID of a protected VM.
Step 3 Log into HX Connect of the secondary site. Select Replication > Remote VMs Tab > protected_vm. Click Migrate.
Step 4 After the Migrate task has completed successfully, log into vSphere Web Client of the secondary site and manually
register the VM.
a) Log into vSphere Web Client Navigator. Select Configuration > Storage.
b) Right-click on the appropriate datastore and click Browse Datastore.
Navigate to the virtualmachine name.vmx file, right-click on the file and click Add to Inventory. Follow the wizard
to manually register the VM.
Low Bandwidth and Temporary Packet Loss - In the event VM migration operation fails with an error
message that contains "THRIFT_EAGAIN (timed out)", retry the VM Migration. The timeout error is due to
temporary network congestion caused by bandwidth saturation or underlying network packet loss.
Step 2 Migrate the remaining virtual machines from the Standalone Protected VMs subpane, which were a part of the protection
group. See Planned Migration, on page 226 for more details.
• Using the HX Connect user interface, you can run a maximum of 5 re-protect tasks in a sequence without
waiting for the previously submitted task to complete.
Step 1 Log into HX connect of the source and the target. The target cluster is where the replicated DP snapshots were copied
to. The source cluster is the cluster where the VMs reside.
Step 2 Select a VM from the remote VM list. Execute the Recover VM workflow on this cluster workflow.
Note If both the target and source clusters are on the same vCenter, then unregister the VM on the source cluster.
This ensures that vCenter no longer has a record of the VM and it stops managing the VM, but it retains the
data for the VM.
Step 3 Select Replication > > Remote VMs tab > > unprotected and click Recover.
Step 4 To recover on the target VM and build a new VM on the local cluster, click the Recover VM button.
Complete the following fields in the Recover VM on this cluster dialog box.
Resource Pool drop-down list Select a location for the new VM to be stored.
Power On/Off radio button By default the Power ON option is selected. The recovered VM is
powered on or left off after it is created as per the selected option.
Map Networks Select to create a map between the source and target cluster networks.
• Source Network―Network name at the source side on which the
VM is connected.
• Target Network―Select target network from the drop-down list,
where the VM has to be connected.
In a situation where Cluster A and B were paired, and cluster B is permanently down, or unavailable for an
extended period of time, it may be necessary to remove the pairing relation on cluster A the proper solution
is to use the stcli dp peer forget --pair-name on cluster A.
Step 1 Recover the Virtual Machines. Perform standalone recovery (Recovering VMs) or group recovery (Recovering VMs in
protection groups). See Recovering Virtual Machines, on page 224 for more details.
Step 2 To clear the existing pairing, run the following command in the HX Connect WebCLI:
stcli dp peer forget --all
Step 3 Unprotect all the local and remote VMs. See Unprotecting Virtual Machines, on page 219 for more details.
Step 4 Use STCLI to clean-up the protection group data.
Remove Protection group (if any)
stcli dp group list
stcli dp group delete --groupid <groupUUID>
Step 5 (Optional) If needed, use the stcli drnetwork cleanup command to change the DR network. For more information see
the Cisco hyperFlex Data Platform CLI Guide for your release.
Step 6 Pair to the new cluster. See the Creating a Replication Pair, on page 205section for more details.
Step 7 Protect the virtual machines.
• Testing recovery―Testing if the recovery methods are working. See Testing Virtual Machine Recovery,
on page 222 for more details.
• Pausing replication―When preparing to upgrade the HX cluster and replication is configured, replication
activity must be paused.
Use the stcli dp schedule pause command.
• Resuming replication―After HX cluster maintenance activities are complete, resume the replication
schedule.
Use the stcli dp schedule resume command.
• Migration―The option to shift VMs from one source cluster to the replication paired target cluster,
making the target cluster the new source cluster for the migrated VMs.
The following image illustrates which configuration is used for Disaster Recovery on HyperFlex if you are
deploying in an ACI setup on a broad scale:
Pausing Replication
Before performing a storfs or platform upgrade, if replication is configured replication activity must be paused.
Resuming Replication
After successfully upgrading the HX Storage Cluster which had replication configured, do the following to
resume the replication schedule.
The previously configured replication schedule for all the protected virtual machines begins.
Replication Page
Displays summary information and links to detailed information related to Replication Configuration, Local
Protection, and Remote Protection.
BANDWIDTH LIMIT field Displays the configured bandwidth allowed for transmitting incoming
and outgoing replication data.
• Blank—The replication network is not configure.
• # Mbps—Configured setting in Mega bits per second (Mbps).
• Maximum―The default setting. Allows the replication network
to use the total available network bandwidth.
Actions drop-down list Click to create or edit the replication network for this HX Storage Cluster
and to test the replication network.
• Test Local Replication Network—<define>
• Edit Replication Network—Edit IP Range and set replication
bandwidth limit.
RECOVERY SETTINGS field Displays the state of the recovery settings configuration.
• Click Configure that appears when recovery settings configuration
is not done, to configure the recovery settings to return the network
to a known working state during recovery or test recovery.
Actions drop-down list Choose any one of the action to perform specific operation on replication
network, recovery settings, and datastore mapping.
• Test Remote Replication Network—To test the pairing between
clusters in a remote replication network.
• Validate Recovery Settings—To validate the configured recovery
settings.
• Edit Recovery Settings—To edit the recovery settings
configuration.
• Edit Datastore Mapping—To edit the mapping between local and
remote datastores.
Protected field Displays the total number of virtual machines that have a replication
snapshot created.
Click the field to display the list of protected virtual machines in the
Local VMs or Remote VMs tab.
Exceeds Interval field Displays the number of replications that took longer than the configured
interval to complete.
For example, if a virtual machine has an interval of every 15 minutes
and replicating its snapshot from the local to the remote cluster took 20
minutes, the replication exceeded the interval.
Click the field to display the list of virtual machines with exceeded
interval in the Local VMs or Remote VMs tab.
Current Replication Failures Displays the current number of replications that did not complete.
field
Click the field to display the list of virtual machines with failed
replication in the Local VMs or Remote VMs tab.
Protection Groups field Displays the total number of protection groups configured for this HX
Storage Cluster.
Click the field to display the list of protection groups and their associated
VMs in the Protection Groups section under the Local VMs or Remote
VMs tab.
The Replication page table provides four tabs: Local VMs, Remote VMs, Replication Activity, and
Replication Pairs. Each of these tabs provide replication protection configuration options.
Remote Cluster column Name of the corresponding remote cluster associated with the protected
virtual machine. This is the recovery cluster for the listed virtual machine.
Start Time column Displays the timestamp for the most recently started replication process.
End Time column Displays the timestamp for the most recently completed replication
process.
Protection Group column If the associated virtual machine belongs to a protection group, the
protection group name is listed. If it does not have a protection group,
the field displays None.
Direction column The direction of the replicated virtual machine. The direction is relative
to the local cluster. The cluster you are logged into is always the local
cluster. The options are:
• Incoming—The virtual machine resides on the remote cluster. It
is replicated from the remote cluster to the local cluster.
• Outgoing—The virtual machine resides on the local cluster. It is
replicated from the local cluster to the remote cluster.
Data Transferred column The size (in Bytes) of the virtual machine that is replicated. When the
replication is in progress, the amount completed is listed. When the
replication is complete, the amount of data transferred is listed in Bytes.
Remote Cluster Status column Status of the remote cluster. Options include: Online, Offline
Replications Outgoing column The number of virtual machines being replicated, transferring their data,
to the remote HX Storage Cluster from this local HX Storage Cluster.
VMs Incoming column The number of virtual machines configured for replication from the
remote HX Storage Cluster to this local HX Storage Cluster. Includes
the number of protection groups on the remote cluster.
Click the field to display the VM replications on the Virtual Machines
page.
Replications Incoming column The number of virtual machines being replicated, transferring their data,
from the remote HX Storage Cluster to this local HX Storage Cluster.
Mapped Datastores Pairs column The number of datastores used for replication on the local cluster.
Click the field to display the list of datastores on the Datastores page.
Create Replication Pair button This button is only available when a replication pair is not configured
for this local cluster. Click the button and complete the Create
Replication Pair dialog box.
Edit button Select the replication pair and click Edit to change the local or remote
datastores to use for replication. Complete the Edit Replication Pair
dialog box.
Delete button Select the replication pair and click Delete. Confirm the action.
Perform this operation when you want to remove the pairing of this local
cluster from the remote cluster.
Note All VMs on both clusters lose their replication
configuration. To apply protection to the VMs, you must
complete all the protection steps, including creating a new
replication pair.
Protection Groups sub table + Create Group button―Opens Create Protection Group dialog box.
Lists protection groups created on the local cluster. You can filter the
virtual machines by All Protected VMs or Standalone Protected VMs.
Displays the following protection group data:
• Group name
• Number of VMs in the group
• Status of VMs: Protected, Recovered, Recovering, Recovery Failed
• Replication interval time, tooltip shows the time of last replication
• To edit the group schedule, click the pen icon. To delete the protection
group, use the trash icon.
Pause button Pause outgoing replication stops all ongoing and new virtual machines
from being protected to the target site.
Virtual Machine Name column Lists the names of the virtual machines protected by replication in the HX
Storage Cluster.
Protection Status column Displays the current status of the virtual machine protected on this cluster:
• Recovering—The virtual machine is restoring from a replication
snapshot on the remote cluster.
VM State—Prepare Failover Started, Prepare Failover Completed
• Recovery Failed—The virtual machine failed to restore from a
replication snapshot on the remote cluster.
VM State—Prepare Failover Failed, Failover Failed
• Recovered—The virtual machine was recently restored from a
replication snapshot on the remote cluster.
VM State—Failover Completed
• Protecting—Reverse protect started for that virtual machine.
VM State—Prepare Reverse Protect Started, Prepare Reverse Protect
Completed, Reverse Protect Started
• Protection Failed—Reverse protect failed for the virtual machine.
VM State—Prepare Reverse Protect Failed, Reverse Protect Failed
• Protected—The virtual machine has at least one snapshot available
for recovery.
VM State—Success
• Active—Protection is configured, but no snapshot is available.
VM State—Active
• Exceed Interval—The last replication process took longer than the
configured interval to complete.
Last Protection Time column Displays the timestamp for the most recently completed replication process.
Direction column Displays the direction of the replicated virtual machine, relative to the local
cluster. The cluster you are logged into is always the local cluster.
• Incoming—The virtual machine resides on the remote cluster. It is
replicated from the remote cluster to the local cluster.
• Outgoing—The virtual machine resides on the local cluster. It is
replicated from the local cluster to the remote cluster.
Protection Group column If the associated virtual machine belongs to a protection group, the
protection group name is listed. If it does not have a protection group, the
field displays None.
Interval column Displays the length of time between the start of each replication. Select an
Interval time sufficient to complete each replication.
For example, an Interval time of Every 1 hour means that a replication
of the VM is started every hour.
Edit Schedule button Select an individually protected VM and click Edit Schedule to modify
the replication interval.
Remove from Group button Select one or more VMs from the same protection group and click Remove
from Group to remove the selected VMs from the group.
The selected VMs continue to be protected individually with the same
replication schedule as the protection group.
Click Remove from Protection Group to confirm.
Add to Group button Click to add virtual machines that are protected to a group. VM schedule
is changed to be a group schedule.
Protection Groups sub table + Create Group button―Opens Create Protection Group dialog box.
Lists protection groups that are created on the remote cluster. You can
filter the virtual machines by All Protected VMs or Standalone Protected
VMs.
Displays the following protection group data:
• Group name
• Number of VMs in the group
• Status of VMs: Protected, Recovered, Recovering, Recovery Failed
• Replication interval time, tooltip shows the time of last replication.
Virtual Machine Name column Displays the name of the virtual machine that is protected by replication
in the HX Storage Cluster.
Protection Status column Displays the current status of the virtual machine protection on this cluster:
• Recovering—The virtual machine is restoring from a replication
snapshot on the remote cluster.
VM State—Prepare Failover Started, Prepare Failover Completed
• Recovery Failed—The virtual machine failed to restore from a
replication snapshot on the remote cluster.
VM State—Prepare Failover Failed, Failover Failed
• Recovered—The virtual machine was recently restored from a
replication snapshot on the remote cluster.
VM State—Failover Completed
• Protecting—Reverse protect started for that virtual machine.
VM State—Prepare Reverse Protect Started, Prepare Reverse Protect
Completed, Reverse Protect Started
• Protection Failed—Reverse protect failed for the virtual machine.
VM State—Prepare Reverse Protect Failed, Reverse Protect Failed
• Protected—The virtual machine has at least one snapshot available
for recovery.
VM State—Success
• Active—Protection is configured, but no snapshot is available.
VM State—Active
• Exceed Interval—The last replication process took longer than the
configured interval to complete.
Last Protection Time column Displays the timestamp for the most recently completed replication process.
Direction column Displays the direction of the replicated virtual machine, relative to the local
cluster. The cluster that you are logged into is always the local cluster.
• Incoming—The virtual machine resides on the remote cluster. It is
replicated from the remote cluster to the local cluster.
• Outgoing—The virtual machine resides on the local cluster. It is
replicated from the local cluster to the remote cluster.
Protection Group column If the associated virtual machine belongs to a protection group, the
protection group name is listed. If it does not have a protection group, the
field displays None.
Interval column Displays the length of time between start of each replication. Select an
interval time sufficient to complete each replication.
For example, an interval time of Every 1 hour means that a replication
of the VM is started every hour.
Recover button Select a VM and click Recover, to take the most recent replication snapshot
of the VM and build a new VM on the local cluster. Ensure that the VM
on the remote cluster is unavailable.
Complete this step after Unprotect, for recovering a VM.
Migrate button Select a VM and click Migrate, to migrate the protected VM from the
source to the target.
Test Recovery button Select a VM and click Test Recovery, to take the most recent replication
snapshot of the VM and builds a new VM on the local cluster.
Note To perform this task, you must be logged in as a user with administrator privileges.
You are required to configure a replication network and a replication pair before protecting virtual machines.
Configure the replication network both on the local and remote cluster. Configure replication network on the
local cluster first, then log into the remote cluster and complete the configuration.
1. Select Replication > Configure Network.
2. Complete the following fields in the VLAN Configuration tab:
Create a new VLAN radio button Click this radio button to create a new VLAN.
Note If you are configuring replication network on edge
cluster, do not use the Create VLAN option. Use the
existing VLAN option and follow the same procedure.
VLAN ID field Click the up or down arrows to select a number for the VLAN ID or
type a number in the field.
This is separate from the HX Data Platform Management traffic
network and Data traffic network.
Important Be sure to use a different VLAN ID number for each
HX Storage Cluster in the replication pair.
Replication is between two HX Storage clusters. Each HX Storage
cluster requires a VLAN dedicated to the replication network.
For example, 3.
When a value is added, the default VLAN Name is updated to include
the additional identifier. The VLAN ID value does not affect a
manually entered VLAN name.
VLAN Name field This field is automatically populated with a default VLAN name
when the Create a new VLAN radio button is selected. The VLAN
ID is concatenated to the name.
For Stretched Cluster, provide Cisco UCS Manager credentials for primary and secondary FIs (site A
and site B).For normal cluster, provide Cisco UCS Manager credential for single FI.
Click Next.
3. Complete the following fields in the IP & Bandwidth Configuration tab:
IP & Bandwidth Configuration Tab
Gateway field Enter the gateway for use by the replication network. This is separate
from the HX Data Platform Management traffic network and Data
traffic network.
For example, 1.2.3.4.
IP Range field Enter a range of IP addresses for use by the replication network.
• The minimum number of IP addresses required is the number
of nodes in your HX Storage Cluster plus one more.
For example, if you have a 4 node HX Storage Cluster, enter a
range of at least 5 IP addresses.
• The from value must be lower than the to value.
For example, From 10.10.10.20 To 10.10.10.30.
• If you plan to add nodes to your cluster, include sufficient
number of IP addresses to cover any additional nodes. You can
add IP addresses at any time.
Add IP Range button Click to add the range of IP addresses entered in IP Range From and
To fields.
4. Click Configure.
Note To perform this task, you must be logged in as a user with administrator privileges.
Add available IP addresses to the configured replication network. There must be one IP address for every
node in the storage cluster, plus one more for management. If you expand your storage cluster, available IP
addresses are consumed.
1. Select Replication > Actions drop-down list > Edit Configuration.
2. In the Edit Network Configuration dialog box, you can edit the range of IPs to use and set the replication
bandwidth limit for replication traffic. The replication network subnet, gateway, and VLAN ID are
displayed for reference only and cannot be edited.
Edit Network Configuration Dialog Box
Gateway field The gateway that is configured for the replication network. This is
value cannot be edited.
Add IP Range field Click to add the range of IP addresses that are entered in IP Range
From and To fields.
Set replication bandwidth limit Enter the maximum network bandwidth that the replication network
check box (Optional) is allowed to consume for outbound traffic.
Valid Range: 10 to 10,000. The default is unlimited, which sets the
maximum network bandwidth to the total available replication network
bandwidth.
The replication bandwidth is used to copy DP snapshots from the
local HX Storage Cluster to the paired remote HX Storage Cluster.
Prepare group recovery stops the replication schedule for all the Virtual Machines in the protection group.
After the replication schedule is stopped for all the VMs, proceed to the Standalone VM tab and recover each
VM.
Resource Pool drop-down list Select a location for the new VM to be stored.
Power On/Off radio button Choose if the recovered VM must be powered on or left powered off after
it is created.
Map Networks Select to create a map between the source and target cluster networks.
• Source Network―the network on the cluster with the VM replication
snapshot.
• Target Network―the network on the cluster where the new VM is
created.
Resource Pool drop-down list Select a location for the test VM to be stored.
Power On/Off radio button Click a button. The recovered VM is powered on or left off after it is
created.
VM Name field Enter a new name for the created test VM.
Test Networks radio button Select which HX Storage Cluster network to use for transferring the
data from the replication snapshot.
Network options include:
• Storage Controller Data Network
• Storage Controller Management Network
• Storage Controller Replication Network
• VM Network
Map Networks radio button Select to create a map between the source and target cluster networks.
Source―the cluster with the VM replication snapshot.
Target―the cluster where the test VM is created.
Virtual Machine Name column Name of the virtual machine protected by replication in the HX Storage
Cluster.
Last Protection Time column Timestamp for when the most recent virtual machine replication process
started.
Direction column The direction of the replicated virtual machine. The direction is relative
to the local cluster. The cluster you are logged into is always the local
cluster. The options are:
• Incoming—The virtual machine resides on the remote cluster. It
is replicated from the remote cluster to the local cluster.
• Outgoing—The virtual machine resides on the local cluster. It is
replicated from the local cluster to the remote cluster.
Interval column The configured interval setting for replicating the virtual machine. To
change this, select the virtual machine row and click Edit Schedule.
Use VMware Tools to quiesce the To have HX Data Platform quiesce the virtual machines before taking
virtual machine check box the replication snapshot, click this checkbox.
This only applies to virtual machines with VMware Tools installed.
Protection Groups
Create Protection Group Dialog Box
Select Replication > Protection Groups > + New Group.
Create Protection Group Dialog Box
Protect virtual machines in this Select how often the virtual machines are to be replicated to the paired
group every field cluster.
The pull-down menu options are: 5 minutes, 15 minutes, 30 minutes, 1
hour, 90 minutes, 2 hours, 4 hours, 8 hours, 12 hours, and 24 hours. The
default value is 1 hour.
Start protecting the virtual Select this radio button if you want to set a specific time for the first
machines at radio button replication operation to start.
Before you start replication ensure:
• At least one virtual machine is added to the protection group.
• The scheduled start time is reached.
Cluster time zone and Current time on cluster are references to help
you to choose the appropriate replication start time. Start time is based
on the local cluster clock. For example:
10 hours, 3 minutes from now with Current time on cluster, 1:56:15PM,
means that the first replication occurs at 11:59:00PM.
The hours, minutes from now states when the first replication will occur.
This is updated when you change the time field setting.
Use VMware Tools to quiesce the Select this check box to take quiesced DP snapshots. Leaving the check
virtual machine check box box in an unchecked state will take crash consistent DP snapshots.
This only applies to virtual machines with VMware Tools installed.
Click Save Changes to save the interval and VMware Tools quiescence settings for the protection group.
View the Protection Groups tab to verify the interval frequency.
Add to an existing protection group drop-down list Select a protection group, click to add virtual
machines that are protected to a protection group.
Edit button Change the datastores assigned the replication pair name.
Select the replication pair and click Edit.
Delete button Remove the replication pair between the local and remote cluster.
Prerequisites: Remove all the dependencies: Remove protection from
all virtual machines. Remove datastore mapping.
Select the replication pair and click Delete.
Remote Cluster column Name of the remote cluster in this replication pair.
Remote Cluster Status column Displays the current status of the remote cluster. This is different that
the general cluster status. Options are:
• Online
• Offline
• Upgrading
• Out of Space
• Shutdown
• Unknown
VMs Outgoing column Number of virtual machines protected and number of protection groups
on the local cluster. Click number to display the outgoing Local VMs.
Replications Outgoing column Number of replication snapshots of the protected virtual machines being
replicated from the local cluster to the remote cluster.
VMs Incoming column Number of virtual machines protected and number of protection groups
on the remote cluster. Click number to display the outgoing Remote
VMs.
Replications Incoming column Number of snapshots of the protected virtual machines being replicated
from the remote cluster to the local cluster.
Mapped Datastore Pairs column Number of datastores mapped to this replication pair. Click number to
display the Datastores page.
Last Protection Time column Timestamp for when the most recent virtual machine replication process
started.
Protection Group column If the associated virtual machine belongs to a protection group, the
protection group name is listed. If it does not have a protection group,
the field displays None.
Interval column The configured interval setting for replicating the virtual machine. To
change the Interval, select the virtual machine row and click Edit
Schedule.
Prerequisites
• Create a datastore on both the local and remote cluster.
• Configure the replication network.
Name Page
Click Next.
User Name and Password fields Enter vCenter single sign-on or cluster specific administrator credentials
for the remote HX cluster.
Click Pair.
HX Data Platform verifies the remote HX Storage Cluster and assigns the replication pair name.
Note Virtual machines to be protected must reside on one of the datastores in the replication pair.
Note • The virtual machines to be protected must be on the datastores you select. Moving virtual machines from
the configured datastores for the replication pair, also removes protection from the virtual machines.
• Moving virtual machine to another paired datastore is supported. If the VMs are moved to unpaired
datastore, replication schedule fails.
To protect VMs using the HX Data Platform disaster recovery feature, click Native Protection and do the
following:
Remote Datastore column Pair the datastores between the HX Storage Clusters.
From the desired Local Datastore row, select a datastore from the
Remote Datastore pull-down menu. This selects both the remote and
local datastores in a single action.
Create New Replication Page > Map Datastores : Other DRO Protection
To protect VMs using SRM through disaster recovery orchestrator (DRO), click Other DRO Protection and
do the following:
Remote Datastore column Pair the datastores between the HX Storage Clusters.
From the desired Local Datastore row, select a datastore from the
Remote Datastore pull-down menu. This selects both the remote and
local datastores in a single action.
Protection Schedule column Choose the shedule for protecting all the VMs in the datastore.
Note The VMs in the datastores that are under other DRO, are protected by SRM.
Note If a new VM is added to the datastore protected by other DRO, the newly added VM is automatically protected
by Cisco HyperFlex. If a VM is added to the datastore protected using native DRO, you have to protect the
VM.
The replication pairs that are edited under Other DRO Protection, are exposed to SRM.
Note Changing the datastores used in a replication pair removes protection from all virtual machines on both the
local and remote clusters.
To protect VMs using SRM through disaster recovery orchestrator (DRO), click the Other DRO Protection
tab and do the following:
Protection Schedule column Choose the shedule for protecting all the VMs in the datastore.
3. Click Finish.
4. Re-protect your virtual machines. Select Virtual Machines > virtual_machines > Protect.
Click Run Test, to test cluster pairing between the clusters in the remote replication network.
c. Click Finish.
Password field Enter the administrator password for the remote HX Storage Cluster.
Note To perform this task, you must be logged in as a user with administrator privileges.
Test Virtual Machine Name (Optional) Use a common prefix to help identify the type and context
Prefix field of the resource.
Notification Setting radio button Select the type of notification prompt sent after a recovery event.
• Choose Normal Mode to get a confirmation prompt with
summary of configuration at the time of recovery, test recovery,
or migration.
Choose Silent Mode to not get a confirmation prompt.
Add Rule button Click to add an additional rule. The default value is 0.
Note This chapter covers the instructions to be executed in SRM Release 8.4. If you are using SRM Release 8.3,
8.2, 8.1, 6.5 or 6.1, refer to the VMware Site Recovery Manager Documentation on the VMware website
Processor At least two 2.0 GHz or higher Intel or AMD x86 processors.
Memory 4 GB
Disk Storage 5 GB
For information on HyperFlex Release support, see the SRA Compatibility Matrix section in this guide. For
additional information on supported platforms and databases, see the Compatibility Matrices for VMware
Site Recovery Manager for your release on the VMware website
(https://docs.vmware.com/en/Site-Recovery-Manager/index.html).
SRM Version
SRM Version
After you install SRM, it remains in evaluation mode until you install an SRM license key. After the evaluation
license expires, existing protection groups remain protected and you can recover them, but you cannot create
new protection groups or add virtual machines to an existing protection group until you obtain and assign a
valid SRM license key. You must obtain and assign SRM license keys after installing SRM. For more
information on the licensing, refer the Site Recovery Manager Licensing section in the Site Recovery Manager
Installation and Configuration guide on the VMware website.
Step 2 Copy the Windows installer of SRA to SRM Windows machines at the protected and recovery sites.
Step 3 Double-click the installer.
Step 4 Click Next on the welcome page of the installer.
Step 5 Accept the EULA and click Next.
Step 6 Click Finish.
Step 3 Log into the Site Recovery Manager Appliance Management Interface as admin.
Click on Storage Replication Adapter tab.
Pairing Sites
Before you can use SRM, you must connect the SRM server instances on the protected and recovery sites.
This is known as site pairing.
Step 1 Log into your protected site vCenter server using vSphere Web Client.
Step 2 In the home page, click Site Recovery plug-in and select sites.
Step 3 Enter the platform service controller address of the SRM remote site (recovery site). Leave the port number as default if
you don’t have any custom port for platform services controller (PSC). Click Next.
The system displays the vCenter server instance in which SRM server is registered on the remote site.
Step 4 Enter the vCenter server single sign-on (SSO) credentials and click Finish.
The connection process proceeds automatically after authentication with the correct privileges established.
For more information, refer the Site Recovery Manager Licensing section in the Site Recovery Manager Installation and
Configuration guide on the VMware website.
When the site pairing is established between protected and recovery site, you will be able to see the remote
site under Sites in the SRM interface. It displays the information about the primary site and paired site.
Configuring Mapping
Mappings allow you to specify how SRM maps virtual machine resources on the protected site to resources
on the recovery site. During recovery, when virtual machines start on the recovery site, the virtual machines
use the resources on the recovery site that you specify in the mappings.
Step 1 Log into the protected site vCenter using vSphere Web Client.
Step 2 Click SRM plugin. Select the protected site and click Configure Inventory Mappings.
Step 3 Click the following options and do the respective mappings:
Option Action
Create resource mappings Map resource pools, standalone hosts, vApps, or clusters on the protected site
to resource pools, standalone hosts, vApps, or clusters on the recovery site. You
can map any type of resource on one site to any type of resource on the other
site.
Note You cannot map individual hosts that are part of clusters to other
resource objects.
Create folder mappings Map datacenters or virtual machine folders on the protected site to datacenters
or virtual machine folders on the recovery site.
Create network mappings Map networks on the protected site to networks on the recovery site.
Configure placeholder datastores Select the datastore from the list to be used as placeholder datastore for the
production cluster.
Important You have to configure placeholder datastores on recovery site as
well.
For more information on inventory mappings, refer Site Recovery Manager Installation and Configuration guide on the
VMware website.
Execute the same procedure for the recovery site to refresh SRA in the recovery site.
What to do next
After refresh, the status of the site is set to OK with a green tick. This action refreshes SRA information,
allowing SRM to discover the SRAs.
Note If the datastore is the target of replication, it is safe to ignore any “Datastore unmounted” alerts.
Discovering Devices
This actions runs the discover devices on the selected array pair.
Step 1 In the vSphere Web Client home page, click Site Recovery plugin.
Step 2 Expand Inventories and click Recovery Plans.
Step 3 Click Objects.
Step 4 Click Create Recovery Plan.
Step 5 Enter the name and description for the recovery plan and select the location for the recovery plan. Click Next.
Step 6 Select the recovery site in which the VMs will be recovered as part of the recovery plan. Click Next.
Step 7 Select the list of protection groups which will be included as part of the recovery plan. Click Next.
Step 8 Specify the test networks to use while running tests of the recovery plan.
You can select the test networks if you have created for the test purposes or you can leave the default auto created test
network called isolated network.
Note After failover, the source datastore is unmounted on HX data platform, due to which, the datastore to VM
mapping returns empty list. So, we cannot unpair the datastores. To unpair the datastore, you have to bring
the VMs on the source datastore to the Active state by invoking the Reprotect operation.
Step 1 In the vSphere Web Client home page, click Site Recovery plugin.
Step 2 Expand Inventories and click Recovery Plans.
Step 3 Select the recovery plan and click Monitor.
Step 4 Click the green play button to run the test recovery.
Caution Be cautious in selecting the test recovery option.
The recovery process runs in the test mode to recover the virtual machines in a test environment on the recovery site.
You have the storage option to choose whether to replicate the recent changes to the recovery site or not. This process
may take several minutes to complete the replication of recent changes to the recovery site. According to your requirement,
you can choose to run the test recovery with recent changes or not.
Step 6 Review the settings and click Finish to initiate the SRM test recovery.
The series of recovery steps involved as part of this recovery plans are displayed with its progress percentage.
Note Once the plan is executed, the datastore on the recovery site will be unmounted and it will stay unmounted.
During this time, you can ignore alerts on the HyperFlex recovery cluster as this is expected behavior.
On completion of the test recovery, the status of the recovery plan is changed to Test Complete and status
of each recovery steps is changed to Success. Also, the virtual machines that are part of the protection group
as part of this recovery plan are powered on in both protected and recovery site without causing any interruption
to the production virtual machines.
After successful execution of test recovery, you can run a recovery plan cleanup to return virtual machines
to their initial state at which it was before running the recovery plan test. The recovery plan status is reset to
Ready state. You can also run a forced cleanup when you experience errors during execution of a recovery
plan. For more information, see Testing the Recovery Plan Cleanup Task, on page 271.
Step 1 In the vSphere Web Client home page, click Site Recovery plugin.
Step 2 Expand Inventories and click Recovery Plans.
Step 3 Select the recovery plan and click Monitor.
Step 4 Click the brush icon to cleanup the test recovery.
Running a cleanup operation on this plan will remove the test environment and reset the plan to ready state.
You can monitor the progress of the test recovery cleanup process. The recovery plan cleanup task returns
virtual machines to their initial state as it was before running the recovery plan test. After cleanup, the virtual
machines recovered as part of the test recovery in the recovered site are in the powered off state. The recovery
plan status is reset to Ready state. The recovery plan is ready to test again.
Executing a Recovery
Step 1 In the vSphere Web Client home page, click Site Recovery plugin.
Step 2 Expand Inventories and click Recovery Plans.
On completion of recovery process, the status of the recovery plan is set to Recovery Complete and the status
of each recovery steps is set to Success. The virtual machines that are part of the protection group of the
recovery plan are powered on in both the protected and recovery sites, without interrupting the production
virtual machines.
Step 1 In the vSphere Web Client home page, click Site Recovery plugin.
Step 2 Expand Inventories and click Recovery Plans.
Step 3 Select the recovery plan and click Monitor.
Step 4 Click Reprotect.
Step 5 Review the settings set for initiating the SRM reprotect and click Finish.
The progress of each reprotect steps is displayed. On completion of reprotection, the status of the recovery
plan is set to Reprotect Successful and the status of each reprotect steps is set to Success. The virtual machines
are protected, and the test recovery and recovery options are enabled.
Note When you execute the Reprotect task, the datastore replication direction does not change to reflect reverse
replication. We recommend you to view the VM direction at Replication > Replication Pairs > Other DRO
Protection to confirm replication direction.
Troubleshooting SRA
Troubleshooting: Using Windows SRM Log Files
To identify the cause of any problems that you may encounter during the day-to-day running of SRM, you
have to collect and review SRM log files. To get VMware support, you have to send the log files to them.
The VMware SRM logs from SRM server are available at: C:\ProgramData\VMware\VMware vCenter Site
Recovery Manager\Logs
SRA specific logs are available under the SRA folder at: C:\ProgramData\VMware\VMware vCenter Site
Recovery Manager\Logs\SRAs\
To download SRM log files from the SRM interface, follow this procedure:
1. In the vSphere Web Client home page, click Site Recovery > Sites.
2. Select a site for which the log file has to be collected.
3. From the Actions menu, select Export SRM Log.
Alternatively, right-click the site and select Export SRM Log.
4. In the Export SRM Log wizard, click Generate Log.
5. Click Download Log.
• If a VM is renamed and a VMDK is added subsequently, the new VMDK is created at [sourceDs]
newVm/newVm.vmdk. HyperFlex recovers this VMDK with the earlier name. In such cases, recovery
can fail if the VMDK is explicitly referenced in the virtualmachine name.vmx file by its path. The data
is recovered accurately but there could be problems with registering the VM to the Virtual Center. Correct
this error by updating the virtualmachine name.vmx file name with the new path.
HX Connect Can perform most Not valid Can perform most Can only view
HX tasks. HX tasks. monitoring
information.
local/ prefix A preferred user.
required for login. Cannot perform HX
Example: tasks.
local/admin A preferred user.
Storage Controller Can perform most Can perform most vc- prefix required Cannot perform HX
VM with hxcli HX tasks. HX tasks. for login. Example: tasks.
command line
vc-hx_admin vc- prefix required
for login. Example:
vc-hx_readonly
HX Data Platform Can perform most Can perform most Can perform most Can only view
Plug-in through HX tasks. HX tasks. HX tasks. vCenter information.
vCenter
A vCenter SSO user. Cannot view HX
Data Platform
Plug-in .
A vCenter SSO user.
HX REST API Can perform most Can perform most Not valid Not valid
HX tasks. HX tasks.
local/ prefix local/ prefix
required for login. required for login.
Example: Example:
local/admin local/root
• Role―Defines an authority level. An application function may be performed by one or more roles.
Examples: Administrator, Virtual Machine Administrator, or Resource Pool Administrator. Role is
assigned to a given Identity.
This file contains other information as well. To filter for audit events, use a script to filter for the word,
Audit.
What to do next
Add the user to an RBAC role group. See Assigning Users Privileges, on page 278.
Step 1 From the Cisco vSphere Web Client, select Navigator Home > Administration > Global Permissions > Manage.
Step 2 Click Add (+) icon to assign roles.
Step 3 Select an Assigned Role.
In the Global Permission Root - Add Permission dialog box, select from the Assigned Role drop down menu. Choose
one:
• Administrator
• Read only
Note The iSCSI features are supported in Cisco HyperFlex Release 4.5(x) and later.
• HyperFlex iSCSI Target Service Overview and Supported Use Cases, on page 279
• HyperFlex iSCSI Best Practices , on page 280
• iSCSI Configuration Overview, on page 280
• iSCSI Network Page, on page 281
• iSCSI Initiator Group, on page 283
• iSCSI Target Page, on page 286
• iSCSI LUN Page, on page 289
• Configuring an iSCSI Initiator (Windows), on page 291
• Configuring an iSCSI Initiator (Linux), on page 292
• Cloning an iSCSI LUN, on page 292
Initiators are currently supported on Windows Server 2016, Windows Server 2019, Ubuntu 18.04 and 20.04,
Oracle Linux 8.2, Oracle Linux 7, Red Hat Enterprise Linux 8.2, Red Hat Enterprise Linux 7.
Note The HyperFlex iSCSI Target Service is not supported on Stretched clusters.
UI Element Description
Caution An ip address range is required for the new VLAN configuration. This range should not already exist in the
cluster. Not adhering to this requirement can result in a cluster outage.
Note After you confirm creation of your iSCSI network, you cannot change some iSCSI network parameters without
TAC assistance.
Note If you configured the hx-storage-data network with 1500 MTU (no Jumbo frames), but you want to utilize
Jumbo frames (as recommended for the iSCSI network), you will need to manually edit the hx-storage-data
network vswitch on all ESXi hosts in the HyperFlex cluster to 9000 MTU.
UI Element Description
IP Range Enter a valid IP Range (this range should not include the iSCSI Storage
IP)
iSCSI Storage IP Enter a valid IP Address for iSCSI Storage (this IP address should not
be an address used in the IP Range field)
Set non default MTU Checkbox to enable setting MTU (Message Transport Unit) manually.
The MTU defines the maximum size of a network frame that you can
send in a single data transmission across the network. The default MTU
size is 9000.
To disable Jumbo Frames, click the Set non default MTU checkbox,
and then use the pull-down to change the value to 1500.
Note If any of the initiators are crossing a router, the router will
need to allow Jumbo Frames.
UI Element Description
Note The option to delete an iSCSI network and retain objects you have configured is also available using the
"hxcli iscsi network delete" command or API. For more information, see the CLI Reference Guide.
UI Element Description
Step 2 Enter a name for your Initiator Group in the Name field.
Step 3 Enter the Initiator IQN in the field provided. If you do not know the Initiator IQN, you can fetch it from the server:
a) For Windows, log into the Windows machine. Go to Server Manager and click on iSCSI Initiator. Go to the
Configuration tab. The Initiator IQN is identified in the Initiator Name field.
b) For Linux, enter the command: "sudo cat /etc/iscsi/initiatorname.iscsi". The Initiator IQN is identified in
the Initiator Name field.
The initiator name is in the file and can be changed with the appropriate rights.
For more information, see the CLI Guide for your release.
Step 5 Repeat the above steps to add multiple IQNs to participate in the same Initiator Group.
Step 6 Click Create Initiator Group.
Step 1 Click on the Edit (pencil) icon next to the Create button.
The Edit Initiator Group window appears.
Note You cannot delete an iSCSI Initiator Group when linked with a Target.
Step 1 Click on the Delete (X) icon next to the Create Initiator Group button. The Delete Initiator Group window appears.
Step 2 Click Delete; or Cancel to cancel your changes.
Step 1 Go to the Targets tab, and select the name of the Target you want to link an Initiator Group.
Step 2 From the Linked Initiator Groups tab, click in the checkbox to Link.
The Link Initiator Groups window appears.
What to do next
After you have linked Initiator Groups with Targets;
• Creating an iSCSI LUN
• Configuring an iSCSI Initiator (Windows)
• Cloning an iSCSI LUN
Step 1 Go to the Targets tab, and select the name of the Target you want to unlink an Initiator Group.
Step 2 From the Link Initiator Groups tab, click in the checkbox to select the Initiator Group(s) to unlink to the Target.
Step 3 Click on the Unlink Initiator Group button.
Step 4 Click Unlink Initiator Group(s) to proceed or Cancel.
Target Data
The following information is required to create a Target.
UI Element Description
UI Element Description
Total LUN Capacity Total storage capacity of LUN used and available (in Tb and in Gb).
LUNs Tab allowing you to create, edit, clone, delete and view LUNs for the
Target.
Linked Initiator Groups Tab allowing you to create and view LUNs for the Target.
Note CHAP based authentication is not supported during iSCSI Discovery phase.
HXDP supports one-way CHAP authentication.
Note When using CHAP and booting from SAN (iSCSI LUN), you must set the CHAP user/password at the top
initiator area in UCS (for example, in UCS manager, navigate to Boot Policy > Set iSCSI Boot Parameters
> Authentication Profile).
What to do next
After you have created an iSCSI Target:
• Linking iSCSI Initiator Group with Targets
Step 1 Click on the Edit (pencil) icon next to the Create button.
The Edit Target window appears.
Step 1 Click on the Delete (X) icon next to the Create Target button.
The Delete Target window appears.
Step 1 Go to the Initiator Groups tab, and select the name of the Initiator Group you want to link to a Target.
Step 2 From the Linked Targets tab, click on the Link button.
The Link Target window appears.
The linked targets you selected appear in the Linked Targets tab.
What to do next
After you have linked iSCSI Targets.
• Creating an iSCSI LUN
• Configuring an iSCSI Initiator (Windows)
• Cloning an iSCSI LUN
Step 1 Go to the Initiator Groups tab, and select the name of the Initiator Group you want to unlink targets.
Step 2 From the Linked Targets tab, click in the checkbox to select the Target(s) that you want to unlink.
Step 3 Click on the Unlink Target button.
Step 4 Click Unlink; or click Cancel to cancel your changes.
UI Element Description
UI Element Description
Step 1 Go to the Targets tab, and select the name of the Target you want to create a LUN.
Step 2 Click in the Create LUN checkbox.
The Create LUN window appears.
Note There is a limit on the number of LUNs that you can expose per target. On Linux systems, the limit is 255
LUNs per target. On Windows systems, the limit is 254 LUNs per target.
Note You can create volumes that are protected by CHAP. The limit that you can create with one storage class for
each target is 255 volumes (Persistent Volume Claims).
What to do next
After you have created an ISCSI LUN, you can Configuring an iSCSI Initiator (Windows)
.
Step 1 Go to the Targets tab, and select the name of the Target you want to edit LUNs.
Step 2 From the LUNs tab, click in the checkbox to select the LUN you want to edit.
Step 3 Click on the Edit icon next to the Create LUN button. The Edit LUN window appears.
Step 4 Edit data for the LUN.
Note The maximum LUN size is 64 TB.
Step 1 Go to the Targets tab, and select the name of the Target you want to delete LUNs.
Step 2 From the LUNs tab, click in the checkbox to select the LUN you want to delete.
Step 3 Click on the Delete (X) icon next to the Clone LUN button. The Delete LUN window appears.
Step 4 Click Delete; or click Cancel to cancel your changes.
Note HXDP supports one-way CHAP authentication. Two-way CHAP authentication is not supported.
Step 1 Log into the Windows machine that you want to configure as an iSCSI Initiator.
Step 2 Go to Server Manager, and click on iSCSI Initiator. Click Yes to continue.
Step 3 Enter the Target’s Hostname or IP address in the Target tab, and then click on Quick Connect.
Discovered Targets appear as "HX Cluster IP(CIP)" and display the Target IQN and status of the Targets.
If there are no issues with the configuration, the status is updated to indicate "Connected".
Step 8 Verify that the iSCSI LUN(s) is attached in the Disk Management tool.
You can now initialize, online and create a volume on the iSCSI LUN(s).
Note With lsblk –scsi, you can verify which device is the target. You can now partition it with gdisk and format
the HX iSCSI drive.
Step 1 Go to the Targets tab, and select the name of the Target you want to clone LUNs.
Step 2 From the LUNs tab, click in the checkbox(es) to select the LUN(s) you want to clone.
Step 3 Click on the Clone LUN icon next to the Edit LUN button. The Clone LUNs window appears.
Step 4 Click in the Application Consistent checkbox to enable. Specify the administrator account (local or AD) username and
password for the Windows machine to verify and authenticate the VSS user.
Note If you do not enable the Application Consistent checkbox, the iSCSI Clone LUN(s) will be Crash Consistent.
Step 5 Enter a name for the new destination target in the New destination target name field.
Step 6 If you want to enable CHAP authentication, select the check box to Enable CHAP Authentication. For each source
LUN(s), enter the destination LUN(s) name in the Destination LUN Name field.
Note CHAP based authentication is not supported during iSCSI Discovery phase.
What to do next
To access cloned LUN(s), link the destination Target with an Initiator Group, and then discover the LUN(s)
by refreshing the iSCSI target from the Initiator window. Use the HxWindowsAgentUtils.exe to change
the disk/volume properties for the destination LUN(s).
Note Agent logs and files extracted from HxWindowsAgentIscsiClone-v4.5.1a-39020.exe appear with build
number 4.5.1a.38547 in properties of the file. This is a version display issue with no impact on functionality
and can be disregarded.
This installs the HyperFlex Windows Agent and HyperFlex VSS Hardware Provider services. Other installatoin
notes are as follows:
• You can view the HyperFlex Windows Agent and HyperFlex VSS Hardware Provider services in Windows
services. The HyperFlex Windows Agent should be in running state and the HyperFlex VSS Hardware
Provider is in stopped state. The HyperFlex VSS Hardware Provider service is started when you request
to clone or backup a LUN.
• You can view MSI installation and other install details in %appdata%\HxWinAgentMsiInstall.log.
• In the installation directory, you may notice some dependent dll’s. Do not delete or update these
dependencies. If deleted, you will need to use the repair option in the installer to restore.
• View service logs in C:\HxWindowsAgent\Logs\ HxAgentService_<DateTime>.log. The Windows
Registry is in location : HKEY_LOCAL_MACHINE\SOFTWARE\HyperFlex. This entry will have the
location of Service logs.
• To verify the Agent version: Go to the Installation directory, right-click on HxWinAgentService.exe and
select Properties. Go to the Details tab and verify the Product Version.
• You can view installation and operation events on the service in the Event viewer with Source as
“HxVssHardwareProvider” and “HxWindowsAgent”. Inbound rules are set with the name “Hx Windows
Agent” on port 10152 and enabled for all IP Addresses in Windows firewall .
Note Agent logs and files extracted from HxWindowsAgentIscsiClone-v4.5.1a-39020.exe appear with build
number 4.5.1a.38547 in properties of the file. This is a version display issue with no impact on functionality
and can be disregarded.
Note Agent logs and files extracted from HxWindowsAgentIscsiClone-v4.5.1a-39020.exe appear with build
number 4.5.1a.39020 in properties of the file. This is a version display issue with no impact on functionality
and can be disregarded.
• When done, the uninstaller removes “HyperFlex Windows Agent” and “HyperFlex VSS Hardware
Provider” services.
• The uninstaller deletes inbound rules with the name “Hx Windows Agent” having port 10152 from
Windows Firewall.
• The uninstaller does not remove Microsoft .NET framework 4.5 (version : 4.5.50709 or later), Microsoft
Visual C++ 2017 Redistributable (x64) (version: 14.10.25017 or later), Microsoft Visual C++ 2017
Redistributable (x86) (version: 14.10.25017 or later) programs.
• The uninstaller does not delete files and folders from the location: C:\HxWindowsAgent\Logs\
HxAgentService_<DateTime>.log and Installation directory\HxCAInstallLogMsi.txt”.
Note This command accepts input parameters including the Windows IP, username and password required to pull
the logs.
The logs are transferred to the following location on the Controller VM machine:
:/var/log/springpath/<WindowsIP> in a file called "HXLogs.zip". The HXLogs.zip file
includes the following information: HX Windows Agent logs, Disk Details from the HxDiskInfo.log, and
system information from HxSystem.log.
Step 1 Go to diskmgmt.msc and right-click on the required Disk <Disk ID> as "Online".
Step 2 Open a command prompt as "Administrator". Run diskpart.exe.
Step 3 Run the command: List Disk.
Step 4 Run the command: Select Disk <Disk ID> (Select the corresponding disk accordingly).
Step 5 Run the command: Detail Disk.
Step 6 If the disk attribute "Read-only" is "Yes", set as "No": "attributes disk clear readonly"
Step 7 Select Volume <Volume ID> (select volume shown as part of detail disk)
Step 8 Run the following commands:
• attributes volume clear READONLY
You should now have a disk with volume accessible and the volume label with read/write permissions.
Note VMware local plugin architecture support is limited to vSphere versions 6.5, 6.7, and 7.0. For more information,
see the vSphere Client Local plugins are deprecated (87880) article on the VMware site.
To get started, use the following table to determine the plugin type for your deployment. Click the link in the
Plugin Type column to advance to the full documentation.
vCenter Plugin Version Cisco HyperFlex Release vCenter/ESXi version Plugin Type
3.0.0 and later 4.0(2f) and later 6.7 U3 and later Cisco HyperFlex
Remote Plugin for
VMware vCenter
1
2.0.0, 2.1.0 and 2.2.0 4.0(2e) and earlier Up to 7.0 U3 Cisco HyperFlex
HTML5 Plugin for
VMware vCenter
1
In alignment with the VMware deprecation of local plugins starting with vSphere 8.0, remote plugin is
the only supported version for HX 4.0(2f) and later
Note Cisco HyperFlex Release 4.5(x) - Perform install, uninstall, and upgrade operations by running
install_vc_plugin on your shell.
Rename Clusters - ✓ ✓
2
Disks View ✓ ✓ ✓
Nodes View ✓ ✓ ✓
HX Datastore Management ✓ ✓ ✓
Network Management - ✓ ✓
iSCSI Management - ✓ ✓
3
Manage Tasks - ✓ ✓
Schedule Snapshot - ✓ ✓
4
Linked Mode - - ✓
2
Requires HXDP Release 4.5(x) or later.
3
Requires HXDP Release 4.5(x) or later.
4
Requires HXDP Release 4.5(x) or later.
5
Requires HXDP Release 5.0(x) or later.
HX Controller VM Password -
-h, --help Optional Shows the help message relative to the listed
command and exits.
Step 1 Download the Cisco HyperFlex HTML plugin for VMware vCenter from the Cisco Software Download site.
Step 2 Copy the HyperFlex-VC-HTML-Plugin-2.2.0.zip file into a temporary directory in one of the controller VMs and unzip.
a) The file transfer may be completed by using sftp cli or any file transfer app such as winscp or filezilla.
To use sftp transfer via a file transfer app copy the file to the /tmp folder on SCVM, using HX admin account.
b) SSH to that SCVM and login with admin account.
c) Change to the /tmp directory using the command "cd /tmp".
Example:
"cd /tmp"
d) Unzip the plugin file HyperFlex-VC-HTML-Plugin-2.2.0.zip using the command unzip.
Example:
unzip HyperFlex-VC-HTML-Plugin-2.2.0.zip
Step 4 Log on to vCenter and a blue banner message appears to confirm that the new plugin is installed.
Step 5 Log out and log in again to vCenter to see the Cisco HyperFlex menus for HTML5 plugin.
Verifying the Cisco HyperFlex HTML5 Plugin Installation from the vSphere
Client
To verify Cisco HyperFlex plugin installation from the vSphere Client UI.
Launch the vSphere client, select Menu > Administration > Solutions > Client Plug-Ins
Step 1 Execute the uninstall command install_vc_plugin -u on the shell and enter the following credentials:
• vCenter FQDN/IP address
• Administrator username and password for the vCenter server
Step 1 Download the Cisco HyperFlex HTML plugin for VMware vCenter from the Cisco Software Download site.
Step 2 Copy the HyperFlex-VC-HTML-Plugin-2.2.x.zip file into a temporary directory in one of the controller VMs and unzip.
a) The file transfer may be completed by using sftp cli or any file transfer app such as winscp or filezilla.
To use sftp transfer via a file transfer app copy the file to the /tmp folder on SCVM, using HX admin account.
b) SSH to that SCVM and login with admin account.
c) Change to the /tmp directory "cd /tmp"
Step 4 Select Y to continue the Upgrade process with controller root and admin password.
Step 5 Logout and log in again into vCenter to see Cisco HyperFlex listed in the vCenter menus.
The Cisco HyperFlex HTML5 Plugin has functionality that is common throughout the plugin. This section
describes the icons and their usage.
Icon Usage
Icon Usage
Cluster Management
Managing Users and Access to HX Clusters
The vCenter plugin requires the user to have administrator privileges. You can create a user and assign
administrator role to that user from Permissions tab on cluster level.
To manage users and access to HX clusters, assign the No Access Role to all the clusters for that user.
Note Administrative Privileges are required for managing users and roles.
Step 4 If you have added new HX Cluster(s) to the vCenter server and they are not appearing in the cluster list, Click the Rescan
icon on top of the cluster list grid to reload the cluster list from HyperFlex. The Scanning Clusters icon indicates that
the cluster table is still populating. The icon disappears when the cluster list is complete.
Rename Cluster
The rename cluster was introduced in HX Release 4.5. To rename a cluster, perform the following steps:
Uptime Length of time that the cluster has been up and running
Deployment Type Type of cluster deployment. Valid options are Standard and Edge.
License Type6 Type of License. Valid options include: Evaluation, Standard, and Enterprise.
Note New users enjoy a 90-day grace period to register the license. After
90-days a "License is not in compliance" appears and product
features are limited.
License Status7 License status. Status include: In-compliance, Out of Compliance, and License
expires in x days, Cluster not registered with Cisco Licensing. Cluster not
registered with Cisco Licensing.
Storage Capacity Bar A graphical representation of the percentage of total storage used. Hover over
the bar to view the amount of storage used.
Total VMs Bar A graphical representation of the total number of VMs in the cluster.
Total Datastore Bar Total number of datastores connected to the cluster. Hover over the bar to view
the number of datastores mounted and unmounted.
6
Added in HX Release 5.0(x)
7
Added in HX Release 5.0(x)
a) The summary view includes four portlets with additional details about the cluster: Status, Network Details, Capacity
and Performance.
Use the arrows to collapse and expand the portlet contents.
General Usage
• Performance charts are visible when the License Status is In-compliance8.
• Click on the Time Interval list to select the length of time viewed in the performance chart.
• Hover over the chart line to display totals for a specific time.
• Click the refresh arrow to refresh the view.
• The Scanning Cluster icon indicates that the cluster table is still populating. The icon disappears when the
cluster list is complete.
• To change the timezone, click the current time interval, complete the Time Range pop-up, and click OK. The
time seen reflects the browser time.
Step 1 Starting on the vSphere web client Summary page, click the discovered HX cluster name to view its summary.
Step 2 In the License Type summary, click the Register Now link. The "Smart Software Licensing Product Registration window
appears.
Step 3 Type the Product Instance Registration Token on the field provided
Note If your registration token is not available, generate a new one by clicking on the Cisco Smart Software
Manager Link and follow the prompts.
General Usage:
• Click on the Time Interval list to select the length of time viewed in the performance chart.
Note The Alarms chart appears with time interval selections of 1 Month or less.
• Use the drop-down Cluster list on the top right to navigate between the clusters.
• Hover over the chart line to display totals for a specific time.
• Click the refresh arrow to refresh the view.
• To change the timezone, click the current time interval, complete the Time Range pop-up, and click OK.
The time seen reflects the browser time.
Disks
To view the Disks details page, perform the following steps:
Usage How the disk is being used. Valid values include: Caching, Persistence, System
Step 5 (Optional) Locate a physical server using the Turn On LED button
Note Beginning with HX Release 5.0(1a), The Turn On/Off LED button functionality requires the license status
to be In-compliance.
a) Click the Turn On LED button to illuminate the LED light on the associated physical server.
b) When finished, click the Turn Off LED button to turn the LED light off.
Nodes
To view node details specific to the Cluster, Host, and VMs, perform the following steps:
Step 5 Click the Node name that you want to view details. The Node Summary portlet appear below the Nodes list.
a) The Node Summary view includes two portlets with additional details about the node: Hyperconverged Nodes and
Disk Overview.
Use the arrows to collapse and expand the portlet contents.
Disk Overview Notes the number of slots in use and the number that are empty.
Legend Legend for icons and colors used in the disk graphics.
Disk graphic Hover over a disk to display details for that disk.
Details include:
• Slot Number and type of usage
• Disk Status: Claimed or Unclaimed
• Capacity
• Storage Usage as a percentage.
• Disk Drive Interface
• Version
Note If you have a 3- or 4-node cluster, only one node will go into maintenance mode.
Network
Network: Create New VLAN
The Network page allows users to create a VLAN without going through UCS. To create a VLAN from the
vSphere client perform the following steps:
VLAN ID To create one VLAN, enter a single numeric ID. A VLAN ID can:
• Be between 1 and 3967
• Be between 4049 and 4093
VLAN Name This name can be between 1 and 32 alphanumeric characters. You cannot use
spaces or any special characters, and you cannot change this name after the
object has been saved
VLAN Configuration Click in the checkbox to Create a new VLAN (Recommended) or Select an
existing VLAN.
To Create a new VLAN, you need to specify the following: VLAN ID, VLAN
Name, UCS Manager host IP or FQDN, Username (Username for authentication
with UCS), Password (Password for authentication with UCS).
To Select an existing VLAN, you need to specify the VLAN ID.
Note To configure the VLAN manually in UCS-M, use the create VLAN
menu option. In the Create VLANs window, leave the checkboxes
as is. In the vNIC templates for HX, attach the VLAN to "vNIC
Template storage-data-a" and "vNIC Template storage-data-b".
This configuration is non-disruptive.
Set Non-Default MTU Checkbox to enable setting MTU (Message Transport Unit) manually. The
MTU defines the maximum size of a network frame that you can send in a single
data transmission across the network. The default MTU size is 9000.
To disable Jumbo Frames, click the Set non default MTU checkbox, and then
use the pull-down to change the value to 1500.
Note If any of the initiators are crossing a router, the router will need
to allow Jumbo Frames.
iSCSI
iSCSI: Targets
After the iSCSI network is created, the iSCSI page appears in the list of navigation tabs. The default view is
Targets, use the Create, Edit, Delete, and Clone LUN buttons to manage iSCSI targets.
Note The iSCSI page only appears in navigaton tabs of clusters with a configured iSCSI network.
Target Name Name of the iSCSI storage resource on the iSCSI server.
CHAP Authentication Authentication scheme that uses a shared secret and a challenge message to
validate the identity of remote clients.
Step 5 Select a Target from the list to display all the LUNs associated with the selected target. in the portlet below the Target
list. Use the Create, Edit, Clone LUN, and Delete buttons to create, edit, clone or delete LUNs in the selected target.
Related Topics
iSCSI LUN Page, on page 289
Creating an iSCSI LUN, on page 290
Editing an iSCSI LUN, on page 291
Cloning an iSCSI LUN, on page 292
Deleting an iSCSI LUN, on page 291
View iSCSI and Datastore Summary from the Configure Tab, on page 344
Initiators Groups List of groups that specify which hosts can access specified LUNs on the cluster.
Step 4 Select an Initiators Group from the list to display the list of Initiators in the details portlet below the list.
Step 5 Click the Initiators button to view the individual Initiators in a group
Step 6 Click the Linked Targets button to view the targets associated with the selected group.
Step 7 Use the Link, and Unlink buttons to link and unlink targets to groups.
Related Topics
iSCSI Initiator Group, on page 283
Creating an iSCSI Initiator Group, on page 284
Editing an iSCSI Initiator Group, on page 285
Deleting an iSCSI Initiator Group, on page 285
Linking iSCSI Initiator Group with Targets, on page 285
Unlinking an iSCSI Initiator Group, on page 286
View iSCSI and Datastore Summary from the Configure Tab, on page 344
iSCSI: LUNs
Use the LUNS Button to Create, Edit, Delete, and Clone LUN buttons to manage LUNs.
Create LUN
Edit LUN
Clone LUN
Step 4 Select a LUN on the list to display the Details Portlet and Perforrmance Charts below the LUN list.
Related Topics
View iSCSI and Datastore Summary from the Configure Tab, on page 344
HX Datastore Management
Managing Datastores
The Datastore page allows users to view datastore details, create, edit, mount, unmount or delete datastores
on a Cluster.
Note Beginning with HX Release 5.0(1a), the Create and Delete Datastore buttons are enabled when the license
status is In-compliance.
Step 5 Click on a Datastore name in the table to view additional details about the Datastore.
SUMMARY is auto-selected and the Details and Trends portlets appear below the table.
General Usage:
• Click on the Time Interval list to select the length of time viewed in the performance chart.
• Hover over the chart line to display totals for a specific time.
• Click the refresh arrow to refresh the view.
• The Scanning Cluster icon indicates that the cluster table is still populating. The icon disappears when the cluster
list is complete.
• To change the timezone, click the current time interval, complete the Time Range pop-up, and click OK. The time
seen reflects the browser time.
Host Name Displays the IP address of the host for the selected datastore.
Related Topics
View iSCSI and Datastore Summary from the Configure Tab, on page 344
Edit Datastore
To edit an existing datastore:
Note You can only change the name of a datastore after it has been Unmounted.
Step 4 Click OK to continue with the Mount (Unmount) action, or click Cancel to exit the Mount (or Unmount) Datastore
window. The datastore status is changed from Mounted to Unmounted or Unmounted to Mounted.
Delete Datastore
To delete a datastore:
Step 4 Click OK to continue with the delete action, or Cancel to exit the Delete Datastore window.
The selected Datastore is deleted from the Datastore table list.
VMs
To view VM details specific to the Cluster, Host, and VMs, perform the following steps:
VMs Summary Usage diagram of user VMs in use. Hover over to view the number of VMs
running, suspended, and off.
Total VMs: The total count of all user VMs.
Note Controller VMs are not included in the summary.
VMs Storage The sum of all user VMs storage. Total storage capacity for all user VMs appears
above image. Hover over the graphic to view the current amount of storage
being consumed.
VMs Memory Amount of point-in-time memory. Total memory capacity is listed, hover over
the graphic to view the current amount of memory used.
CPU Total vCPU Count - Total CPU count for all VMs in the cluster.
CPU Usage - Number of cycles per second a given CPU is using.
Alarms Displays Alarms for the VMs during the last week (7days).
Click View to navigate to the Alarms Details view.
• Triggered Time-Date and time the alarm occurred.
• Description-Alarm Description.
Events Displays Events for theVMs during the last week (7days).
Click View to navigate to the Events Details view.
• DateTime - Date and time the event occurred.
• Description- Event description.
Time list Specify the measurement of time to show the top 15 VMs. List options include:
1 Hour, 1 Day, or 1 Week.
Metrics List Select the metric type used to populate the table. Options include: CPU, Memory,
Disk Latency, Network, and Space.
• CPU, Memory, Disk Latency, Network - only report metrics for VMs in
a running condition. VMs that are switched off are not included.
• Space - counts all VMs regardless of their state.
Name VM Name - Clicking on the VM name redirects users to the graph or monitoring
page of VM being viewed in vCenter.
State Current state of the VM. Valid values are Running, Off and Suspended.
Memory Amount of host physical memory consumed for backing up guest physical
memory pages.
Disk Latency Highest latency value across all disks used by the host
Network Throughput Network utilization during the interval (combined transmit and receive rates).
View Metrics Link to view CPU, Memory, Disk Latency, and Network Throughput
performance tables for that specified VM. The usage values displayed are for
the five-minute average for all.
Use the hover feature to display the matrices simultaneously and evaluate any
visible spikes in the data.
Events
To view events specific to the Cluster, Node, Host, VM or Disk, perform the following steps:
Entity Type Entity affected. Values include: All, Cluster, Node, Virtual Machine, and Disk
Severity Event severity level. Valid values include: All, Info, Warning, Error, and Critical
Step 5 (Optional) Use filters to limit the results that appear in the Events Table.
Filter Type a keyword in the Filter option to filter the table contents seen in the browser
Step 6 In the list of events, Click on the event name that you want more information about.
The details appear below the Events table. Details include:
• Description
• Event Name
• Entity Type
• Severity
• DateTime
Alarms
To view alarms specific to the Cluster, Host, and VMs, perform the following steps:
Note Acknowledged alarms on HX Connect or HTML plugin does not acknowledge the equivalent vCenter alarm.
Severity Alarm severity level. Valid values include: All, Info, Warning, and Error
Step 5 (Optional) Use filters to limit the results that appear in the Alarms Table.
Entity Type All, Cluster, Node, Virtual Machine, Disk and Datastore
Filter Type a keyword in the Filter option to filter the table contents seen in the browser
Step 6 Click the Acknowledged button to acknowledge that the alarm has been seen.
Clicking the Acknowledged button auto-enters who acknowledged the alarm in the Acknowledged by field.
Step 7 Click the Reset To Green button to remove the alarm from the list.
Tasks
View asynchronous tasks that are happening on the platform to validate maintenance, perform the following
steps:
Entity Type Type of task, Valid values include: NODE, DP_Summary, Virtual Machine,
Disk and Datastore.
Success indicator Description of the action and the status of the task when it completed. A check
icon preceding the description identifies that the task was successful. Review
this list to identify where a task failed.
Step 4 (Optional) Use the Entity Type list to filter the table results.
Step 1 Access the Hosts & Clusters either from the vSphere Menu or the Shortcut link.
Step 2 Right click on the cluster and select Cisco HyperFlex > Upgrade. Upgrade Launches HyperFlex Connect and takes the
user directly to the upgrade page to complete the upgrade process.
Step 3 Right click on the cluster and select Cisco HyperFlex > Create Datastore.
Related Topics
Create New Datastore, on page 329
Step 1 Access Hosts & Clusters either from the vSphere Menu or the Shortcut link.
Step 2 Click on the cluster name and select the Hosts tab.
The summary page appears.
Step 3 Right click on the host and select Cisco HyperFlex > > Maintenance Mode > Enter (or Exit) MainenanceMode to
enter or Exit HyperFlex Maintenance Mode.
Step 1 Access Hosts & Clusters either from the vSphere Menu or the Shortcut link.
Step 2 Click on a cluster name and select the Summary tab. The summary page appears.
Step 3 Scroll down and use the arrow in each portlet to show or hide the portlet details.
Related Topics
View the HX Cluster Summary, on page 308
Step 1 Access Hosts & Clusters either from the vSphere Menu or the Shortcut link.
Step 2 Click on a cluster name and select the Monitor tab.
Step 3 Scroll down the Monitor navigation panel and locate Cisco HyperFlex.
Step 4 Click on Storage Capacity or Performance to display the related Cisco HyperFlex HTML5 Plugin chart.
Related Topics
View Cluster and Datastore Performance Charts , on page 314
Step 1 Access Hosts & Clusters either from the vSphere Menu or the Shortcut link.
Step 2 Click on a cluster name and select the Configure tab.
Step 3 Scroll down the Monitor navigation panel and locate Cisco HyperFlex.
Step 4 Click on iSCSI Summary or Datastore Summary to display the related Cisco HyperFlex HTML5 Plugin page.
Step 5 Use the buttons to perform all maintenance tasks as defined in the related topics.
Related Topics
Managing Datastores, on page 327
Create New Datastore, on page 329
Edit Datastore, on page 330
Mount or Unmount Datastore, on page 331
Delete Datastore, on page 332
iSCSI: Targets , on page 321
iSCSI: Initiator Groups, on page 323
iSCSI: LUNs, on page 325
Step 1 Right click on the virtual machine. Select Cisco HyperFlex > Snapshot Now.
Step 2 The Take VM Native Snapshot window appears. Complete the following fields:
• Name- Snapshot name
• Description - Description of the snapshot
• Quiesce guest file system - Check box.
Step 3 Click OK to create a VM Snapshot. You will see the snapshot task active in the background. After the Snapshot is
complete it will be listed in the Snapshot Manager.
ReadyClones
• Access the VMs and Templates either from the vSphere Menu or the Shortcut link.
Step 1 Right click on the virtual machine. Select Cisco HyperFlex > ReadyClones.
Step 2 The Ready Clones window appears. Complete the General Settings Fields:
• Number of Clones- Valid entry 1-256
• Customization specifications - If configured, select from the list
• Resource Pool - If configured, select from the list
• Use same name for Guest Name- Uncheck to provide a guest name
Schedule Snapshot
Step 1 Right click on the virtual machine. Select Cisco HyperFlex > Schedule Snapshot .
The Schedule Snapshot window appears.
On Use the Check boxes to select the day(s) of the week for snapshots are to be taken
On Use the Check boxes to select the day(s) of the week for snapshots are to be taken
On Use the Check boxes to select the starting day for the weekly snapshot.
Related Topics
Edit Datastore, on page 330
Delete Datastore
From the Datastore level users have the ability to delete existing datastores.
Related Topics
Delete Datastore, on page 332
Prerequisites
The following is required for using the Remote Plugin for VMware:
• Supported Cisco HyperFlex Release: 5.0(2a) and later (ESXi 6.7 U3 and later).
Rename Clusters ✓
9
Disks View ✓
Nodes View ✓
HX Datastore Management ✓
Network Management ✓
iSCSI Management ✓
10
Manage Tasks ✓
Schedule Snapshot ✓
11
Linked Mode ✓
9
Requires HXDP Release 4.5(x) or later.
10
Requires HXDP Release 4.5(x) or later.
11
Requires HXDP Release 4.5(x) or later.
12
Requires HXDP Release 5.0(x) or later.
Note The configuration and feature functionality for the Remote and Local plugin are identical. For more information
on any feature see the Cisco HyperFlex HTML5 Plugin for VMware vCenter , on page 300 section of this
guide.
Step 1 Download the Cisco HyperFlex HTML plugin OVA for VMware vCenter from the Cisco Software Download site.
Step 2 Login to the vCenter and select the host you want to deploy the Remote Plugin appliance.
Step 3 Right click on the host and click on Deploy OVF Template. Deploy the OVA in a vCenter with appropriate static/DHCP
IP settings.
Step 4 Navigate to and select the local OVF downloaded from Cisco Software Download site.
Recommended configuration settings for the Remote plugin appliance:
• RAM: 4G
• Cores:2
• Datastore: 1 with 50 GB of minimum space
• Valid network adapter
Step 6 Copy and paste the IP address of the appliance in a browser window.
Example:
https://<appliance_ip>
The UI will open.
Step 7 Reset the default user (vcp-admin) password through command line with the following steps:
a) Open the appliance VM console and log in to the appliance with the default credentials.
b) Type passwd vcp-admin and press Enter.
c) Type the old and new password as directed in the UI and complete the password reset process.
Step 8 Log in to the HX vCenter Remote Plug-in Appliance with the new credentials.
Step 9 To register the remote plugin with a vCenter, click the Register button. The Add vCenters window appears.
Step 10 In the Add vCenters window click the Add button and type the vCenter IP/FQDN, Port number, user name and
password in the designated fields. Click Next.
If the vCenter is reachable and credentials are valid, the UI will advance to the Register Plugin page.
Step 11 Click Register and wait for the registration process to complete.
Step 12 Use the hyperlink provided in the registration UI to log on to vCenter.
Step 13 To view the Cisco HyperFlex HTML5 plugin options in the vSphere UI, log out and log in again to vCenter.
Step 1 Download the newer version of the vCenter remote plugin (hyperflex-remote-vcenter-plugin-x.x.x-xxx.tar.gz) upgrade
package from the Cisco Software Download site.
Step 2 Copy it into /tmp directory on the plugin appliance VM.
Step 3 Execute the upgrade script by using the hx-plugin-upgrade <UpgradePackagePath> command.
Step 4 Review the messages carefully and select Y to continue the upgrade process.
Step 5 Once the vCenter application upgrade is complete, update the vCenter extension to the latest version on each vCenter
registered with the plugin server.
a) vCenter Extension Upgrade: Log in to the plugin appliance UI and select the vCenter.
b) Click on Update and provide the credentials and click the Update button.
Encryption Support
Remote Plugin Encryption Support
To enable remote plugin encryption, perform the following steps, For more information on enabling Software
Encryption on your cluster, see Enabling HyperFlex Software Encryption Workflow.
SUMMARY STEPS
1. Log in to the vCenter appliance using pre-configured user credentials.
2. Enter the hx-plugin-supportbundle command. This will prompt you for the vCenter for which you want
to generate the support bundle.
3. Provide the root credentials or skip the vCenter logs and continue with support bundle generation of
vCenter appliance logs.
4. The default support bundle download location is: /var/log/plugin_support/.
DETAILED STEPS
Help support for the support bundle generation tool is hx-plugin-supportbundle -help or
hx-plugin-supportbundle -h
Note:
• Configuration option is Common/Global. It applies to both fabrics and uses the same configuration parameters in
both cases.
• *There is no specific recommendation for VM data VLANs. You can create your own VLANs for the VM data
traffic. By default, the HXDP installer will not create VLANs for the VM data traffic.
• Installer sets the VLANs as non-native by default. Ensure to configure the upstream switches to accommodate the
non-native VLANs.
Note The 8th digit is set to A or B. "A" is set on vNICs pinned to Fabric Interconnect A. And "B" is set on vNICs
pinned to Fabric Interconnect B.
Step 1 Open a web browser and enter the IP address for Cisco UCS Manager . Enter the login credentials.
Step 2 In Cisco UCS Manager, navigate to LAN tab > Pools > root > Sub-org > hx-cluster > MAC Pools.
Step 3 Right-click MAC Pools and select Create MAC Pool.
Step 4 In the Define Name and Description page of the Create MAC Pool wizard, complete the required fields as shown
below:
Name Description
First MAC Address field The first MAC address in the block.
Step 3 Use the following CLI commands to create the three new vSwitches:
# esxcli network vswitch standard add -v vswitch-hx-storage-data
Step 4 The default vSwitch vSwitch0 created during installation of ESXi needs to be renamed to vswitch-hx-inband-mgmt for
the Hx Data Platform node set up scripts to work properly. Use the following command to rename the switch and then
reboot the host so that the vmkernel re-reads its configuration file to use the new name.
# sed -i 's/vSwitch0/vswitch-hx-inband-mgmt/g' /etc/vmware/esx.conf
# reboot
Step 5 You can verify the creation and renaming of the vSwitches after a host reboot with the following command:
# esxcli network vswitch standard list
Confirm that you see the four previously listed vSwitches in the command output. Only the switch-hx-inband-mgmt
vSwitch will have Uplinks and Port groups listed. The HX Data Platform installer scripts perform the rest of the network
configuration.
Note • The HX Data Platform can be configured with VMware DVS or Cisco Nexus 1000v for specific non-HX
dependent networks:
vMotion networks
and virtual machine networks
• For further details, see Cisco Nexus 1000v documentation.
To migrate non-HX dependent vSwitches and associated port groups to DVS or N1Kv networks, follow the
listed steps:
Step 2 Migrate the vSwitch, VMNetwork. Perform the following steps to migrate VMNetwork from legacy vSwitch to DVS.
a) Select vCenter Inventory Lists > Datacenters > datacenter > Related Objects > Distributed Switches.
b) Select the DVSwitch-VMNetwork vSwitch. Click the Add and Manage Hosts icon. This starts the Add and Manage
Hosts wizard.
c) On the Select task page, select Add Hosts. Click Next.
d) On the Select hosts page, click Add New Hosts. Select all hosts in the cluster. Click Next.
e) On the Select network adapter tasks page, select Manage physical adapters and Migrate virtual machine networking.
Click Next.
f) On the Manage physical network adapters page, the physical adapters part of vswitch-hx-vm-network:VM Network
are assigned to the DVSwitch-VMNetwork.
g) Under the On other switches/unclaimed list, select the vmnic corresponding to the In Use by Switch,
vswitch-hx-vm-network.
h) Click Assign uplink.
i) Select Auto-assign.
j) Click OK. The page refreshes with the newly assigned vmnic listed under On this switch.
k) The Analyze impact page shows the impact of this migration. Verify the impact is all green. Click Next.
l) On the Migrate VM networking page, select the VMs to migrate to the new network, DVPortGroup-VMNetwork.
Next
Select all the VMs, except the controller VMs, stCtlVM, from all the hosts. Select the DVPortGroup-VMNetwork.
Click Next.
Note The list of VMs for each host includes all the VMs, including the controller VMs. DO NOT select any
controller VMs. Migrating the controller VMs will break your storage cluster.
m) On the Ready to complete page, confirm the summary of the migration. Click Finish.
Note Post migration system generates several network related alarms. Verify and clear the alarms.
Step 3 Migrate the vSwitch to vmotion pg. Perform the following steps to migrate vmotion pg from legacy vSwitch to DVS.
a) Select vCenter Inventory Lists > Datacenters > datacenter > Related Objects > Distributed Switches.
b) Select the DVSwitch-Vmotion vSwitch. Click the Add and Manage Hosts icon. This starts the Add and Manage
Hosts wizard.
c) On the Select task page, select Add Hosts. Click Next.
d) On the Select hosts page, click Add New Hosts. Select all hosts in the cluster. Click Next.
e) On the Select network adapter tasks page, select the tasks Manage physical adapters and Manage VMkernel adapters.
Click Next.
f) On the Manage physical network adapters page, the physical adapters part of vmotion:vmotion pg are assigned to
the DVSwitch-Vmotion.
Under the On other switches/unclaimed list, select the vmnic corresponding to the In Use by Switch, vmotion. Click
Assign uplink, select Auto-assign, and click OK. The page refreshes with the newly assigned vmnic listed under On
this switch. Click Next.
g) On the Manage VMkernel network adapters page, migrate the VMkernel adapter to the port group,
DVPortGroup-Vmotion.
For each host, under the On other switches list, select the VMKernel adapter corresponding to the In Use by Switch,
vmotion. Click Assign port group. Select the destination port group, DVPortGroup-Vmotion. Click OK. The page
refreshes with the Reassigned VMkernel network adapters, listing the Source Port Group and Destination Port Group.
h) Select the hosts to migrate to the new network, DVPortGroup-Vmotion. Click Next.
i) On the Ready to complete page, confirm the summary of the migration, click Finish.
Step 4 Post migration step. Verify there is no impact on the VMs with respect to IO, Network connectivity and VM Migration.