0% found this document useful (0 votes)
20 views166 pages

SG 248561

This draft document provides an overview and guidance on IBM Storage Virtualize Version 8.7, focusing on its features, enhancements, and configuration processes. It includes detailed chapters on system initialization, configuration, and management of storage systems using IBM FlashSystem and IBM SAN Volume Controller. The document is intended for users looking to implement and optimize IBM Storage Virtualize in their environments.

Uploaded by

cfonseca230379
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
20 views166 pages

SG 248561

This draft document provides an overview and guidance on IBM Storage Virtualize Version 8.7, focusing on its features, enhancements, and configuration processes. It includes detailed chapters on system initialization, configuration, and management of storage systems using IBM FlashSystem and IBM SAN Volume Controller. The document is intended for users looking to implement and optimize IBM Storage Virtualize in their environments.

Uploaded by

cfonseca230379
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 166

Front cover

Draft Document for Review July 22, 2024 9:33 am SG24-8561-00

Unleash the Power of Flash:


Getting Started with IBM Storage Virtualize
Version 8.7 on IBM Storage FlashSystem
and IBM SAN Volume Controller

Andy Harchen
Hartmut Lonzer
Jon Herd
Jonathan Wilkie
Vasfi Gucer

Redbooks
Draft Document for Review July 16, 2024 3:23 pm 8561edno.fm

IBM Redbooks

Unleash the Power of Flash: Getting Started with IBM


Storage Virtualize Version 8.7

July 2024

SG24-8561-00
8561edno.fm Draft Document for Review July 16, 2024 3:23 pm

Note: Before using this information and the product it supports, read the information in “Notices” on
page xv.

First (July 2024)

This edition applies to IBM Storage Virtualize Version 8.7.0.

This document was created or updated on July 16, 2024.

© Copyright International Business Machines Corporation 2024. All rights reserved.


Note to U.S. Government Users Restricted Rights -- Use, duplication or disclosure restricted by GSA ADP Schedule
Contract with IBM Corp.
Draft Document for Review July 22, 2024 9:24 am 8561TOC.fm

Contents

Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii

Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi

Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii

Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi

Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvii
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xix

Chapter 1. Introduction and system overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1


1.1 IBM Storage Virtualize . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.1 Benefits of IBM Storage Virtualize . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 IBM Storage Virtualize V8.7 supported product list. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2.1 IBM Storage Virtualize V8.7.0 current product features . . . . . . . . . . . . . . . . . . . . . 4
1.3 Changes and enhancements in IBM Storage Virtualize V8.7.0 . . . . . . . . . . . . . . . . . . . 6
1.3.1 File system awareness for ransomware detection . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.3.2 GUI updates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.3.3 Flash Grids and storage partition migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3.4 Automatic Patch Updating and Automatic Drive Downloading . . . . . . . . . . . . . . . 10
1.3.5 Management IP changes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.3.6 Remote Copy support on IBM Storage Virtualize V8.7 . . . . . . . . . . . . . . . . . . . . . 12
1.4 Preparation and upgrading to IBM Storage Virtualize V8.7.0 . . . . . . . . . . . . . . . . . . . . 13

Chapter 2. Initial configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15


2.1 Prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.2 System initialization. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.2.1 System initialization process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.3 System setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
2.3.1 System Setup wizard . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
2.3.2 Adding an enclosure in IBM FlashSystem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
2.3.3 Adding a node or hot spare node in IBM SAN Volume Controller systems . . . . . 51
2.3.4 Business continuity with policy-based high availability . . . . . . . . . . . . . . . . . . . . . 55
2.3.5 Configuring quorum disks or applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 56
2.3.6 Configuring the local Fibre Channel port masking . . . . . . . . . . . . . . . . . . . . . . . . 59
2.3.7 Automatic configuration for IBM SAN Volume Controller back-end storage . . . . . 61

Chapter 3. Step-by-step configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67


3.1 The Storage Virtualize GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
3.1.1 Accessing the GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
3.1.2 Brief introduction to the GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
3.2 Network configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
3.2.1 Management IP addresses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69
3.2.2 Service IP addresses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
3.2.3 Additional Ethernet ports. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70

© Copyright IBM Corp. 2024. iii


8561TOC.fm Draft Document for Review July 22, 2024 9:24 am

3.2.4 Portsets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
3.3 Pools and managed disks configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
3.3.1 Provisioning policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
3.3.2 Types of pools. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
3.3.3 Ransomware threat detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72
3.3.4 Creating storage pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
3.3.5 Creating RAID array managed disks in a storage pool . . . . . . . . . . . . . . . . . . . . . 74
3.3.6 Adding external managed disks into a storage pool . . . . . . . . . . . . . . . . . . . . . . . 75
3.3.7 Child pools . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
3.4 Configuring volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
3.4.1 Creating volume groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
3.4.2 Creating volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
3.4.3 Virtual volumes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
3.5 Configuring hosts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
3.5.1 Host attachment overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
3.5.2 Fibre channel host connectivity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
3.5.3 Ethernet host connectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
3.5.4 Host objects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
3.5.5 Mapping volumes for host access. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
3.6 Snapshots and replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
3.6.1 Volume group snapshots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
3.6.2 Asynchronous policy-based replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

Chapter 4. Verifying configuration and basic operations . . . . . . . . . . . . . . . . . . . . . . . 91


4.1 Verifying the configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
4.1.1 System Health Dashboard . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
4.1.2 Verifying network configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
4.1.3 Verifying storage configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
4.1.4 Verifying volume configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
4.1.5 Verifying host configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
4.2 Additional settings and basic operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
4.2.1 Security settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
4.2.2 Audit log . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
4.2.3 Support settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro. . . . . . . . . . . . . . . . . . 99


5.1 IBM Storage Insights overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
5.1.1 IBM Storage Insights: Information and registration . . . . . . . . . . . . . . . . . . . . . . . 101
5.2 IBM Storage Insights monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
5.2.1 Component health. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
5.2.2 Capacity monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
5.2.3 Performance monitoring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5.2.4 Logging support tickets by using IBM Storage Insights . . . . . . . . . . . . . . . . . . . 106
5.2.5 Managing existing support tickets by using IBM Storage Insights . . . . . . . . . . . 112
5.2.6 Enhancements to IBM Storage Insights Pro . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115

Chapter 6. Storage Virtualize troubleshooting and diagnostics. . . . . . . . . . . . . . . . . 117


6.1 Troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
6.1.1 Storage Insights . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
6.1.2 Using the GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
6.1.3 Recommended actions and fix procedure. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120
6.1.4 Storage Virtualize failure recovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
6.1.5 Using the command-line interface. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
6.2 Collecting diagnostic data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125

iv Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561TOC.fm

6.2.1 IBM Storage Virtualize systems data collection . . . . . . . . . . . . . . . . . . . . . . . . . 125


6.2.2 Drive data collection: drivedumps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
6.2.3 Host multipath software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
6.2.4 More data collection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137

Related publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139


IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
Online resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139
Help from IBM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 139

Contents v
8561TOC.fm Draft Document for Review July 22, 2024 9:24 am

vi Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561LOF.fm

Figures

1-1 IBM FlashSystems and SVC Family . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3


1-2 Async policy-based replication and partition base HA user experience improvements . 7
1-3 Volume group tile and assigning ownership groups to volume groups . . . . . . . . . . . . . . 8
1-4 GUI performance panel. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1-5 IBM Storage Virtualize upgrade support matrix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2-1 Technician port FlashSystem 9500. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2-2 Technician port FlashSystem 7300. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2-3 Technician port FlashSystem 5200. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2-4 Technician port FlashSystem 5300. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2-5 Technician port FlashSystem 5045. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2-6 Technician port FlashSystem 5015. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2-7 Technician port IBM SAN Volume Controller 2145-SV3 . . . . . . . . . . . . . . . . . . . . . . . . 19
2-8 Technician port IBM SAN Volume Controller 2145-SV2 . . . . . . . . . . . . . . . . . . . . . . . . 19
2-9 Logging in to Service Assistant by way of the technician port . . . . . . . . . . . . . . . . . . . 21
2-10 System Initialization: Canister detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2-11 System Initialization: Initialize the first enclosure . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2-12 System Initialization: Initialize the first IBM SAN Volume Controller node . . . . . . . . . 22
2-13 System Initialization: Enter Management IP . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
2-14 System Initialization: Web-server restart timer counting down from 5 minutes. . . . . . 23
2-15 System Initialization completed. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
2-16 Logging in for the first time . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
2-17 Initial Setup Window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
2-18 Setup Call Home . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2-19 Transmission Types for Call Home . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
2-20 Setup Internal Proxy Server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
2-21 Connection Test to the Support Center . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
2-22 System Location . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2-23 Summary page . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
2-24 System Setup Welcome page. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
2-25 Accept License Agreement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
2-26 Change password . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
2-27 System Name . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36
2-28 License Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
2-29 DNS Server setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
2-30 Date and Time . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
2-31 Activate Encryption License . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
2-32 Encryption licensed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
2-33 Change Call Home settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
2-34 Setup Support Assistance. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
2-35 System communicating with named IBM Support servers . . . . . . . . . . . . . . . . . . . . . 44
2-36 Remote support access settings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
2-37 Automatic Configuration for Virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
2-38 Summary Page . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
2-39 System Initialization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
2-40 Setup completed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
2-41 Dashboard . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
2-42 Add Enclosure button . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
2-43 Selecting the control enclosure to add . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50

© Copyright IBM Corp. 2024. vii


8561LOF.fm Draft Document for Review July 22, 2024 9:24 am

2-44 Add Node button . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52


2-45 Adding a node . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
2-46 IBM SAN Volume Controller is adding node to the cluster . . . . . . . . . . . . . . . . . . . . . 53
2-47 Node added . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
2-48 Download IPv4 quorum button . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
2-49 Download IP quorum application window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 57
2-50 IP quorum application that is deployed and connected . . . . . . . . . . . . . . . . . . . . . . . 58
2-51 Changing the quorum mode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
2-52 Applying a port mask by using a GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
2-53 Modify Connection dialog box . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60
2-54 Automatic Configuration wizard enablement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
2-55 Automatic configuration: Add Enclosure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
2-56 Defining a host cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 63
2-57 Hosts inside an IBM SAN Volume Controller host cluster. . . . . . . . . . . . . . . . . . . . . . 64
2-58 Begin the automatic configuration process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
2-59 Automatic pool configuration. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
2-60 Pools configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
2-61 Automatic configuration running commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
2-62 Automatic configuration complete . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
3-1 Welcome page with the dashboard. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
3-2 Ethernet ports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
3-3 Add IP address . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
3-4 Portsets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
3-5 Create Pool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
3-6 Create Pool panel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
3-7 Add Storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
3-8 RAID array . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
3-9 Child pools with different purposes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
3-10 Create Child Pool . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
3-11 Add Host . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 84
3-12 Snapshot Policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
3-13 Suspend Policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86
3-14 Create Partnership . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
3-15 Partnership Created . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
3-16 Setup policy-based replication wizard. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
4-1 System Health. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
4-2 Portset Mappings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
4-3 Portset IP addresses. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
4-4 Host with asymmetrical logins. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
4-5 Audit log . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
5-1 IBM Storage Insights System overview (classic view) . . . . . . . . . . . . . . . . . . . . . . . . 101
5-2 IBM Storage Insights System overview (Carbon enhanced view) . . . . . . . . . . . . . . . 102
5-3 Component Health overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
5-4 Ports in error . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
5-5 Capacity area of the IBM Storage Insights system overview . . . . . . . . . . . . . . . . . . . 103
5-6 Capacity planning for one system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5-7 System overview: Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
5-8 IBM Storage Insights: Performance view . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5-9 Filtered performance graph. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
5-10 Performance List View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
5-11 Get Support (see highlighted area) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
5-12 Get Support window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
5-13 Create Ticket wizard . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 108

viii Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561LOF.fm

5-14 Add a note or attachment window. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109


5-15 Selecting a Severity Level window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
5-16 Review the ticket window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
5-17 Update ticket . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
5-18 View tickets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
5-19 Adding a log package to the ticket . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
5-20 Confirming the log upload . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
5-21 Log upload completed and processing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
5-22 IBM Storage insights Pro and IBM Flash Grid integration. . . . . . . . . . . . . . . . . . . . . 116
6-1 Events icon in the GUI . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
6-2 System Health expanded section in the dashboard . . . . . . . . . . . . . . . . . . . . . . . . . . 120
6-3 Recommended actions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
6-4 Monitoring → Events window . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
6-5 Properties and Sense Data for an event . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124
6-6 Upload Support Package details. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128

Figures ix
8561LOF.fm Draft Document for Review July 22, 2024 9:24 am

x Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561LOT.fm

Tables

1-1 IBM Storage Virtualize V8.7 supported product list . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3


1-2 IBM FlashSystem current products feature summary comparison chart . . . . . . . . . . . . 4
6-1 Useful AIX lspath commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
6-2 Useful AIX lsmpio commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
6-3 Useful Windows mpclaim.exe commands. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
6-4 Useful Windows PowerShell cmdlets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133

© Copyright IBM Corp. 2024. xi


8561LOT.fm Draft Document for Review July 22, 2024 9:24 am

xii Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561LOE.fm

Examples

2-1 Reenabling the onboard Ethernet port 2 as the technician port . . . . . . . . . . . . . . . . . . 17


2-2 Listing the I/O groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
2-3 Listing the candidate control enclosures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
2-4 Adding a control enclosure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
2-5 Adding an expansion enclosure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
2-6 Listing I/O groups . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
2-7 Listing the candidate nodes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
2-8 Adding a node as a spare . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
2-9 Adding a node to an I/O group . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
2-10 Single IO-group (two nodes) and one spare . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
2-11 Two IO-groups (four nodes) configured- no spare . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
2-12 Starting the IP quorum application on the Windows operating system . . . . . . . . . . . . 57
2-13 Viewing the local port mask . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
2-14 Setting a local port mask by running the chsystem command . . . . . . . . . . . . . . . . . . 61
6-1 The svc_livedump command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
6-2 preplivedump and lslivedump commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
6-3 The triggerdrivedump command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
6-4 Output for the multipath -ll command . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
6-5 Output of esxcli storage core path list command . . . . . . . . . . . . . . . . . . . . . . . . . . . . 134
6-6 Output of esxcli storage core path list -d <naaID> . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
6-7 Output for esxcli storage nmp device list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136

© Copyright IBM Corp. 2024. xiii


8561LOE.fm Draft Document for Review July 22, 2024 9:24 am

xiv Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 16, 2024 3:23 pm 8561spec.fm

Notices

This information was developed for products and services offered in the US. This material might be available
from IBM in other languages. However, you may be required to own a copy of the product or product version in
that language in order to access it.

IBM may not offer the products, services, or features discussed in this document in other countries. Consult
your local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not
infringe any IBM intellectual property right may be used instead. However, it is the user’s responsibility to
evaluate and verify the operation of any non-IBM product, program, or service.

IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, MD-NC119, Armonk, NY 10504-1785, US

INTERNATIONAL BUSINESS MACHINES CORPORATION PROVIDES THIS PUBLICATION “AS IS”


WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED
TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT, MERCHANTABILITY OR FITNESS FOR A
PARTICULAR PURPOSE. Some jurisdictions do not allow disclaimer of express or implied warranties in
certain transactions, therefore, this statement may not apply to you.

This information could include technical inaccuracies or typographical errors. Changes are periodically made
to the information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.

Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the
materials for this IBM product and use of those websites is at your own risk.

IBM may use or distribute any of the information you provide in any way it believes appropriate without
incurring any obligation to you.

The performance data and client examples cited are presented for illustrative purposes only. Actual
performance results may vary depending on specific configurations and operating conditions.

Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.

Statements regarding IBM’s future direction or intent are subject to change or withdrawal without notice, and
represent goals and objectives only.

This information contains examples of data and reports used in daily business operations. To illustrate them
as completely as possible, the examples include the names of individuals, companies, brands, and products.
All of these names are fictitious and any similarity to actual people or business enterprises is entirely
coincidental.

COPYRIGHT LICENSE:

This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in
any form without payment to IBM, for the purposes of developing, using, marketing or distributing application
programs conforming to the application programming interface for the operating platform for which the sample
programs are written. These examples have not been thoroughly tested under all conditions. IBM, therefore,
cannot guarantee or imply reliability, serviceability, or function of these programs. The sample programs are
provided “AS IS”, without warranty of any kind. IBM shall not be liable for any damages arising out of your use
of the sample programs.

© Copyright IBM Corp. 2024. xv


8561spec.fm Draft Document for Review July 16, 2024 3:23 pm

Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business Machines
Corporation, registered in many jurisdictions worldwide. Other product and service names might be
trademarks of IBM or other companies. A current list of IBM trademarks is available on the web at “Copyright
and trademark information” at https://www.ibm.com/legal/copytrade.shtml

The following terms are trademarks or registered trademarks of International Business Machines Corporation,
and might also be trademarks or registered trademarks in other countries.
Redbooks (logo) ® HyperSwap® IBM Research®
AIX® IBM® IBM Spectrum®
DS8000® IBM Cloud® Interconnect®
Easy Tier® IBM FlashCore® PowerHA®
FlashCopy® IBM FlashSystem® Redbooks®

The following terms are trademarks of other companies:

Intel, Intel Xeon, Intel logo, Intel Inside logo, and Intel Centrino logo are trademarks or registered trademarks
of Intel Corporation or its subsidiaries in the United States and other countries.

ITIL is a Registered Trade Mark of AXELOS Limited.

The registered trademark Linux® is used pursuant to a sublicense from the Linux Foundation, the exclusive
licensee of Linus Torvalds, owner of the mark on a worldwide basis.

Microsoft, Windows, and the Windows logo are trademarks of Microsoft Corporation in the United States,
other countries, or both.

OpenShift, Red Hat, are trademarks or registered trademarks of Red Hat, Inc. or its subsidiaries in the United
States and other countries.

VMware, VMware vSphere, and the VMware logo are registered trademarks or trademarks of VMware, Inc. or
its subsidiaries in the United States and/or other jurisdictions.

Other company, product, or service names may be trademarks or service marks of others.

xvi Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561pref.fm

Preface

IBM Storage Virtualize (formerly IBM Spectrum Virtualize) simplifies infrastructure


management for block storage across diverse workloads, on-premises, off-premises, or in
hybrid cloud environments. This core offering of the IBM Storage portfolio enables rapid
deployment and streamlines management for SAN Volume Controller and IBM FlashSystem®
systems, including support for hybrid multicloud deployments.

This IBM Redbooks® publication, focused on IBM Storage Virtualize Version 8.7, guides
users through new features, upgrades, and configuration for both new and existing systems.
It's intended for pre-sales and post-sales technical support personnel, as well as storage
administrators.

Authors
This book was produced by a team of specialists from around the world.

Andy Harchen is a subject matter expert (SME) who is based


at the EMEA TLS Remote Technical Support Storage, IBM
Germany. He has over 25 years of experience in onsite and
remote technical support with a focus on disk storage and
virtualization solutions. In his current role, he delivers technical
support for IBM Storage Virtualize Products (IBM SAN Volume
Controller and IBM FlashSystem). He is a member of the
Predictive Support Team which provide proactive support
based on cloud technology such as Storage Insights,
supported by AI to detect and prevent impacts and outages of
systems.

Hartmut Lonzer brings 45 years of technical and sales


expertise at IBM to his role as a Storage Advisory Partner
Technical Specialist for DACH. Previously, he served as an
OEM Alliance Manager for Lenovo at IBM Germany. Based at
the company's headquarters in Ehningen, his focus lies on the
IBM FlashSystem Family and IBM SAN Volume Controller
products, with experience dating back to their introduction.

© Copyright IBM Corp. 2024. xvii


8561pref.fm Draft Document for Review July 19, 2024 5:44 pm

Jon Herd is an IBM Senior Executive Advocate working for the


TLS EMEA Remote Technical Support and Client Care team
based in IBM Germany. He covers the United Kingdom, Ireland
and beyond, advising customers on a portfolio of IBM storage
products, including FlashSystem products. He also works as a
senior advisor to the TLS EMEA RTS/CC management on new
products, strategy and new technologies that might affect the
TLS business. Jon has been with IBM for more than 45 years,
and has held various technical roles, including Europe, Middle
East, and Africa (EMEA) level 2 support on mainframe servers
and technical education development. He has written many
IBM Redbooks on the FlashSystem products and is an IBM
Redbooks Platinum level author. He holds IBM certifications in
Product Services profession at a thought leader L3 level, and is
a Technical Specialist at an experienced L1 level. He also is a
certified Chartered Member of the British Computer Society
(MBCS - CITP), a Certified Member of the Institution of
Engineering and Technology (MIET), and a Certified Technical
Specialist of the Open Group (TOG).

Jonathan Wilkie is an Advanced Subject Matter Expert/L3


support representative for IBM Spectrum Virtualize and IBM
FlashSystem. He has more than 20 years of experience in IBM
storage technical support. Over his career, he has provided
technical support for Shark, DS4000, DS6000, and IBM
DS8000® products. He has been supporting IBM Storage
Virtualize-based products since 2010.

Vasfi Gucer leads projects for the IBM Redbooks team,


leveraging his 20+ years of experience in systems
management, networking, and software. A prolific writer and
global IBM instructor, his focus has shifted to storage and cloud
computing in the past eight years. Vasfi holds multiple
certifications, including IBM Certified Senior IT Specialist, PMP,
ITIL V2 Manager, and ITIL V3 Expert.

Thanks to the following people for their contributions to this project:

Elias Luna, Andrew Greenfield


IBM USA

Lucy Harris, Evelyn Perez, Chris Bulmer, Chris Canto, Daniel Dent, Bill Passingham,
Nolan Rogers, David Seager, Russell Kinmond
IBM UK

Sushil H Sharma, Ramakrishna Vadla


IBM India

Diana Laura Silva Gallardo


IBM Mexico

xviii Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561pref.fm

Now you can become a published author, too!


Here’s an opportunity to spotlight your skills, grow your career, and become a published
author—all at the same time! Join an IBM Redbooks residency project and help write a book
in your area of expertise, while honing your experience using leading-edge technologies. Your
efforts will help to increase product acceptance and customer satisfaction, as you expand
your network of technical contacts and relationships. Residencies run from two to six weeks
in length, and you can participate either in person or as a remote resident working from your
home base.

Find out more about the residency program, browse the residency index, and apply online at:
ibm.com/redbooks/residencies.html

Comments welcome
Your comments are important to us!

We want our books to be as helpful as possible. Send us your comments about this book or
other IBM Redbooks publications in one of the following ways:
򐂰 Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
򐂰 Send your comments in an email to:
[email protected]
򐂰 Mail your comments to:
IBM Corporation, IBM Redbooks
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400

Stay connected to IBM Redbooks


򐂰 Find us on LinkedIn:
https://www.linkedin.com/groups/2130806
򐂰 Explore new Redbooks publications, residencies, and workshops with the IBM Redbooks
weekly newsletter:
https://www.redbooks.ibm.com/subscribe
򐂰 Stay current on recent Redbooks publications with RSS Feeds:
https://www.redbooks.ibm.com/rss.html

Preface xix
8561pref.fm Draft Document for Review July 19, 2024 5:44 pm

xx Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

Chapter 1. Introduction and system


overview
This chapter dives into IBM Storage Virtualize V8.7, covering its new features, supported
systems, and upgrade considerations.

This chapter has the following sections:


򐂰 “IBM Storage Virtualize” on page 2
򐂰 “IBM Storage Virtualize V8.7 supported product list” on page 3
򐂰 “Changes and enhancements in IBM Storage Virtualize V8.7.0” on page 6
򐂰 “Preparation and upgrading to IBM Storage Virtualize V8.7.0” on page 13

© Copyright IBM Corp. 2024. 1


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

1.1 IBM Storage Virtualize


IBM Storage Virtualize (formerly IBM Spectrum Virtualize) simplifies managing block storage
for various workloads, on-premises or in the cloud. It runs on IBM FlashSystem and SAN
Volume Controller hardware, offering data protection, rapid cloud deployment, and
performance for analytics. IBM Storage Virtualize provides a way to manage and protect huge
volumes of data from mobile and social applications, enable rapid and flexible cloud services
deployments, and deliver the performance and scalability that is needed to gain insights from
the latest analytics technologies.

Note: For more information, see IBM Storage FlashSystem and IBM SAN Volume
Controller.

With the introduction of the IBM Storage family, the software that runs on IBM SAN Volume
Controller and on IBM Storage FlashSystem (IBM FlashSystem) products is called IBM
Storage Virtualize. The name of the underlying hardware platform has not changed.

1.1.1 Benefits of IBM Storage Virtualize


IBM Storage Virtualize delivers benefits that improve storage infrastructure in many ways,
including the following examples:
򐂰 Reduces the cost of storing data by increasing the use and accelerating applications to
speed business insights. To achieve this goal, the solution provides the following functions:
– Uses data reduction technologies to increase the amount of data that you can store in
the same space
– Enables rapid deployment of cloud storage for disaster recovery (DR) along with the
ability to store copies of local data
– Moves data to the most suitable type of storage based on policies that you define by
using IBM Storage Control to optimize storage
– Improves storage migration performance so that you can do more with your data
򐂰 Protects data from theft or inappropriate disclosure while enabling a high availability (HA)
strategy that includes protection for data and application mobility and DR. To achieve this
goal, the solution provides the following functions:
– Uses software-based encryption to improve data security
– Provides fully duplexed copies of data and automatic switchover across data centers to
improve data availability
– Eliminates storage downtime with nondisruptive movement of data from one type of
storage to another type.
򐂰 Simplifies data by providing a data strategy that is independent of your choice of
infrastructure, which delivers tightly integrated functions and consistent management
across heterogeneous storage. To achieve this goal, the solution provides the following
functions:
– Integrates with virtualization tools, such as VMware vCenter to improve agility with
automated provisioning of storage and easy deployment of new storage technologies
– Enables supported storage to be deployed with Kubernetes and Docker container
environments, including Red Hat OpenShift
– Consolidates storage, regardless of the hardware vendor for simplified management,
consistent functions, and greater efficiency

2 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

– Supports common capabilities across storage types, which provide flexibility in storage
acquisition by allowing a mix of vendors in the storage infrastructure

Note: These benefits are a subset of the list of features and functions that are available
with IBM Storage Virtualize software.

Figure 1-1 on page 3 shows the current IBM FlashSystem and IBM SAN Volume Controller
Family.
i

Figure 1-1 IBM FlashSystems and SVC Family

Note: IBM Storage Virtualize for Public Cloud is not currently supported on IBM Storage
Virtualize V8.7. This functionality is planned for a future release.

1.2 IBM Storage Virtualize V8.7 supported product list


Table 1-1 shows the IBM Storage Virtualize V8.7 supported product list and whether the
product is still currently sold or is designated as End-of-Marketing (EOM).

Table 1-1 IBM Storage Virtualize V8.7 supported product list


Product Machine Type Model Comment

FS9500/R 4666, 4983 AH8, UH8 Current Product

FS7300 4657 924, U7D Current Product

FS5300 4662 7H2 Current Product

FS5200 4662 6H2, UH6 Current Product

Chapter 1. Introduction and system overview 3


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

Product Machine Type Model Comment

FS5045 4680 3P2, 3P4 Current Product

FS5015 4680 2P2, 2P4 Current Product

SVC 2145, 2147 SA2, SV3 Current Product

SVC 2145, 2147 SV2 EOM 01/2023

FS9200/R 9846, 9848, 4666 AG8, UG8 EOM 07/2022

FS7200 2076, 4664 824, U7C EOM 07/2022

FS9100 9846, 9848 AF8, UF8 EOM 07/2022

FS5000 (FS5015, 2072 2N2, 2N4, 3N2, 3N4 EOM 12/2023


FS5035)

Note: This version of the IBM Redbooks includes systems that can run IBM Storage
Virtualize V8.7. Some products that are listed in the book are no longer sold by IBM but
can still run the V8.7 software. Where this is applicable, it is mentioned in the text.

1.2.1 IBM Storage Virtualize V8.7.0 current product features


This is a brief summary of the technical specifications for this solution. It provides a concise
overview of the essential features and specifications.

Table 1-2 shows the IBM Storage FlashSystem Family feature summary and comparison, for
currently marketed products.

Table 1-2 IBM FlashSystem current products feature summary comparison chart
SVC 5015 5045 5300 7300 9500

Machine 2145 4680 4680 4662 4657 4983


Type 2147 4666

Controller SA2 (No 2P2 (12-drive) 3P2 (12-drive) 7H2 924 (24-drive) AH8
Models Drives) 2P4 (24-drive) 3P4 (24-drive) (12-drive) (48-drive)
SV3 (No
Drives)

Expansion N/A 12H (12-drive) 12H 12G 12G AFF


Models 24H (24-drive) (12-drive) (12-drive) (12-drive) (24-drive)
92H (92-drive) 24H 24G 24G A9F
(24-drive) (24-drive) (24-drive) (92-drive)
92H 92G 92G
(92-drive) (92-drive) (92-drive)

Processors 2 Intel Xeon 2 Intel Xeon 2 Intel Xeon 2 Intel Xeon 2 Intel Xeon 4 Intel Xeon
CPUs CPUs CPUs CPUs CPUs CPUs
SV3 24 cores 2 cores each 6 cores each 12 cores each 10 cores each 24 cores each
each
SA2 8 Cores
each

4 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

SVC 5015 5045 5300 7300 9500

Memory SA2 32GB 32GB 64GB 256GB 1TB


128GB 64GB 64GB 256GB 756GB 2TB
386GB 512GB 1.5TB 3TB
764GB
SV3
512GB
1TB
1.5TB

Height 2U 2U 2U 1U 2U 4U

Connectivity N/A 1 Gb/s iSCSI 10 Gb/s iSCSI 25/10 Gb/s 10 Gb/s iSCSI N/A
(standard) iSCSI or
NVMe/TCP

Connectivity 64 Gb/s FC or 16 Gb/s FC 16 Gb/s FC 64 Gb/s FC or 32 Gb/s FC or 64 Gb/s FC or


(optional) NVMe/FC 10 Gb/s iSCSI 10 Gb/s iSCSI NVMe/FC NVMe/FC NVMe/FC
(SV3) 12 Gb/s SAS 12 Gb/s SAS 32 Gb/s FC or 25/10 Gb/s 32 Gb/s FC or
32 Gb/s FC or NVMe/FC iSCSI or NVMe/FC
NVMe/FC 10 Gb/s iSCSI NVMe /TCP 100 Gb/s
100 Gb/s or NVMe iSCSI or
iSCSI or /TCP NVMe /TCP
NVMe /TCP 25/10 Gb/s
25/10 Gb/s iSCSI or
iSCSI or NVMe /TCP
NVMe /TCP

Max ports 12 8 8 16 24 48

Max IOPS SA2 1M 140k 400k 700k 1M 2.5M


(4K read SV3 2.5M
miss)

Max SA2 45GB/s 8 GB/s 12 GB/s 28 GB/s 45 GB/s 100 GB/s


bandwidth SV3 100GB/s
(256K read
miss)

Warranty and 2145 One year 9x5 One year 9x5 One year 9x5 One year 9x5 One year
Support Enterprise standard standard standard standard 24x7 standard
Class Support 1-5 Expert 1-5 Expert 1-5 Expert 1-5 Expert 1-5 Expert
and a one Care Basic, Care Basic, Care Basic, Care Basic, Care
year warranty. Advanced or Advanced or Advanced or Advanced or Advanced or
2147 Premium Premium Premium Premium Premium
Enterprise
Class Support
and a
three-year
warranty.

Chapter 1. Introduction and system overview 5


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

SVC 5015 5045 5300 7300 9500

Dimensions Control Control Control Control Control Control


Enclosure enclosure enclosure enclosure enclosure enclosure
• Height:8.7 • Height: 8.7 • Height: 8.7 • Height: 4.3 • Height: 8.8 • Height:
cm (3.4 in.) cm (3.4 in.) cm (3.4 in.) cm (1.7 in.) cm (3.5 in.) 17.43 cm (6.8
• Width 44.6 • Width: 48.3 • Width: 48.3 • Width: 44.6 • Width: 48.3 in.)
cm (17.6 in) cm (19.0 in.) cm (19.0 in.) cm (17.5 in.) cm (19.0 in.) • Width: 44.6
• Depth 82.6 • Depth: 55.6 • Depth: 55.6 • Depth: 77 • Depth: 85 cm (17.6 in.)
cm (32.5 in) cm (21.9 in.) cm (21.9 in.) cm (30.3 in.) cm (33.5 in.) • Depth: 82.6
cm (32.6 in.)

Weight 25 kg (55 lb) Fully Fully Fully Fully Fully


to 30 kg (65 configured (12 configured configured configured configured
lb) depending drives): (12 drives): (12 drives): (24 drives): (48 drives):
on 28.3Kg 28.3Kg 19.5Kg 46.6Kg 70.5Kg
configuration Fully Fully
configured (24 configured
drives): (24 drives):
27.3Kg 27.3Kg

1.3 Changes and enhancements in IBM Storage Virtualize


V8.7.0
IBM Storage Virtualize V8.7.0 provides more features and updates to the IBM Storage
Virtualize family of products, which include IBM FlashSystems and the IBM SAN Volume
Controller.

1.3.1 File system awareness for ransomware detection


Volumes can be used by many different applications, operating systems and file systems.
This might pose a challenge for AI-powered ransomware detection. Knowing the specific file
system used on each volume can significantly improve I/O pattern analysis, a key element in
ransomware detection.

However, storage administrators often lack complete visibility into volume usage. Applications
and different teams may employ volumes for diverse purposes, often resorting to cryptic
volume names that don't reflect the actual use case. This lack of clear information hinders
efficient storage management and ransomware detection strategies. IBM Support can also
benefit from understanding what file systems are in which volumes in some recovery
scenarios.

IBM Storage Virtualize V8.7.0 provides this new file level awareness for ransomware
detection as follows:
1. Every 12 hours, the file system is automatically updated for each volume.
– Can also updated by analyzevdisk or analyzevdiskbysystem CLI commands.
2. Background reads are sent to a volume.
3. Open source libraries used to determine file system.
4. Output is displayed in lsvdiskanalysis's file_system field.
– 15 character max for field.

6 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

– Can display multiple file systems.


5. File system used by inferencing engine to improve ransomware detection.

Enhancement to IBM Storage Insights for threat detection


A key part of monitoring your system includes the detection of potential ransomware attacks.
To ensure that you have the latest storage metadata for detecting those types of attacks,
compression and cyber resiliency statistics for volumes are collected every 5 minutes.

With these statistics, IBM Storage Insights builds a historical model of a storage system and
uses its built-in intelligence and formulas to identify when and where ransomware attacks
might be occurring. For more information about statistics, see Storage Insights Overview.

For more information on what is new with Storage Virtualize V8.7: The following blog
post IBM Storage Virtualize 8.7.0 including Flash Grid (Barry Whyte and Andrew Martin:
IBM Storage blog) provides a good overview of the new features in Version 8.7.

1.3.2 GUI updates


IBM Storage Virtualize V8.7.0 provides these three areas of GUI enhancements:

Async policy-based replication and partition base HA user experience


improvements
Figure 1-2 shows the async policy-based replication and partition base HA user experience
improvements.

Figure 1-2 Async policy-based replication and partition base HA user experience improvements

Volume group tile and assigning ownership groups to volume groups


Figure 1-3 on page 8 shows the volume group tile and assigning ownership groups to volume
group enhancement.

Chapter 1. Introduction and system overview 7


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

Figure 1-3 Volume group tile and assigning ownership groups to volume groups

Performance panel carbonization


Storage Virtualize introduces GUI performance panel modernization using Carbon v11
components and Carbon Charts:
– Tabs based implementation for future scalability of the charts to a incorporate a
growing number of statistics.
– Charts by user customization across the user sessions.
– Responsive flexible layout of 1 column, 2 column or 3 columns.
– Improved usability and accessibility through drag and drop feature.
– New option for power and temperature charts to display statistics at systems, node or
enclosure level.
– Easy to compare the IOPS, bandwidth and latency through single view.
– Restore to default feature to remove all user customization and restore default view.

Figure 1-4 shows the newly designed GUI performance panel.

8 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

Figure 1-4 GUI performance panel

1.3.3 Flash Grids and storage partition migration


A Flash Grid is a collection of single I/O group FlashSystem or SVC systems that looks and
feels like a single storage solution, providing single pane of glass management and
non-disruptive migration

The current I/O group structure presents several limitations that hinder performance
scalability and flexibility:
򐂰 Limited scalability: A maximum of four I/O groups restricts the overall performance
achievable by a single system.
򐂰 Hardware compatibility challenges: Compatibility requirements between I/O groups
complicate hardware upgrades.
򐂰 Disruptive upgrades: System-wide upgrades are needed for both software and
hardware, leading to downtime and complexity.
򐂰 Non-linear object limits: Volume, snapshot, and host counts are limited per system, not
per I/O group, hindering scalability.
򐂰 Feature restrictions: Several advanced features, like policy-based HA, vVol replication,
and storage partitions, are only available on single I/O group systems

Flash Grid addresses these limitations by offering a more granular and flexible approach to
storage management.

Flash Grid brings a novel approach to storage management, and Storage Virtualize 8.7.0
marks the first phase of its implementation. Here are the key features available in this release:
򐂰 CLI-driven Flash Grid management: Initial configuration and management are primarily
done through the command-line interface (CLI), which uses AI-assisted storage partition

Chapter 1. Introduction and system overview 9


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

migration utilizing artificial intelligence to optimize the migration of storage partitions


across Flash Grid member systems.
򐂰 Scalability for performance and flexibility: Flash Grid allows grouping up to eight
systems, providing significant scalability for I/O and storage capacity.
򐂰 Broad hardware compatibility: All NVMe-based FlashSystem products and SVC models
are supported within a Flash Grid, promoting hardware choice and future upgrades.
򐂰 Centralized management and ownership: A single system, designated as the Flash
Grid owner, manages the membership of other systems within the grid.
򐂰 Tiering for optimized performance: Flash Grid empowers you to combine different
hardware within a single grid, enabling you to tier storage based on performance and
capacity needs.
򐂰 Independent software updates: Each member system in a Flash Grid can receive
software updates independently, offering greater flexibility and reduced downtime for
maintenance.

Note: It is planned to make the Flash Grid implementation, monitoring and management
as part of the GUI in a future release. There will also be tighter integration with IBM
Storage Insights to give AI capable operation to storage partitions migration across
systems in the Flash Grid.

1.3.4 Automatic Patch Updating and Automatic Drive Downloading


Storage Virtualize V8.7 simplifies patching with two new features

Automatic Patch Updating


Automatic Patch Updating is a mechanism that enables security, or other patches, to be
scheduled to automatically update on a user’s system, as new patch versions are published.

A patch is a lightweight update to a function or service, which can be installed on a user’s


system. A patch install never requires a node reboot or reset.

Note: A patch install may restart a Linux service when installed. It can be installed on all
platform types and is small in size.

A process for creating and publishing patches is already in place on Storage Virtualize 8.6.0
and patches are stored on IBM FixCentral. IBM Cloud® Call Home is used to access patches.
Newer versions of IBM Storage Virtualize code can include older patches released in
previous Storage Virtualize versions.

When developers identify a need, they create patches (such as bug fixes, security updates
and so forth) to address issues in Storage Virtualize. These patches, along with existing ones,
are published on IBM FixCentral, a centralized repository for software updates.

The Automatic Patch Updating has the following benefits to the clients systems:
򐂰 The enhanced Patching framework empowers you to schedule automatic patch
application for your Storage Virtualize systems. This eliminates the need for complex full
PTF (Program Temporary Fix) or concurrent code upgrades, saving you valuable time and
effort.
򐂰 Benefits users whose systems have patches that might need be frequently updated.

10 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

Note: An example could be Ransomware Threat Detection, where the inference data
files could be regularly changed.

򐂰 Users can set up their systems in the knowledge that updating of vital patches will happen
seamlessly, in the background.
򐂰 Automatic Patch Updating can be set up on a user’s system, either using the GUI, or via
CLI commands.

Once configured, automatic patching performs daily checks on IBM Fix Central. If any
selected patches are available for download, they will be automatically downloaded and
applied to your system.

Important: Automatic Patch Updating leverages IBM Cloud Call Home to access patch
information and lists. Therefore, ensuring a functioning IBM Cloud Call Home is a
prerequisite before configuring automatic updates.

Automatic Drive Firmware Download


Automatic Drive Firmware Download is a mechanism that utilizes the new patch
infrastructure to enable drive firmwares to be stored on a cluster and ensure a standard drive
firmware level is maintained.

Note: This change is for FCM drives only.

The Automatic Drive Firmware Download process is as follows:


򐂰 Building on the patching infrastructure, FCM drive firmwares are now built as patches.
򐂰 You can use the applysoftware <firmware_patch> CLI command to copy the firmware
patch to all cluster nodes, streamlining the update process.
򐂰 When a new drive is added to the cluster, it automatically checks for and applies the latest
available firmware, ensuring your storage remains up-to-date.
򐂰 Firmware patches are located on IBM FixCentral, alongside other software updates for
Storage Virtualize.
򐂰 Newer versions of Storage Virtualize may include older firmware patches, eliminating the
need to search for and apply them individually.

With Automatic Drive Firmware Download you can be confident that any FCM Field
Replaceable Unit (FRU) replacements or new additional drives added to your array will be
compatible with your system's firmware.

Example scenario
Let us consider the following scenario:
1. User has a FCM4 array using 4.1.4 firmware version. A drive fails and requires
replacement.
2. A FRU arrives with version 4.0.4. The user performs the Dynamic Drive Pool (DMP)
operation and replaces the failed drive with the replacement FRU.
3. As the drive attempts to rejoin the array, Automatic Drive Download functionality
seamlessly takes over, verifies that version 4.1.4 is available and upgrades the drive.

Chapter 1. Introduction and system overview 11


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

1.3.5 Management IP changes


IBM Storage Virtualize 8.7 gives more flexibility in assigning and managing the system
management ports on the IBM FlashSystem and SVC machines.

Some of these changes are:


򐂰 Ability to configure system IP address on any port and with VLAN.
򐂰 Requirement of more routable data IP addresses:
– Have 4 routable data IP addresses per port per node.
򐂰 System defined default management portset for system IP addresses.
򐂰 Limiting number of system IP addresses to 2.
򐂰 System IP address on any port other than default port 1 and 2.
򐂰 VLAN support.

Changes in GUI for system IP addresses:


򐂰 Management IP address panel is changed.
򐂰 Adding new management IP addresses.

Reclaim space of thin-provisioned volumes in standard pools


This feature supplies a new mechanism to automatically reclaim thin-provisioned volumes in
standard pools. It automatically recovers space in standard pools after hosts submit unmap
I/O (or overwrite with zeros). This feature was developed in response to feedback from
customers who previously relied on manual processes and close monitoring.

1.3.6 Remote Copy support on IBM Storage Virtualize V8.7


IBM Storage Virtualize V8.7.0 will be the final release supporting all Remote Copy based
features, including HyperSwap®, Metro Mirror, Global Mirror and Global Mirror with Change
Volumes. When planning new deployments, consider using policy-based replication and
policy-based high availability to avoid causing an unnecessary migration.

Note: Remote Copy will be supported long-term on V8.7.0 for as long as the hardware has
a valid support contract. This also includes the following functions:
򐂰 Global Mirror
򐂰 Global Mirror with Change Volumes
򐂰 HyperSwap
򐂰 Metro Mirror
򐂰 Migration relationships
򐂰 HyperSwap and Metro Mirror 3-site solutions

Important: Entry-level IBM FlashSystem 5015 and 5035 will not have replication
capabilities if upgraded beyond V8.7.0.

򐂰 Global Mirror and Global Mirror with Change Volumes are replaced by policy-based
replication.
– A migration procedure is available in the product documentation.
򐂰 HyperSwap is replaced by policy-based HA.

12 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:24 am 8561ch01.fm

򐂰 Migration is using storage partition migration.

1.4 Preparation and upgrading to IBM Storage Virtualize V8.7.0


In order to run IBM Storage Virtualize V8.7.0 on your selected hardware, there are some
tasks and checks that need to be done prior to implementing this level of IBM Storage
Virtualize software.

Firstly, you need to confirm that your current hardware is able to support IBM Storage
Virtualize V8.7.0.

Figure 1-5 on page 13 shows the matrix of supported hardware versus the IBM Storage
Virtualize software levels.

The “from” level is your current IBM Storage Virtualize software level and the “to” level will be
IBM Storage Virtualize 8.7.0.

Figure 1-5 IBM Storage Virtualize upgrade support matrix

Take time to look at the above matrix and confirm your IBM Storage Virtualize hardware can
upgrade to the IBM Storage Virtualize 8.7.0 level.

Secondly, you will need to check that any outstanding issues or errors are corrected or fixed
prior to the upgrade. If required, place a service call with IBM for assistance with getting these
resolved.

Also, be aware that there are certain limits with features at IBM Storage Virtualize 8.7.0 that
might be applicable to your configuration. You need to take these into account before
upgrading to IBM Storage Virtualize 8.7.0. See here for the list:

Chapter 1. Introduction and system overview 13


8561ch01.fm Draft Document for Review July 22, 2024 9:24 am

V8.7.0.x Configuration Limits for IBM FlashSystem and SAN Volume Controller.

Finally, you will need to schedule time to do the upgrade to the new IBM Storage Virtualize
V8.7.0 software. This upgrade is concurrent, but it is recommended that it is done at a time of
low I/O activity as the upgrade works in parallel to the normal running so at low I/O times the
upgrade will happen sooner.

For upgrade process including the pre checks and code download can be found here in the
IBM Documentation section: Updating System Software.

It is also important to check that the drive firmware on your IBM FlashSystem is also kept up
to the latest level. This process is done separately to the IBM Storage Virtualize software
updating.

Details of the drive firmware upgrade process can be found here: Updating Drive Firmware.

14 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Chapter 2. Initial configuration


This chapter describes the initial configuration of the following actual systems:
򐂰 IBM FlashSystem 9500
򐂰 IBM FlashSystem 7300
򐂰 IBM FlashSystem 5300
򐂰 IBM FlashSystem 5200
򐂰 IBM FlashSystem 5045
򐂰 IBM FlashSystem 5015
򐂰 IBM SAN Volume Controller

It also provides step-by-step instructions for the initial setup process and defines the baseline
system settings. These settings are typically applied during the implementation phase, before
volume creation and provisioning.

This chapter includes the following topics:


򐂰 “Prerequisites” on page 16
򐂰 “System initialization” on page 17
򐂰 “System setup” on page 24

© Copyright IBM Corp. 2024. All rights reserved. 15


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

2.1 Prerequisites
Before initializing and setting up the system, ensure that the following prerequisites are met:
򐂰 The physical components fulfill all the requirements and are correctly installed, including:
– The FlashSystem control enclosures or IBM SAN Volume Controller nodes are
physically installed in the racks.
– The Ethernet and Fibre Channel (FC) cables are connected.
– The expansion enclosures (if available) are physically installed and attached to the
control enclosures that use them.
– The control enclosures or IBM SAN Volume Controller nodes and optional expansion
enclosures are powered on.

Note: IBM SAN Volume Controller nodes need enough time to charge the batteries.
How long it takes to recharge depends on how long it was waiting idle in stock and
not in production. You cannot start the nodes without a fully charged battery.

򐂰 The web browser that is used for managing the system is supported by the management
GUI. For the list of supported browsers, see Web browser requirements to access the
management GUI.
򐂰 The required information for remote management of the system, including:
– The IPv4 (or IPv6) addresses that are assigned for the system’s management
interfaces:
• The unique cluster IP address, which is the address that is used for the
management of the system.
• Unique service IP addresses, which are used to access node service interfaces.
You need one address for each IBM SAN Volume Controller node or IBM
FlashSystem node (two per control enclosure).
• The IP subnet mask for each subnet that is used.
• The IP gateway for each subnet that is used.
– The licenses that might be required to use specific functions. Whether these licenses
are required depends on the hardware that is used. For more information, see IBM
Documentation.
– Information that is used by a system when performing Call Home functions, such as:
• The company name and system installation address.
• The name, email address, and phone number of the storage administrator whom
IBM can contact if necessary.
– The following information is optional:
• The Network Time Protocol (NTP) server IP address.
• The Simple Mail Transfer Protocol (SMTP) server IP address, which is necessary
only if you want to enable Call Home or want to be notified about system events
through email.
• The IP addresses for Remote Support Proxy Servers, which are required only if you
want to use them with the Remote Support Assistance feature.

16 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Note: IBM FlashSystem 9500, and IBM SAN Volume Controller are installed by an IBM
System Services Representative (IBM SSR). Provide all the necessary information to the
IBM SSR by completing the following planning worksheets:
򐂰 IBM FlashSystems
򐂰 IBM SAN Volume Controller

After the IBM SSR completes their portion of the setup, see 2.3, “System setup” on
page 24 to continue the setup process.

2.2 System initialization


This section provides step-by-step instructions about how to create the system cluster.
Demonstration videos

You can view the following demonstration videos. Although the videos are based on IBM
Storage Virtualize V8.6, they are still applicable to V8.7.
򐂰 IBM Storage Virtualize V8.6 Initial setup: SSR configuration tasks
򐂰 IBM Storage Virtualize V8.6 Initial setup: Customer configuration tasks
򐂰 IBM Storage Virtualize V8.6 Initial setup: Setting up a cluster from the service IP

To start the initialization procedure, connect a workstation directly to the technician port. The
technician port is a dedicated 1 Gb Ethernet (GbE) port located on the rear of each storage
node canister. On all platforms except IBM FlashSystem 5015, it can only be used for
initializing or servicing the system. It cannot be connected to an Ethernet switch because it
supports only a direct connection to a workstation, and it remains disconnected after the
initial setup is complete.

On IBM FlashSystem 5015, the technician port is enabled initially. However, the port is
switched to internet Small Computer Systems Interface (iSCSI) host attachment mode after
the setup wizard is complete.

To re-enable an onboard Ethernet port on a system to be used as the technician port, refer to
the command shown in Example 2-1.

Example 2-1 Reenabling the onboard Ethernet port 2 as the technician port
IBM_IBM FlashSystem 9100:superuser>satask chserviceip -techport enable -force

Chapter 2. Initial configuration 17


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

The location of the technician port on an IBM FlashSystem 9500 is shown in Figure 2-1.

Figure 2-1 Technician port FlashSystem 9500

The location of the technician port of an IBM FlashSystem 7300 is shown in Figure 2-2.

Figure 2-2 Technician port FlashSystem 7300

The location of the technician port of an IBM FlashSystem 5200 is shown in Figure 2-3

Figure 2-3 Technician port FlashSystem 5200

The location of the technician port of an IBM FlashSystem 5300 is shown in Figure 2-4

Figure 2-4 Technician port FlashSystem 5300

The location of the technician port of an IBM FlashSystem 5045 is shown in Figure 2-5.

Figure 2-5 Technician port FlashSystem 5045

18 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

The location of the technician port of an IBM FlashSystem 5015 is shown in Figure 2-6.

Figure 2-6 Technician port FlashSystem 5015

The location of a technician port on the IBM SAN Volume Controller 2145-SV3 is shown in
Figure 2-7.

Figure 2-7 Technician port IBM SAN Volume Controller 2145-SV3

The location of a technician port on the IBM SAN Volume Controller 2145-SV2 is shown in
Figure 2-8.

Figure 2-8 Technician port IBM SAN Volume Controller 2145-SV2

The technician port runs an IPv4 DHCP server, and it can assign an address to any device
that is connected to this port. Ensure that your workstation Ethernet adapter is configured to
use a DHCP client if you want the IP to be assigned automatically.

If you prefer not to use DHCP, you can set a static IP on the Ethernet port from the
192.168.0.x/24 subnet; for example, 192.168.0.2 with the netmask 255.255.255.0.

The default IP address of a technician port on a node canister is 192.168.0.1. Do not use this
IP address for your workstation.

Note: Ensure that the technician port is not connected to the organization’s network. No
Ethernet switches or hubs are supported on this port.

2.2.1 System initialization process


Before initialization, each node canister in a new system remains in the candidate state and
cannot process I/O (Input/Output).

Chapter 2. Initial configuration 19


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

During initialization, the nodes within a single control enclosure are joined into a cluster. This
cluster is later configured to process data. (For an IBM SAN Volume Controller system, the
cluster initially consists of only one node.)

If your system has multiple control enclosures or IBM SAN Volume Controller nodes, only
initialize the first enclosure or node. The remaining enclosures or nodes can be added to the
cluster later using the cluster management interface (GUI or CLI) after the initial setup.

During initialization, you must specify an IPv4 or IPv6 system management address. This
address is assigned to Ethernet port 1 on each node and is used to access the management
GUI and CLI. You can configure additional IP addresses after the system is initialized.

Note: Do not perform the system initialization procedure on more than one node canister
of one control enclosure. After initialization completes, use the management GUI or CLI to
add control enclosures to the system.

To initialize a new system, complete the following steps:


1. Connect your workstation to a technician port of any canister of the control enclosure or
the IBM SAN Volume Controller system. Ensure that you obtained a valid IPv4 address
with DHCP.
2. Open a supported web browser and browse to https://install. The browser is
automatically redirected to the System Initialization wizard. You can also use the IP
address https://192.168.0.1 if you are not automatically redirected.

Warnings about untrusted certificates: During system initialization, you might see
warnings about untrusted certificates. This happens because the system uses
self-signed certificates, which are not verified by a well-known authority.

However, if you are directly connected to the service interface, there is no intermediary
that could impersonate the system with a fake certificate. Therefore, you can safely
accept the certificates in this scenario.

If the system is not in a state that allows initialization, the system does not start the
System Initialization wizard, and you are redirected to the Service Assistant interface. Use
the displayed error codes to troubleshoot the problem.

20 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

3. The window that is used to log in to Service Assistant opens (Figure 2-9). This window is
the first step of initializing the system. Enter the default superuser password (passw0rd)
and click Log in.

Figure 2-9 Logging in to Service Assistant by way of the technician port

4. The System Initialization wizard shows the detected canisters, as shown in Figure 2-10.
Click Proceed to continue (This window is not shown for IBM SAN Volume Controller
nodes).

Figure 2-10 System Initialization: Canister detection

Chapter 2. Initial configuration 21


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

5. The System Initialization wizard shows the enclosure assignment. Select As the first
enclosure in a new system, as shown in Figure 2-11. Click Next.

Figure 2-11 System Initialization: Initialize the first enclosure

If you select As an additional enclosure in an existing system, you are prompted to


disconnect from the technician port and use the GUI of the system to which the new nodes
are to be added.
For IBM SAN Volume Controller systems, the initialization window might differ (see
Figure 2-12). You will likely be prompted to add nodes directly, rather than enclosures.
If you select As an additional node in an existing system, the process will instruct you
to disconnect from the technician port and use the system's GUI for further configuration.

Figure 2-12 System Initialization: Initialize the first IBM SAN Volume Controller node

22 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

6. Enter the management IP address information for the new system as shown in
Figure 2-13. Set the IP address, network mask, and gateway. Click Next.

Figure 2-13 System Initialization: Enter Management IP

7. A window that includes a restart timer opens (Figure 2-14). When the timeout is reached,
the window is updated to reflect success or failure. Failure occurs if the system is
disconnected from the network, which prevents the browser from updating with the IBM
FlashSystem web server.

Figure 2-14 System Initialization: Web-server restart timer counting down from 5 minutes

Chapter 2. Initial configuration 23


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

8. The System Initialization completed wizard is shown in Figure 2-15. Click Finish.

Figure 2-15 System Initialization completed

Follow the instructions, and direct your browser to the management IP address to access
the system GUI after you click Finish.
System Setup is also available directly from the technician port. The System Setup wizard
is available through both the management IP address and the technician port.

2.3 System setup


This section provides instructions about how to define the basic settings of the system by
using the System Setup wizard.

2.3.1 System Setup wizard


Complete the System Setup wizard to define the basic settings of the system. After the
initialization is complete, you are redirected to a management GUI from your workstation, or
you browse to the management IP address of a freshly initialized system from another
workstation.

The first time that you connect to the management GUI, you can be prompted to accept
untrusted certificates because the system certificates are self-signed. If your company policy
requests certificates that are signed by a trusted certificate authority (CA), you can install
them after you complete the System Setup.

To finish the System Setup wizard, complete the following steps:

24 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

1. Log in to system GUI. Until the wizard is complete, you can use only the superuser
account, as shown in Figure 2-16. Click Sign in.

Note: The default password for the superuser account is passw0rd (with the number
zero, not the uppercase O). The default password must be changed by using the
System Setup wizard or after the first CLI login. The new password cannot be set to the
default password.

Figure 2-16 Logging in for the first time

Chapter 2. Initial configuration 25


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

2. The Initial Setup starts with the Welcome page, as shown in Figure 2-17. Click Next.

Figure 2-17 Initial Setup Window

3. The next step in the initial setup is the setup of Call Home. See Figure 2-18 on page 27.
Call Home provides multiple benefits. It enables automatically creating tickets at IBM if
errors occur, which improves the speed and efficiency by which calls are handled. Call
Home also enables Storage Insights and Remote Support.

Note: Refer to “Call Home” on page 97 for more information on Call Home.

26 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-18 Setup Call Home

On IBM FlashSystem 9500 systems and IBM SAN Volume Controller systems, an IBM
SSR configures Call Home during installation. Verify that all the entered data is correct.
All IBM FlashSystem products and IBM SAN Volume Controller systems support the
following methods of sending Call Home notifications to IBM:
– Cloud Call Home
– Call Home with email notifications
Cloud Call Home is the default and preferred option for a system to report event
notifications to IBM Support. With this method, the system uses RESTful application
programming interfaces (APIs) to connect to an IBM centralized file repository that
contains troubleshooting information that is gathered from customers. This method
requires no extra configuration.
The system can also be configured to use email notifications for this purpose. If this
method is selected, you are prompted to enter the SMTP server IP address.
If both methods are enabled, Cloud Call Home is used, and the email notifications method
is kept as a backup.
If either of these methods is selected, the system location and contact information must be
entered. This information is used by IBM to provide technical support. All fields in the form
must be completed. In this step, the system also verifies that it can contact the Cloud Call
Home servers.
4. If you click Next you will be asked for the Transmission Type for Call Home.
5. System Setup prompts to select which transmission types to be used for Call Home. See
Figure 2-19 on page 28.

Note: It is not recommended to select I don’t want to use Call Home.

Chapter 2. Initial configuration 27


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-19 Transmission Types for Call Home

6. Select your choice. In our case we choose Send using Cloud services. Click Apply and
Next to setup the Internal Proxy Server. See Figure 2-20 on page 29. Enter the requested
information.

28 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-20 Setup Internal Proxy Server

7. After you setup the Proxy Server the system checks the connection to the Support Center.
See Figure 2-21 on page 30.

Chapter 2. Initial configuration 29


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-21 Connection Test to the Support Center

8. Enter all required information for the System Location. After clicking Next you should fill in
the Contact information. Figure 2-22 on page 31 shows the panel for the contact
information. Use the Company contact information to comply with privacy regulations. IBM
may use the contact data if you allow it. To complete the registration click Apply and Next.

30 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-22 System Location

9. The System provides you the Summary information. If all is correct click Finish. See
Figure 2-23 on page 32.

Chapter 2. Initial configuration 31


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-23 Summary page

10.The system saves the entered information and asks you to log in again. After login, you
will be guided to the System Setup page. See Figure 2-24 on page 33.

32 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-24 System Setup Welcome page

11.Clicking Next brings you to the License agreement page. Carefully read the license
agreement. Select I agree with the terms in the license agreement if you want to
continue the setup, otherwise the system will stop the setup. See Figure 2-25 on page 34.

Chapter 2. Initial configuration 33


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-25 Accept License Agreement

12.You are asked to change the password shown in Figure 2-26 on page 35. Enter a new
password for superuser. A valid password is 8 - 64 characters and cannot begin or end
with a space. Also, the password cannot be set to match the default password.

Note: All configuration changes that are made by using the System Setup wizard are
applied immediately, including the password change. The user sees the system running
commands during the System Setup wizard.

34 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-26 Change password

13.To confirm press Apply and Next.


Next step is to provide a System Name as shown in Figure 2-27 on page 36. Avoid the use
of an underscore (_) in a system name. Although permitted here, it cannot be used in
domain name server (DNS) shortnames and fully qualified domain names (FQDNs).
Therefore, such naming might cause confusion and access issues. The following
characters can be used: A - Z, a - z, 0 - 9, and - (hyphen).

Note: In a 3-Site Replication solution, ensure that the system name is unique for all
three clusters to prepare the IBM Storage Virtualize clusters at Master, AuxNear, and
AuxFar sites to work. The system names must remain different for the life of the 3-site
configuration.

For more information about 3-Site Replication, see IBM Spectrum Virtualize 3-Site
Replication, SG24-8504.

If required, the system name can be changed by running the chsystem -name
<new_system_name> command. The system can also be renamed by using the
management GUI by clicking Monitoring → System Hardware and selecting System
Actions → Rename System.

Chapter 2. Initial configuration 35


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-27 System Name

14.Click Apply and Next.


15.After you provide the System Name you will be asked to enter the required, additional
licenses for each function. Figure 2-28 on page 37 shows you an example.

36 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-28 License Functions

The window for this step in the system setup might look different, depending on the
systems that are used. Also, the way the license is enabled depends on the system that is
used.

Note: Encryption uses a key-based licensing scheme.

16.When done, click Apply and Next.


17.DNS can be configured on the system, as shown in Figure 2-29 on page 38. DNS helps
the system to resolve the names of the computer resources that are in the external
network if they are not indicated by an IP address.

Chapter 2. Initial configuration 37


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-29 DNS Server setup

18.Following the DNS Setup you are asked to set Date and Time. Enter the date and time
settings. In the example that is shown in Figure 2-30 on page 39, the date and time are set
by using manually settings. Generally, use an NTP server so that all of your storage area
network (SAN) and storage devices have a common timestamp. This practice facilitates
troubleshooting and prevents time stamp-related errors.

38 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-30 Date and Time

19.When done, click Apply and Next.


20.If you have purchased an Encryption License, activate it here. See Figure 2-31 on
page 40.

Chapter 2. Initial configuration 39


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-31 Activate Encryption License

If encryption is not planned now, select No and then click Next. You can enable this
feature later.

Note: When encryption is enabled on the system, encrypted storage pools can be
created. If the system is a single control enclosure system where all FCM-drives should
be in the same storage pool, encryption must be enabled before creating the storage
pool. If a storage pool is created before encryption is enabled, any data in that pool
must be migrated to an encrypted storage pool, if the data must be encrypted.

If you purchased the encryption feature, you are prompted to activate your license
manually or automatically. The encryption license is key-based and required for each
control enclosure.
You can use automatic activation if the workstation that you use to connect to the GUI and
run the System Setup wizard has Internet access. If no Internet connection is available,
use manual activation and follow the instructions.

40 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-32 Encryption licensed

21.After the encryption license is activated, you see a green check mark for each enclosure,
as shown in Figure 2-32 on page 41. After all the control enclosures show that encryption
is licensed, click Next.
22.If you wish to modify your previously entered Call Home settings, you can do so here. See
Figure 2-33 on page 42.

Chapter 2. Initial configuration 41


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-33 Change Call Home settings

23.If you want to use the Support Assistance offering from IBM you can choose your
options here (Figure 2-34 on page 43). If you enabled at least one Call Home method, the
Support Assistance configuration window opens. The Support Assistance function
requires Call Home; therefore, if it is disabled, Support Assistance cannot be used. Click
Next to continue.

Note: Refer to “Remote Support Assistance” on page 98 for more information.

42 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-34 Setup Support Assistance

With the Support Assistance feature, you allow IBM Support to perform maintenance
tasks on your system with support personnel onsite or remotely.
If an IBM SSR is onsite, the SSR can log in locally with your permission and a special user
ID and password so that a superuser password does not need to be shared with the IBM
SSR.
You can also enable Support Assistance with remote support to allow IBM Support
personnel to log in remotely to the machine with your permission through a secure tunnel
over the Internet.
If you allow remote support, you are provided with the IP addresses and ports of the
remote support centers and an opportunity to provide proxy server details (if required) to
allow the connectivity, as shown in Figure 2-35 on page 44. Click Apply and Next.

Chapter 2. Initial configuration 43


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-35 System communicating with named IBM Support servers

You can also allow remote connectivity at any time or only after obtaining permission from
the storage administrator, as shown in Figure 2-36 on page 45.

44 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-36 Remote support access settings

24.Click Apply and Next.


25.If you install your system below an IBM San Volume Controller, you can use Automatic
Configuration for Virtualization. System Setup offers this option (IBM FlashSystem
products only) to automatically configure the system if it is used as FC-attached, back-end
storage for IBM SAN Volume Controller. If you plan to use the system in stand-alone mode
(that is, not behind an IBM SAN Volume Controller), leave Automatic Configuration turned
off, as shown Figure 2-37 on page 46. Click Next to continue.

Chapter 2. Initial configuration 45


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-37 Automatic Configuration for Virtualization

For more information about how to enable Automatic configuration for IBM SAN Volume
Controller on a running system after the System Setup wizard, see 2.3.7, “Automatic
configuration for IBM SAN Volume Controller back-end storage” on page 61.
26.On the Summary page, the settings that were selected by the System Setup wizard are
shown. If corrections are needed, you can return to a previous step by clicking Back.
Otherwise, click Finish to complete the system setup wizard shown Figure 2-38 on
page 47.

46 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-38 Summary Page

When the system setup wizard completes, your IBM FlashSystem consists only of the control
enclosure that includes the node canister that you used to initialize the system and its partner,
and the expansion enclosures that are attached to them.

In the case of an IBM SAN Volume Controller, your system consists of only one node in the
cluster, which might see other candidate nodes in the service GUI if they are connected to
SAN and zoned together.

If you have other control and expansion enclosures or IBM SAN Volume Controller nodes, you
must add them to complete the System Setup.

For more information about how to add a control or expansion enclosure, see 2.3.2, “Adding
an enclosure in IBM FlashSystem” on page 49.

For more information about how to add a node or hot spare node, see 2.3.3, “Adding a node
or hot spare node in IBM SAN Volume Controller systems” on page 51.

If no other enclosures or nodes are to be added to this system, the System Setup process is
complete and you can click Finish to be returned to the login window of the IBM
FlashSystem.

All the required steps of the initial configuration are complete. If needed, you can configure
other global functions, such as system topology, user authentication, or local port masking
before configuring the volumes and provisioning them to hosts.

Chapter 2. Initial configuration 47


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

27.With pressing Finish you confirm your choices. See Figure 2-39.

Figure 2-39 System Initialization

Click Close as shown in Figure 2-39.


28.After Setup is completed you will be guided to the Management GUI. See Figure 2-40 on
page 48.

Figure 2-40 Setup completed

29.Clicking Close and Finish takes you to the Dashboard (see Figure 2-41 on page 49).

48 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-41 Dashboard

The tasks that are described next are used to define global system configuration settings.
Often, they are performed during the System Setup process. However, they can also be
performed later, such as when the system is expanded or the system environment is
reconfigured.

2.3.2 Adding an enclosure in IBM FlashSystem


This procedure is the same whether you are configuring the system for the first time or
expanding it. When performed by using the system GUI, the same steps are used for adding
expansion or control enclosures.

Before beginning this process, ensure that the new control enclosure is correctly installed and
cabled to the system.

For FC node-to-node communication, verify that the correct SAN zoning is set.

For node-to-node communication over RDMA-capable Ethernet ports, ensure that the IP
addresses are configured and a connection between nodes can be established.

To add an enclosure to the system, complete the following steps:


1. In the GUI, select Monitoring → System Hardware. When a new enclosure is detected
by a system, the Add Enclosure button appears on the system next to System Actions, as
shown in Figure 2-42.

Figure 2-42 Add Enclosure button

Chapter 2. Initial configuration 49


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Note: If the Add Enclosure button does not appear, review the installation instructions
to verify that the new enclosure is connected and set up correctly.

2. Click Add Enclosure, and a list of available candidate enclosures opens, as shown in
Figure 2-43. To light the Identify light-emitting diode (LED) on a selected enclosure, select
Actions → Identify. When the required enclosure (or enclosures) is chosen, click Next.

Figure 2-43 Selecting the control enclosure to add

3. Review the summary in the next window and click Finish to add the expansion enclosure
or the control enclosure and all expansions that are attached to it to the system.

Note: When a new control enclosure is added, the software version that is running on
its nodes is upgraded or rolled back to match the system software version. This process
can take up to 30 minutes or more, and the enclosure is added only when this process
completes.

4. After the control enclosure is successfully added to the system, a success message
appears. Click Close to return to the System Overview window and check that the new
enclosure is visible and available for management.
5. To perform the same procedure by using a CLI, complete the following steps. For more
information about the detailed syntax for each command, see Command-line interface.
a. When you add control enclosures, check for unpopulated I/O groups by running the
lsiogrp command. Because each control enclosure includes two nodes, it forms an
I/O group.
Example 2-2 shows that only io_grp0 has nodes. Therefore, a new control enclosure
can be added to io_grp1.

Example 2-2 Listing the I/O groups


IBM_IBM FlashSystem:ITSO-FS9500:superuser>lsiogrp
id name node_count vdisk_count host_count site_id site_name
0 io_grp0 2 0 0
1 io_grp1 0 0 0
2 io_grp2 0 0 0
3 io_grp3 0 0 0
4 recovery_io_grp 0 0 0

50 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

b. To list control enclosures that are available to add, run the


lscontrolenclosurecandidate command, as shown in Example 2-3. To list the
expansion enclosures, run the lsenclosure command. Expansions that have the
managed parameter set to no can be added.

Example 2-3 Listing the candidate control enclosures


IBM_IBM FlashSystem:ITSO-FS9500:superuser>lscontrolenclosurecandidate
serial_number product_MTM machine_signature
78E005D 9848-AF8 4AD2-EA69-8B5E-D0C0

c. Add a control enclosure by running the addcontrolenclosure command, as shown in


Example 2-4. The command triggers only the process, which starts in the background
and can take 30 minutes or more.

Example 2-4 Adding a control enclosure


IBM_IBM FlashSystem:ITSO-FS9500:superuser>addcontrolenclosure -iogrp 1 -sernum
78E005D

d. To add an expansion enclosure, change its managed status to yes by running the
chenclosure command, as shown in Example 2-5.

Example 2-5 Adding an expansion enclosure


IBM_IBM FlashSystem:ITSO-FS9500:superuser>lsenclosure
id status type managed IO_group_id IO_group_name product_MTM serial_number
1 online control yes 0 io_grp0 9848-AF8 78E006A
2 online expansion no 0 io_grp0 9848-AFF 78CBVF5

IBM_IBM FlashSystem:ITSO-FS9500:superuser>chenclosure -managed yes 2

2.3.3 Adding a node or hot spare node in IBM SAN Volume Controller systems
This procedure is the same whether you are configuring the system for the first time or
expanding it later. The same process is used to add a node to an I/O group, or a hot spare
node.

Before beginning this process, ensure that the new control enclosure is correctly installed and
cabled to the system.

For FC node-to-node communication, verify that the correct SAN zoning is set.

For node-to-node communication over RDMA-capable Ethernet ports, ensure that the IP
addresses are configured and a connection between nodes can be established.

To add a node to the system, complete the following steps:


1. In the GUI, select Monitoring → System Hardware. When a new enclosure is detected
by a system, the Add Node button appears on the System - Overview window next to
System Actions, as shown in Figure 2-44.

Chapter 2. Initial configuration 51


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-44 Add Node button

Note: If the Add Node button does not appear, review the installation instructions to
verify that the new node is connected and set up correctly.

2. Click Add Node. A form that you can use to assign nodes to I/O groups opens, as shown
in Figure 2-45. To light the Identify light-emitting diode (LED) on a node, click the LED icon
that is next to a node name. When the required node (or nodes) is selected, click Finish.

Figure 2-45 Adding a node

The Monitoring → Systems Hardware window now changes and shows that the node is
added, as shown in Figure 2-46.

52 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-46 IBM SAN Volume Controller is adding node to the cluster

The node is added, as shown in Figure 2-47.

Figure 2-47 Node added

Note: When a node is added, the software version that is running is upgraded or rolled
back to match the cluster software version. This process can take 30 minutes or more
to complete. The node is added only after this process completes.

Chapter 2. Initial configuration 53


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

To perform the same procedure by using a CLI, complete the following steps. For more
information about the detailed syntax for each command, see this IBM Documentation web
page.
1. When adding nodes, check for unpopulated I/O groups by running lsiogrp. Each
complete I/O group has two nodes. Example 2-6 shows that only io_grp0 has nodes;
therefore, a new control enclosure can be added to io_grp1.

Example 2-6 Listing I/O groups


IBM_2145:ITSO-SVC:superuser>lsiogrp
id name node_count vdisk_count host_count site_id site_name
0 io_grp0 2 0 0
1 io_grp1 0 0 0
2 io_grp2 0 0 0
3 io_grp3 0 0 0
4 recovery_io_grp 0 0 0

2. To list nodes that are available to add to the I/O group, run the lscnodecandidate
command, as shown in Example 2-7.

Example 2-7 Listing the candidate nodes


BM_2145:ITSO-SVC:superuser>lsnodecandidate
id panel_name UPS_serial_number UPS_unique_id hardware
serial_number product_mtm machine_signature
500507680C00D98F 78KKLD0 500507680C00D98F SV1 78KKLD0
2145-SV1 3F25-557E-21E6-2B7D
500507680C00D98A 78KKCH0 500507680C00D98A SV1 78KKCH0
2145-SV1 702D-D5FE-76AA-4034

3. Add a node by running the addnode command, as shown in Example 2-8. The command
triggers only the process, which starts in the background and can take 30 minutes or
more.

Example 2-8 Adding a node as a spare


IBM_2145:ITSO-SVC:superuser>addnode -panelname 78KKLD0 -spare
Node, id [3], successfully added

Example 2-9 shows same command, but used to add a node to an I/O group io_grp1.

Example 2-9 Adding a node to an I/O group


IBM_2145:ITSO-SVC:superuser>addnode -panelname 78KKCH0 -name node3 -iogrp 1
Node, id [4], successfully added

4. Check the nodes in the system by using CLI. As shown in Example 2-10, the IBM SAN
Volume Controller is configured with two nodes, which forms one IO-group. A spare node
is configured for the IO-group.

Example 2-10 Single IO-group (two nodes) and one spare


IBM_2145:ITSO-SVC:superuser>lsnode
id name UPS_serial_number WWNN status IO_group_id IO_group_name config_node
UPS_unique_id hardware iscsi_name iscsi_alias panel_name
enclosure_id canister_id enclosure_serial_number site_id site_name
1 node1_78KKLC0 500507680C00D990 online 0 io_grp0 yes
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node178kklc0 78KKLC0
2 node2_78KKCD0 500507680C00D982 online 0 io_grp0 no
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node278kkcd0 78KKCD0

54 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

3 spare1 500507680C00D98F spare no


SV1 78KKLD0

A two-IO-group system with no spare is shown in Example 2-11.

Example 2-11 Two IO-groups (four nodes) configured- no spare


IBM_2145:ITSO-SVC:superuser>lsnode
id name UPS_serial_number WWNN status IO_group_id IO_group_name config_node
UPS_unique_id hardware iscsi_name iscsi_alias panel_name
enclosure_id canister_id enclosure_serial_number site_id site_name
1 node1_78KKLC0 500507680C00D990 online 0 io_grp0 yes
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node178kklc0 78KKLC0
2 node2_78KKCD0 500507680C00D982 online 0 io_grp0 no
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node278kkcd0 78KKCD0
3 node3_78KKCH0 500507680C00D98A online 1 io_grp1 no
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node378kkch0 78KKCH0
4 node4_78KKLD0 500507680C00D98F online 1 io_grp1 no
SV1 iqn.1986-03.com.ibm:2145.itso-svc.node478kkld0 78KKLD0

The administrator might want to rename the nodes to feature consistent names. This
process can be done by clicking Monitoring → System Hardware → Node Actions →
Rename.

2.3.4 Business continuity with policy-based high availability


Business continuity ensures an organization can deliver services even during disruptions.
While some applications might tolerate temporary outages, major disasters can cause
significant downtime and data loss, leading to immense recovery costs. Organizations should
minimize data loss and downtime to lessen business impact and financial strain.

From a storage perspective, business continuity involves maintaining data consistency and
availability for uninterrupted application access, achieved through two key concepts: disaster
recovery (DR) and high availability (HA). DR focuses on replicating data to remote locations
for recovery, while HA prioritizes continuous data accessibility.

Disasters can range from entire site outages to data corruption or theft. Data protection relies
on local or remote backups. IBM Storage Virtualize offers functionalities to safeguard your
data against various threats, such as hardware failures, software errors, or cyberattacks.
Policy-based replication and policy-based high availability protect against site failures by
automatically failing over to a secondary site, ensuring business continuity. While not covered
here, Storage Virtualize offers additional features like snapshots and Safeguarded snapshots
to protect against data corruption or cyberattacks.

Note: Policy-based high availability is not supported by the IBM FlashSystem 5015.

For more information about this topic refer to IBM Redbooks Ensuring Business Continuity: A
Practical Guide to Policy-Based Replication and Policy-Based High Availability for IBM
Storage Virtualize Systems, SG24-8569.

Chapter 2. Initial configuration 55


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

2.3.5 Configuring quorum disks or applications


Quorum devices are required for a system to hold a copy of important system configuration
data. An internal drive of an IBM FlashSystem, a managed disk (MDisk) from FC-attached
external back-end storage, or a special application that is connected over an IP network can
work as a quorum device.

One of these items is selected for the active quorum role, which is used to resolve failure
scenarios where half the nodes on the system become unavailable or a link between
enclosures is disrupted. The active quorum determines which nodes can continue processing
host operations. It also avoids a “split brain” condition, which occurs when both halves of the
system continue I/O processing independently of each other.

For IBM FlashSystem products with a single control enclosure and IBM SAN Volume
Controller systems with a standard topology, quorum devices are automatically selected from
the internal drives or assigned from an MDisk, respectively. No special configuration actions
are required. This function also applies for IBM FlashSystem products with multiple control
enclosures, a standard topology, and virtualizing external storage.

For policy-based HA (or former HyperSwap and Enhanced Stretched Cluster) topology
systems, configure an active quorum device on a third, independent site. Because of the
costs associated with deploying a separate FC-attached storage device on a third site, an
IP-based quorum device can be used for this purpose.

Without a third arbitration site (quorum server), a tie-breaker mechanism must be chosen for
the two existing sites. During a network outage between the sites, the pre-configured "winner"
will continue operating and processing I/O requests. The "loser" site will be unavailable until
the connection is restored. IP quorum settings, within the configuration options, determine the
preferred site for handling these scenarios. If a site outage occurs at the winning site, the
system stops processing I/O requests until this site is recovered or the manual quorum
override procedure is used.

On IBM FlashSystem products in a standard topology system with two or more control
enclosures and no external storage, none of the internal drives can be the active quorum
device. For such configurations, it is a best practice to deploy an IP-based quorum application
to avoid a “split brain” condition.

Creating and installing an IP quorum application


To create and install an IP quorum application, complete the following steps:
1. Select Settings → System → IP Quorum to download the IP quorum application, as
shown in Figure 2-48. If you use IPv6 for management IP addresses, the Download IPv6
Application button is available and the IPv4 option is disabled. In our example, we select
Download IPv4 Application.

56 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

Figure 2-48 Download IPv4 quorum button

2. Click Download... and a window opens, as shown in Figure 2-49. It provides an option to
create an IP application that is used for tie-breaking only, or an application that can be
used as a tie-breaker and to store recovery metadata.

Figure 2-49 Download IP quorum application window

An application that does not store recovery metadata requires less channel bandwidth for
a link between the system and the quorum app, which might be a decision-making factor
for using a multi-site HA system.
For a full list of IP quorum app requirements, see IP quorum application configuration.
3. Click OK. The ip_quorum.jar file is created. Save the file and transfer it to a supported
AIX, Linux, or Windows host that can establish an IP connection to the service IP address
of each system node. Move it to a separate directory and start the application, as shown in
Example 2-12.

Example 2-12 Starting the IP quorum application on the Windows operating system
C:\IPQuorum>java -jar ip_quorum.jar
=== IP quorum ===
Name set to null.
Successfully parsed the configuration, found 2 nodes.
Trying to open socket
Trying to open socket

Chapter 2. Initial configuration 57


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Handshaking
Handshaking
Waiting for UID
Creating UID
*Connecting
Connected to 10.0.0.42
Connected to 10.0.0.41

Note: Add the IP quorum application to the list of auto-started applications at each start
or restart or configure your operating system to run it as an auto-started service in the
background. The server hosting the IP quorum application must reside within the same
network subnet as the IBM FlashSystem for proper communication. Up to five IP
quorums can be deployed in your environment.

The IP quorum log file and recovery metadata are stored in the same directory with the
ip_quorum.jar file.
4. Check that the IP quorum application is successfully connected and running by verifying
its Online status by selecting Settings → System → IP Quorum, as shown in
Figure 2-50.

Figure 2-50 IP quorum application that is deployed and connected

Configuring the IP quorum mode


On a standard topology system, only the Standard quorum mode is supported. No other
configuration is required. On a policy-based HA topology, you can configure the following
tie-breaker scenarios (a tie occurs when half of the nodes that were a member of the system
are present):
򐂰 If the quorum mode is set to Standard, both sites have an equal chance to continue
working after the tie breaker.
򐂰 If the quorum mode is set Preferred, during a disruption, the system delays processing
tie-breaker operations on non-preferred sites, which leaves more time for the preferred site
to win. If during an extended period a preferred site cannot contact the IP quorum
application (for example, if it is destroyed), a non-preferred site continues working.
򐂰 If the quorum mode is set to Winner, the selected site is always the tie-breaker winner. If
the winner site is destroyed, the remaining site can continue operating only after manual
intervention.

58 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

The Preferred quorum mode is supported by an IP quorum only.

To set a quorum mode, select Settings → System → IP Quorum and then click Quorum
Setting. The Quorum Setting window opens, as shown in Figure 2-51.

Figure 2-51 Changing the quorum mode

2.3.6 Configuring the local Fibre Channel port masking


With FC port masking, you control the use of FC ports. By applying a mask, you restrict
node-to-node communication or replication traffic on selected ports.

To set the FC port mask by using the GUI, complete the following steps:
1. Select Settings → Network → Fibre Channel Ports. In a displayed list of FC ports, the
ports are grouped by a system port ID. Each port is configured identically across all nodes
in the system. You can click the arrow next to the port ID to expand a list and see which
node ports (N_Port) belong to the selected system port ID and their worldwide port names
(WWPNs).

Chapter 2. Initial configuration 59


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

2. Right-click a system port ID that you want to change and select Modify Connection, as
shown in Figure 2-52.

Figure 2-52 Applying a port mask by using a GUI

By default, all system ports can send and receive traffic of any kind, including the following
examples:
– Host traffic
– Traffic to virtualized back-end storage systems
– Local system traffic (node to node)
– Partner system (remote replication) traffic
The first two types are always allowed, and you can control them only with SAN zoning.
The other two types can be blocked by port masking.
3. In the Modify Connection dialog box (see Figure 2-53 on page 60), you can choose which
type of traffic a port can send; for example, Remote if the port is dedicated to Remote
Replication traffic. Click Modify when done.

Figure 2-53 Modify Connection dialog box

The following types of traffic are allowed for each choice:


򐂰 Any: A port can work with all types of traffic.
򐂰 Local: Remote replication traffic is blocked on this port.

60 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

򐂰 Remote: Local node-to-node traffic is blocked.


򐂰 None: Local and remote systems traffic is allowed, but system-to-host and
system-to-back-end storage communication still exists.

Port masks can also be set by using the CLI. Local and remote partner port masks are
internally represented as a string of zeros and ones. The last digit in the string represents port
one. The previous digits represent ports two, three, and so on.

If the digit for a port is set to 1, the port is enabled for the specific type of communication. If it
is set to 0, the system does not send or receive traffic that is controlled by a mask on the port.

To view the current port mask settings, run the lssystem command, as shown in
Example 2-13. The output shows that all system ports allow all types of traffic.

Example 2-13 Viewing the local port mask


IBM_IBM FlashSystem:ITSO-FS9500:superuser>lssystem |grep mask
local_fc_port_mask 1111111111111111111111111111111111111111111111111111111111111111
partner_fc_port_mask 1111111111111111111111111111111111111111111111111111111111111111

To set the local or RPORT mask, run the chsystem command. Example 2-14 shows the mask
setting for a system with four FC ports on each node and that has RC relationships. Masks
are applied to allow local node-to-node traffic on ports 1 and 2, and replication traffic on ports
3 and 4.

Example 2-14 Setting a local port mask by running the chsystem command
IBM_IBM FlashSystem:ITSO-FS9500:superuser>chsystem -localfcportmask 0011
IBM_IBM FlashSystem:ITSO-FS9500:superuser>chsystem -partnerfcportmask 1100
IBM_IBM FlashSystem:ITSO-FS9500:superuser>lssystem |grep mask
local_fc_port_mask 0000000000000000000000000000000000000000000000000000000000000011
partner_fc_port_mask 0000000000000000000000000000000000000000000000000000000000001100

The mask is extended with zeros, and all ports that are not set in a mask have the selected
type of traffic blocked.

Note: When replacing or upgrading your node hardware, consider that the number of FC
ports and their arrangement might be changed. If so, make sure that any configured port
masks are still valid for the new configuration.

2.3.7 Automatic configuration for IBM SAN Volume Controller back-end


storage
If a system is supposed to work as FC-attached back-end storage for IBM SAN Volume
Controller, you can enable Automatic Configuration for Virtualization during the initial System
Setup or anytime later by selecting Settings → System → Automatic Configuration.

Automatic Configuration for Virtualization is intended for a new system. If host, pool, or
volume objects are configured, all the user data must be migrated out of the system and
those objects must be deleted.

The Automatic Configuration for Virtualization wizard starts immediately after you complete
the initial setup wizard if you set Automatic Configuration to On.

Chapter 2. Initial configuration 61


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Complete the following steps:


1. Click Settings → System → Automatic Configuration. Then, select Automatic
Configuration ON and click Save, as shown in Figure 2-54.

Figure 2-54 Automatic Configuration wizard enablement

2. If any control or expansion enclosures must be included as part of the external storage to
be virtualized, you can add them. If you do not have more enclosures to add, this part of
the prerequisite steps can be skipped.
Click Add Enclosure to start the adding process, or click Skip to move to the next step
(see Figure 2-55).

Figure 2-55 Automatic configuration: Add Enclosure

Note: You can turn off the Automatic Configuration for Virtualization wizard at any step
by clicking the dotted symbol in the upper right corner.

62 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

3. The wizard checks whether the IBM SAN Volume Controller is correctly zoned to the
system. By default, newly installed systems run in N_Port ID Virtualization (NPIV) mode
(Target Port Mode). The system’s virtual (host) WWPNs must be zoned for IBM SAN
Volume Controller. On the IBM SAN Volume Controller side, physical WWPNs must be
zoned to a back-end system independently of the NPIV mode setting.
4. Create a host cluster object for IBM SAN Volume Controller. Each IBM SAN Volume
Controller node has its own worldwide node name (WWNN). Make sure to select all
WWNNs that belong to nodes of the same IBM SAN Volume Controller cluster.
Figure 2-56 shows that because the system detected an IBM SAN Volume Controller
cluster with dual I/O groups, four WWNNs are selected.

Figure 2-56 Defining a host cluster

5. When all nodes of an IBM SAN Volume Controller cluster (including the spare cluster) are
selected, you can change the host object name for each one, as shown in Figure 2-57 on
page 64. For convenience, name the host objects to match the IBM SAN Volume
Controller node names or serial numbers.

Chapter 2. Initial configuration 63


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 2-57 Hosts inside an IBM SAN Volume Controller host cluster

6. Click Automatic Configuration and check the list of internal resources that are used, as
shown in Figure 2-58.

Figure 2-58 Begin the automatic configuration process

64 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch02.fm

7. If the system uses compressed drives (FCM drives), you are prompted to enter your
expected compression ratio (or total capacity that is to be provisioned to IBM SAN Volume
Controller), as shown in Figure 2-59. If IBM SAN Volume Controller uses encryption or
writes data that is not compressible, set the ratio to 1:1 and then, click Next.

Figure 2-59 Automatic pool configuration

8. Review the pool (or pools) configuration, as shown in Figure 2-60, and click Proceed to
trigger the commands that applies it.

Figure 2-60 Pools configuration

Chapter 2. Initial configuration 65


8561ch02.fm Draft Document for Review July 19, 2024 5:44 pm

9. When the Automatic Configuration for Virtualization wizard completes, you see the
window that is shown in Figure 2-61. After clicking Close, you can proceed to the IBM
SAN Volume Controller GUI and configure a new provisioned storage.

Figure 2-61 Automatic configuration running commands

10.You can export the system volume configuration data in .csv format by using this window
or anytime later by selecting Settings → System → Automatic Configuration.

Figure 2-62 Automatic configuration complete

66 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

Chapter 3. Step-by-step configuration


This chapter describes the steps used to quickly implement a Storage Virtualize system
utilizing the most commonly used functions. Where applicable it will provide links to further
reading for advanced functions.

In this chapter we discuss the Storage Virtualize GUI, steps needed for network configuration,
creating pools and assigning storage, configuring hosts, basic snapshot and asynchronous
replication configuration.

This chapter has the following chapters:


򐂰 “The Storage Virtualize GUI” on page 68
򐂰 “Network configuration” on page 69
򐂰 “Pools and managed disks configuration” on page 71
򐂰 “Configuring volumes” on page 78
򐂰 “Configuring hosts” on page 79
򐂰 “Snapshots and replication” on page 84

© Copyright IBM Corp. 2024. All rights reserved. 67


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

3.1 The Storage Virtualize GUI


Throughout this chapter we will be using the Storage Virtualize GUI. The GUI is a built-in
software component within the IBM Storage Virtualize Software. Multiple users can be logged
in to the GUI. However, because no locking mechanism exists, be aware that the last action
that is entered from the GUI is the action that takes effect if two users change the same object
simultaneously.

3.1.1 Accessing the GUI


To access the IBM GUI, enter the IP address that was set during the initial setup process into
your web browser. You can connect from any workstation that can communicate with the
system.

Recommendation: It is a recommended practice for each user to have their own unique
account

The default user accounts can be disabled for use or their passwords changed and kept
secured for emergency purposes only. This approach helps to identify any personnel who are
working on the systems and track all important changes that are done by them. The
superuser account is intended to be used for initial configuration and servicing the system
only. For more information on user accounts see Users documentation.

3.1.2 Brief introduction to the GUI


After a successful login, the Welcome window opens and displays the system dashboard.
See Figure 3-1.

Figure 3-1 Welcome page with the dashboard

68 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

Task menu
The IBM Storage Virtualize GUI task menu is always available on the left side of the GUI
window. To browse by using this menu, click the action and choose a task that you want to
display.

Notifications icons and help


Three notification icons are in the upper navigation area of the GUI.
򐂰 The left icon indicates warning and error alerts that were recorded in the Event log.
򐂰 The middle icon shows running jobs and suggested tasks.
򐂰 The third rightmost icon offers a help menu with content that is associated with the current
tasks and the currently opened GUI menu.

Performance
This section provides important information about latency, bandwidth, input/output operations
per second (IOPS), and CPU utilization. All this information can be viewed at the system or
canister levels. A “Node comparison” view shows the differences in characteristics of each
node. The performance graph is updated with new data every 5 seconds. The granularity of
the metrics can be adjusted from seconds to days. For more detailed performance charts
select Monitoring → Performance.

Capacity
This section shows the current utilization of attached storage. It also shows provisioned
capacity and capacity savings.

System Health
This section indicates the status of all critical system components, which are grouped in three
categories: Hardware, logical, and connectivity components. When you click Expand, each
component is listed as a subgroup. You can then go directly to the section of GUI where the
component that you are interested in is managed.

3.2 Network configuration


The network configuration panel is accessed by selecting Settings → Network. Here you will
be able to configure or change configuration for the management IP, service IP, host
attachment IPs, back-end storage IPs, replication IPs, priority flow control, iSCSI, DNS,
internal proxy server, and portsets.

It also allows you to view node to node Ethernet connectivity, fibre channel connectivity,
NVMe connectivity, and fibre channel ports.

3.2.1 Management IP addresses


During the system initialization one management IP address is set. A second management IP
can be added if desired. Starting in 8.7.0 the management IPs no longer are required to use
ports 1 and 2. The two management IPs can be configured on any ports. The management
IPs are included in a default management portset. The management IPs are not tied to a
single node. Whichever node is the config node will respond to requests to the management
IP.

Chapter 3. Step-by-step configuration 69


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

Note: The system will always use the management IP on the lowest numbered port for
outbound communication, for example, Cloud Call Home, e-mail notifications, DNS lookup.

For more information on configuring ports in a FlashSystem storage unit, refer to the IBM
Redpaper The Definitive Guide to FlashSystem 5300 Port Configuration, REDP-5734.

3.2.2 Service IP addresses


On each node, port id 1 is assigned a default service IP. The first node in an enclosure is
assigned 192.168.70.121 and the second node in an enclosure is assigned 192.168.70.122.
These IPs should be changed to addresses that are accessible on the network.

By connecting to a service IP address with a browser or SSH client, you can access the
Service Assistant Interface, which can be used for maintenance and service tasks. The
service IPs are also used for some system functions, for example to access a key server,
access IP quorum, or for remote support assistance.

3.2.3 Additional Ethernet ports


The Ethernet ports menu is for configuring ports to be used for host attachment, replication,
and virtualizing back-end storage via iSCSI.

To configure an IP address on a port, select a port and go to Actions → Manage IP


addresses. See Figure 3-2.

Figure 3-2 Ethernet ports

On the next screen select Add IP address to configure the IP address and add to a portset.
See Figure 3-3.

70 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

Figure 3-3 Add IP address

3.2.4 Portsets
Portsets are groupings of logical addresses that are associated with the specific traffic types.
The system comes with one Fibre Channel and five Ethernet portsets defined. They are used
for host attachment, system management, remote copy, and back-end storage virtualization.
For further details on portsets see Portsets documentation.

Figure 3-4 shows how to create or modify portsets.

Figure 3-4 Portsets

3.3 Pools and managed disks configuration


This section describes how the storage system manages physical storage resources. All
storage resources under system control are managed by using storage pools.

Storage pools aggregate internal and external capacity as managed disks (Mdisks) and
provide containers in which you can create volumes that can be mapped to host systems.
Storage pools make it easier to dynamically allocate resources, maximize productivity, and
reduce costs.

A storage pool is created as an empty container with no storage assigned to it. Storage is
then added in the form of MDisks. MDisks can be redundant array of independent disks

Chapter 3. Step-by-step configuration 71


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

(RAID) arrays that are created by using internal storage, such as drives and flash modules, or
logical units (LUs) that are provided by external storage systems. A single storage pool can
contain both types of MDisks, but a single MDisk can be part of only one storage pool.
MDisks themselves are not visible to host systems.

Arrays are assigned to storage pools at creation time. Arrays cannot exist outside of a storage
pool and they cannot be moved between storage pools. It is possible to delete an array by
removing it from a pool and re-create it within a new pool.

External MDisks can exist within or outside of a pool. The MDisk object remains on a system
if it is visible from external storage, but its access mode changes depending on whether it is
assigned to a pool.

3.3.1 Provisioning policies


Consider utilizing a provisioning policy for pools. If a provisioning policy is assigned, any
volumes created from the pool are provisioned based on the capacity savings method defined
in the policy. A policy can be created and assigned or unassigned to a pool at any time, but it
effects only volumes created while the policy was active To create a provisioning policy select
Policies → Provisioning policies → Create policy

For more information see Provisioning policy documentation.

Note: Provisioning policy will not change any parameters of volumes that already exist in
the pool when a policy is assigned. If you already have volumes in the pool, after assigning
a provisioning policy you might need to change volumes capacity savings settings
manually.

3.3.2 Types of pools


The system supports standard pools and data reduction pools (DRPs). Both support parent
pools and child pools.

Child pools are created from capacity that is assigned to a parent pool instead of created
directly from MDisks. When a child pool is created from a standard pool, the capacity for a
child pool is reserved from the parent pool. This capacity is no longer reported as available
capacity of the parent pool. In terms of volume creation and management, child pools are
similar to parent pools. Child pools created from DRPs are quota-less. Their capacity is not
reserved but is shared with a parent pool.

DRPs use a set of techniques, such as compression and deduplication, that can reduce the
amount of usable capacity that is required to store data. Data reduction can increase storage
efficiency and performance, and reduce storage costs, especially for flash storage. These
techniques can be used in addition to compression on Flash Core Modules (FCMs).

In standard pools, there can be no compression on a pool layer, but data is still compressed
on the FCM layer if the pool contains drives with this technology.

For more information see Pools documentation.

3.3.3 Ransomware threat detection


Ransomware threat detection is automatically enabled provided the following requirements
are met.

72 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

򐂰 The pool must be created at Storage Virtualize code level 8.6.2 or higher.
򐂰 The pool consists of only FCM4 drives with firmware 4.1 or higher configured in a single
DRAID6 array.
򐂰 Each node contains at least 128GB RAM.
򐂰 Volumes are in a standard pool or fully allocated within a DRP.

3.3.4 Creating storage pools


If you want to create an encrypted pool, the encryption license must be installed and
encryption enabled before creating the pool. A pool cannot be changed to encrypted after
creation. For more information see Encryption documentation.

To create a storage pool, complete the following steps:


1. Select Pools → MDisks by Pools and click Create Pool or select Pools → Pools and
click Create → Create Pool.

Figure 3-5 shows the Create Pool menu.

Figure 3-5 Create Pool

Both alternatives open the dialog box that is shown in Figure 3-6.

Figure 3-6 Create Pool panel

Chapter 3. Step-by-step configuration 73


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

2. Select the Data reduction option if you wish to create a DRP. Leaving it clear creates a
standard storage pool.

Note: Limitations, capacity requirements, and performance characteristics of DRPs are


different from standard pools. Verify with your system architect or IBM representative
that your system was sized to be used with DRP and with its reduction features before
creating a DRP.

The size of the extents is selected at creation time and cannot be changed later. The
extent size controls the maximum total storage capacity that is manageable per system
(across all pools). For DRPs, the extent size also controls the maximum pool stored
capacity per IO group.
Refer to the Configuration Limits for IBM FlashSystem and SAN Volume Controller

Note: Do not create DRPs with small extent sizes. For more information, see this IBM
Support alert.

If an encryption license is installed and enabled, you can select whether the storage pool is
encrypted. The encryption setting of a storage pool is selected at creation time and cannot be
changed later. By default, if encryption is licensed and enabled, the encryption check-box is
selected.

Enter the name for the pool and click Create.

Naming rules: When you choose a name for a pool, the following rules apply:
򐂰 Names must begin with a letter.
򐂰 The first character cannot be numerical.
򐂰 The name can be a maximum of 63 characters.
򐂰 Valid characters are uppercase letters (A - Z), lowercase letters (a - z), digits (0 - 9),
underscore (_), period (.), hyphen (-), and space.
򐂰 Names must not begin or end with a space.
򐂰 Object names must be unique within the object type. For example, you can have a
volume that is named ABC and a storage pool that is called ABC, but not two storage
pools that are both called ABC.
򐂰 The default object name is valid (object prefix with an integer).
򐂰 Objects can be renamed at a later stage.

The new pool is created and is included in the list of storage pools. It has no storage in it, so
its capacity is zero. Storage in a form of disk arrays or externally-virtualized MDisks must be
assigned to the pool before volumes can be created.

3.3.5 Creating RAID array managed disks in a storage pool


To create a RAID array and assign it to a pool, select Pools → Pools → Select an already
created pool → Actions → Add Storage The Add Storage menu is shown in Figure 3-7 on
page 75.

74 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

Figure 3-7 Add Storage

This will open a new panel with a suggested RAID array configuration based on the installed
drives. See Figure 3-8 on page 75

Figure 3-8 RAID array

3.3.6 Adding external managed disks into a storage pool


Controllers are external storage systems that provide storage resources that are used as
MDisks. The system supports external storage controllers that are attached through internet
Small Computer Systems Interface (iSCSI) and through Fibre Channel (FC).

A key feature of the system is its ability to consolidate disk controllers from various vendors
into storage pools. The storage administrator can manage and provision storage to
applications from a single user interface and use a common set of advanced functions across
all of the storage systems under the control of the system.

This concept is called External Virtualization, which makes your storage environment more
flexible, cost-effective and easy to manage.

System layers
The system layer affects how the system interacts with a system and other external systems
that run IBM Storage Virtualize software. To virtualize another system using Storage

Chapter 3. Step-by-step configuration 75


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

Virtualize software, one system must be in the replication layer and one system must be in the
storage layer. For more information see System layers documentation.

External storage systems


IBM Storage Virtualize based systems support a wide range of storage controllers. They can
be attached vis Fibre channel or ISCSI. To check the compatibility of a system use the IBM
System Storage Interoperation Center (SSIC).

For detailed instructions on configuring an external storage system review the External
storage documentation.

Once external LUs are discovered by the IBM Storage Virtualize system they will be visible in
Pools → MDisks by pools under Unassigned MDisks. Select the MDisks that are to be added
to a pool and select Actions → Assign.

When you add MDisks to pools, you must assign them to the correct storage tiers. It is
important to set the tiers correctly if you plan to use the Easy Tier® feature. The use of an
incorrect tier can mean that the Easy Tier algorithm might make wrong decisions and thus
affect system performance.

The storage tier setting can also be changed after the MDisk is assigned to the pool. For more
information see Easy tier documentation.

3.3.7 Child pools


A child pool is a storage pool that is created within another storage pool. The storage pool in
which the child storage pool is created is called the parent storage pool. Unlike a parent pool,
a child pool does not contain MDisks. Its capacity is provided by the parent pool.

A child pool cannot be created within another child pool. Multiple child pools can be created
within a single parent pool.

Multiple child pools can be created from a single parent pool for different uses. Each child
pool can use a different provisioning policy. Child pools can also be linked to a remote pool for
policy-based replication. See Figure 3-9.

Figure 3-9 Child pools with different purposes

Child pools created from standard pools and from data reduction pools have a significant
difference:
򐂰 A child pool with a standard pool as a parent has a type child_thick. Child pools of
Standard pools have a fixed capacity, which is taken, or reserved, from the parent pool.
Free capacity of a parent pool reduces when a child pool is created. Volumes in a child
pool of a standard pool cannot occupy more capacity that is assigned to the child.

76 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

򐂰 A child pool with DRP as a parent, has type child_quotaless. Quotaless child pools share
its free and used capacity with the parent pool and do not have their own capacity limit.
Free capacity of a DRP does not change when a new quotaless child pool is created.

The capacity of a child_thick type pool is set at creation time, but can be modified later
non-disruptively. The capacity must be a multiple of the parent pool extent size and must be
smaller than the free capacity of the parent pool.

Child pools of a child_thick type can be used to implement the following configurations:
򐂰 Limit the capacity that is allocated to a specific set of volumes
It can also be useful when strict control over thin-provisioned volume expansion is needed.
For example, you might create a child pool with no volumes in it to act as an emergency
set of extents so that if the parent pool uses all its free extents, you can use the ones from
the child pool.
򐂰 As a container for VMware vSphere virtual volumes (VVOLs). Data reduction pools are not
supported as parent pools for VVOL storage.
򐂰 Migrate volumes from non-encrypted parent storage pool to encrypted child pools. When
you create a child pool of type child_thick after encryption is enabled, an encryption key is
created for the child pool, even when the parent pool is not encrypted. You can then use
volume mirroring to migrate the volumes from the non-encrypted parent pool to the
encrypted child pool.
Encrypted child_quotaless type child pools can be created only if the parent pool is
encrypted. The data reduction child pool inherits an encryption key from the parent pool.

Creating a child storage pool


To create a child pool, complete the following steps:

Select Pools → Pools. Right-click the parent pool that you want to create a child pool from
and select Create Child Pool. The Create Child Pool panel opens. See Figure 3-10.

Figure 3-10 Create Child Pool

Chapter 3. Step-by-step configuration 77


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

3.4 Configuring volumes


In IBM Storage Virtualize, a volume is storage space that is provisioned out of a storage pool
and presented to a host as a logical unit (LU), which is also called a logical disk.

This section describes how to create and provision volumes on IBM Storage Virtualize
systems. For more information on volumes and the various volume types see Volume
documentation.

3.4.1 Creating volume groups


Volume groups refer to a collection or grouping of volumes that share common characteristics
or are organized together for specific reasons. A volume group is a group of volumes. Volume
groups can be formed based on various factors:
򐂰 Volumes with similar SLA requirements, such as performance targets, availability, or data
protection policies can be grouped together within a volume group. This ensures that the
volumes within the group are managed and treated according to the same service level
guidelines.
򐂰 Some applications or data sets may require mutual consistency among multiple volumes.
In such cases, these volumes can be grouped within a volume group to ensure that they
are synchronized and maintain consistency in terms of data updates or access.
򐂰 Volumes residing on the same server can be grouped together within a volume group. This
grouping facilitates efficient management and administration of the volumes within the
server environment, which allows for streamlined operations and centralized control.

It is important to note that volume groups are distinct from consistency groups, although in
some cases, the underlying system may use a consistency group concept internally when
managing volume groups

Volume groups are used with the following functions:


򐂰 Safeguarded Copy function
򐂰 Policy-based replication
򐂰 Snapshot function

To create a volume group select Volumes → Volume groups → Create Volume Group.

Note: If you plan on using policy-based replication, configuring it and assigning a


replication policy to the volume group before creating volumes within the group will allow
you to skip the initial copy of replicated data to the remote site.

3.4.2 Creating volumes


If a volume group was created, select the volume group and select Actions → Create New
Volumes. If you are creating volumes outside a volume group, select Volumes →
Volumes → Create Volumes.

On the next panel, you will define the volume properties. However, if the pool has a
pre-assigned provisioning policy, the capacity savings option will be locked and reflect the
policy's settings.

There is a toggle on the screen for Advanced settings mode which allow manual selection of
I/O group and preferred node parameters.

78 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

The newly created volumes will automatically start formatting. This is a background process
and the volume is immediately available for host access. The default format speed is 2 MiB/s
per volume. To increase the format rate for a volume (if desired), right-click the volume and
select Modify Mirror Sync Rate. Then, choose the desired rate.

Use caution to not over utilize the systems resources by formatting too many volumes at too
high rate. If experiencing a system performance problem after increasing the mirror sync rate,
it can be decreased in the same manner.

Note: If you are using a system with IBM FlashCore® Modules the data written to the
system will be compressed automatically. There is no requirement to also create the
volumes as compressed.

3.4.3 Virtual volumes


The system provides native support for VMware vSphere APIs for Storage Awareness
(VASA) through a VASA Provider (also known as a Storage Provider), which sends and
receives information about storage that is used by VMware vSphere to the vCenter Server.
Through VASA, the system also supports VMware Virtual Volumes (also known as vVols),
which allows VMware vCenter to automate the creation, deletion and mapping of volumes.

For more information about configuring vVols with IBM Storage Virtualize, see IBM Storage
Virtualize and VMware: Integrations, Implementation and Best Practices, SG24-8549.

Also see VMWare virtual volumes documentation.

3.5 Configuring hosts


A host system can be defined as any networked computer server, virtual or physical, that
provides workloads and services to the storage.

This section describes the processes that are required to attach a supported host system to
IBM Storage Virtualize storage system through various supported interconnect protocols.

For more information see Hosts documentation.

3.5.1 Host attachment overview


IBM Storage Virtualize family supports various open system host types (from IBM and
non-IBM vendors).

These hosts can connect to the storage systems through any of the following protocols:
򐂰 Fibre Channel Protocol (FCP)
򐂰 Fibre Channel over Ethernet (FCoE)
򐂰 iSCSI
򐂰 SAS
򐂰 iSCSI Extensions for Remote Direct Memory Access (RDMA) (iSER)
򐂰 Non-Volatile Memory Express (NVMe) over Fibre Channel (FC-NVMe)
򐂰 NVMe over Remote Direct Memory Access (NVMe over RDMA)
򐂰 NVMe over Transmission Control Protocol (NVMe over TCP)

Chapter 3. Step-by-step configuration 79


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

Note: Specific host operating systems can be connected directly to the IBM Storage
Virtualize storage system without the use of SAN switches. For more information, see the
IBM System Storage Interoperation Center (SSIC).

To enable multiple access paths and enable correct volume presentation, a host system must
have a multipathing driver installed.

In addition, the multipathing driver serves the following purposes:


򐂰 Protection from:
– Fabric path failures, including port failures on IBM Storage Virtualize system nodes.
– A host bus adapter (HBA) failure (if two HBAs are used).
– Failures if the host is connected through two HBAs across two separate fabrics.
򐂰 Load balancing across the host HBAs.

For more information about the native operating system multipath drivers supported for IBM
Storage Virtualize systems, see the SSIC.

For more information about how to attach specific supported host operating systems to the
storage systems, see Host attachment documentation.

Note: If a specific host operating system is not mentioned in the SSIC, contact your IBM
representative or IBM Business Partner to submit a special request for support.

3.5.2 Fibre channel host connectivity


In this section we discuss Fibre channel host connectivity.

N_Port ID Virtualization
IBM Storage Virtualize systems utilize N_Port ID Virtualization (NPIV) which is a method for
virtualizing a physical FC port that is used for host I/O.

NPIV mode creates a virtual worldwide port name (WWPN) for every physical system FC
port. This WWPN is available for host connection only. During node maintenance, restart, or
failure, the virtual WWPN from that node is transferred to the same port of the other node in
the I/O group.

Ensure that the FC switches give each physically connected system port the ability to create
four more NPIV ports.

When performing zoning configuration, virtual WWPNs are used for host communication only;
that is, “system to host” zones must include virtual WWPNs. Internode, intersystem, and
back-end storage zones must use the WWPNs of physical ports. Ensure that equivalent ports
(with the same port ID) are on the same fabric and in the same zone.

Important: IBM i Systems attached to FlashSystem or SVC must be converted to use


FlashSystem NPIV before upgrading to 8.7 or higher. For FlashSystem or SVC systems
which have NPIV in "disabled" or "transitional" state, and have any IBM i hosts, a modified
procedure must be used when enabling NPIV, to avoid loss of host access to data.See this
link for details.

80 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

To view the virtual WWPNs to be used in system to host select Settings → Network → Fibre
Channel Ports. Expand the twisty for each port. Columns will indicate WWPN, Host IO
Permitted, and Protocol type. SCSI is for Fibre Channel Protocol (FCP).

Note: The NPIV WWPNs do not become active until there is at least one online volume.

Host zones
A host must be zoned to an I/O group to access volumes that are presented by this I/O group.

The preferred zoning policy is single initiator zoning. To implement it, create a separate zone
for each host bus adapter (HBA) port, and place one port from each node in each I/O group
that the host accesses in this zone. A typical fibre channel host will have two ports zoned to
each IO group creating a total of four paths. For deployments with more than 64 hosts that are
defined in the system, this host zoning scheme must be used.

Note: Cisco Smart Zoning and Brocade Peer Zoning are supported, with which you can
insert target ports and multiple initiator ports in a single zone for easy of management but
act the same as though each initiator and target are configured in isolated zones. The use
of these zoning techniques is supported for host attachment and storage virtualization. As
a best practice, use normal zones when configuring ports for clustering or for replication
because these functions require the port to be an initiator and a target.

Consider the following rules for zoning hosts over SCSI or FC-NVMe:
򐂰 For any volume, the number of paths through the SAN from the host to a system must not
exceed eight. For most configurations, four paths to an I/O group are sufficient.
򐂰 Balance the host load across the system’s ports. For example, zone the first host with
ports 1 and 3 of each node in the I/O group, zone the second host with ports 2 and 4, and
so on. To obtain the best overall performance of the system, the load of each port must be
equal. Assuming that a similar load is generated by each host, you can achieve this
balance by zoning approximately the same number of host ports to each port.
򐂰 Spread the load across all system ports. Use all ports that are available on your machine.
򐂰 Balance the host load across HBA ports. If the host has more than one HBA port per
fabric, zone each host port with a separate group of system ports.

All paths must be managed by the multipath driver on the host side. Make sure that the
multipath driver on each server can handle the number of paths that is required to access all
volumes that are mapped to the host.

3.5.3 Ethernet host connectivity


You can attach your IBM Storage Virtualize system to iSCSI, iSER, NVMe over RDMA, and
NVMe over TCP hosts by using the Ethernet ports of the system.

The same ports can be used for iSCSI and iSER host attachment concurrently; however, a
single host can establish an iSCSI or session, but not both

Hosts connect to the system through IP addresses, which are assigned to the Ethernet ports
of the node. If the node fails, the address becomes unavailable and the host loses
communication with the system through that node.

To allow hosts to maintain access to data, the node-port IP addresses for the failed node are
transferred to the partner node in the I/O group. The partner node handles requests for its

Chapter 3. Step-by-step configuration 81


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

own node-port IP addresses and for node-port IP addresses on the failed node. This process
is known as node-port IP failover.

In addition to node-port IP addresses, the iSCSI name and iSCSI alias for the failed node are
transferred to the partner node. After the failed node recovers, the node-port IP address and
the iSCSI name and alias are returned to the original node.

iSCSI
iSCSI is a protocol that uses the Transmission Control Protocol and Internet Protocol
(TCP/IP) to encapsulate and send SCSI commands to storage devices that are connected to
a network. iSCSI is used to deliver SCSI commands from a client interface, which is called an
iSCSI Initiator, to the server interface, which is known as the iSCSI Target. The iSCSI payload
contains the SCSI CDB and, optionally, data. The target carries out the SCSI commands and
sends the response back to the initiator.

NVMe over Remote Direct Memory Access


IBM Storage Virtualize can be attached to an NVMe host through NVMe over Remote Direct
Memory Access (RDMA). NVMe over RDMA uses RDMA over Converged Ethernet (RoCE)
v2 as the transport protocol. RoCE v2 is based on user datagram protocol (UDP).

RDMA is a host-offload, host-bypass technology that allows an application (including storage)


to make data transfers directly to and from another application’s memory space. The
RDMA-capable Ethernet NICs (RNICs), and not the host, manage reliable data transfers
between source and destination.

RNICs can use RDMA over Ethernet by way of RoCE encapsulation. RoCE wraps standard
InfiniBand payloads with Ethernet or IP over Ethernet frames, and is sometimes called
InfiniBand over Ethernet. The following main RoCE encapsulation types are available:
򐂰 RoCE V1
This type uses dedicated Ethernet Protocol Encapsulation (Ethernet packets between
source and destination MAC addresses by using EtherType 0x8915).
򐂰 RoCE V2:
– This type uses dedicated UDP over Ethernet Protocol Encapsulation (IP UDP packets
by using port 4791 between source and destination IPs; UDP packets are sent over
Ethernet by using source and destination MAC addresses)
– This type is not compatible with other Ethernet options, such as RoCE v1.

Note: Unlike RoCE V1, RoCE V2 is routable.

NVMe over TCP


IBM Storage Virtualize can be attached to an NVMe host through NVMe over Transmission
Control Protocol (TCP). NVMe over TCP is a transport that allows NVMe performance without
any constraint to the data center infrastructure.

NVMe over TCP needs more CPU resources than protocols using RDMA. Each NVMe/TCP
port on FlashSystem supports multiple IPs and multiple VLANs. Generally, NVMe-TCP runs
on all switches and is routable.

For operating system support and multipathing, see IBM System Storage Interoperation
Center (SSIC).

82 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

3.5.4 Host objects


Before a host can access the storage capacity, it must first be presented to the storage
system as a host object.

A host object is configured by using the GUI or command line interface (CLI) and must
contain the necessary credentials for host-to-storage communications. After this process is
completed, storage capacity can be mapped to that host in the form of a volume.

IBM Storage Virtualize supports configuring the following host objects:


򐂰 Host
򐂰 Host cluster

A host cluster object groups clustered servers and treats them as a single entity. This
configuration allows multiple hosts to access the same volumes through one shared mapping.

Note: Any volume that is mapped to a host cluster is automatically assigned to all of the
members in that cluster with the same SCSI ID.

A typical use case for a host cluster object is to group multiple clustered servers with a
common operating system, such as IBM PowerHA® and Microsoft Cluster Server, and enable
them to have shared access to common volumes.

To create a host object select Hosts → Hosts → Add Host. The Add Host panel will display.
See Figure 3-11.

Chapter 3. Step-by-step configuration 83


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 3-11 Add Host

Tip: The Host port drop down will show FCP initiator WWPNs currently logged into the
system. If an expected WWPN is missing, check switch zoning and rescan the storage
from the hosts. Some operating systems may log out if no LUNs are mapped to the host. If
that is the case select Enter Unverified WWPN and enter the host WWPNs manually.

3.5.5 Mapping volumes for host access


In order for the host or host cluster to access the volumes they must be mapped to the host or
host cluster. To perform the mappings, select the volumes you wish to map → Actions →
Map to Host or Host Cluster → Select the host or host cluster → Next. In most cases
leave the radio button for the system to assign SCSI LU IDs.

Note: In most cases all volumes within a volume group would be mapped to the same host
or host cluster and the mapping can easily be done within the volume group view.

3.6 Snapshots and replication


In this section we discuss configuration of snapshots and replication.

84 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

3.6.1 Volume group snapshots


Before the introduction of volume group snapshots, when volumes were dependent on each
other, creating point-in-time copies required taking snapshots at the exact same time. This
was achieved through the use of consistency groups (CG), which consisted of a group of
mappings that had to be started simultaneously. Configuring the mappings and targets for
Flashcopies within a consistency group was a complex process that had to be repeated each
time a new point-in-time copy was created. This complexity posed limitations on usage and
made it challenging to incorporate new functions.

The purpose of the volume group snapshot management model is to simplify the
implementation of standard FlashCopy® operations. It achieves this by offering a more
straightforward setup process and separating the snapshot and clone features. With volume
group snapshots, administrators can create snapshots of volume groups with more ease and
efficiency, without the need for complex consistency group configurations.

Snapshots cannot be mapped to a host. In order to access the data on a snapshot, create a
thin-clone of the snapshot and map it to a host.

Demonstration videos: The following demonstration videos are available:


򐂰 IBM Storage Virtualize V8.6: Handling Snapshots using the GUI.
򐂰 IBM Storage Virtualize V8.6: Handling snapshots using the command line interface.

For more information see Snapshots documentation.

Triggering volume group snapshots


The process for triggering a volume group snapshot involves a streamlined version of the
FlashCopy mapping trigger process. They can be triggered via the GUI, on a schedule using
a snapshot policy, or by an external application such as Copy Services Manager (CSM).

Volume group snapshot policy


A volume group snapshot policy is designed to automate the creation and deletion of
snapshots based on predefined schedules, which eliminates the need for external
applications.
򐂰 Users can choose from predefined snapshot policies or create custom policies tailored to
their specific needs.
򐂰 The snapshot policies configured with the volume group snapshot scheduler are reusable
and can be applied to multiple volume groups as needed.
򐂰 The volume group snapshot scheduler comes with default snapshot policy parameters,
offering convenient options for most use cases.
򐂰 Users can specify the creation frequency of snapshots in minutes, hours, weeks, days, or
months. The minimum creation frequency allowed is 60 minutes. Additionally, the retention
of snapshots can be specified in terms of days.
򐂰 When a snapshot policy is assigned to a volume group there is an option to select
Safeguarded. Safeguarded snapshots can only be deleted before their expiration time by a
security administrator.

Note: When using a snapshot policy, after the initial snapshot, snapshots are triggered
based on the frequency defined. This means that the time of day the snapshot is triggered
may shift forward and backward with daylight savings time changes.

Chapter 3. Step-by-step configuration 85


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

To create, view, or assign a snapshot policy select Policies → Snapshot policies. See
Figure 3-12.

Figure 3-12 Snapshot Policies

You can also suspend or unassign a policy from within the volume group. See Figure 3-13.

Figure 3-13 Suspend Policy

3.6.2 Asynchronous policy-based replication


Asynchronous policy-based replication provides a variable, greater-than-zero recovery point
that aims to achieve the best possible recovery point for the current conditions. This type of
replication ensures mutual consistency between all volumes in the volume group.

To quickly configure policy based replication between two systems, both will require at least
one IP address created and assigned to a replication portset along with at least one pool with
storage created. Multiple IPs can be added to a replication portset. If there is a second
independent inter-site link between the systems a second portset can be used and added to
the partnership.

86 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

1. On the primary system select Copy Services → Partnerships → Create Partnership. If


using IP select IP and enter the partner IP address the select Test Connection. If the
partner meets requirements for policy based replication the Use policy-based replication
selection can be checked. Fill out the rest of the panel and select Create. Repeat these
steps on the partner system. See Figure 3-14 on page 87.

Figure 3-14 Create Partnership

2. Once the partnership shows a green dot and configured select Setup policy-based
replication. See Figure 3-15 on page 88.

Chapter 3. Step-by-step configuration 87


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

Figure 3-15 Partnership Created

3. Select Setup policy-based replication to be guided through the rest of the configuration.
See Figure 3-16.

Figure 3-16 Setup policy-based replication wizard

For more information see IBM Documentation Asynchronous disaster recovery replication.

Also see Policy-based replication Redpaper.

88 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:44 pm 8561ch03.fm

Synchronous replication is provided by policy-based high availability which is beyond the


scope of this document. For more information on high availability see High Availability.

Chapter 3. Step-by-step configuration 89


8561ch03.fm Draft Document for Review July 19, 2024 5:44 pm

90 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:41 pm 8561ch 04.fm

Chapter 4. Verifying configuration and basic


operations
This chapter provides steps to verify the configuration along with tips to resolve common
implementation problems. It also provides additional information about functions that should
be considered when implementing a new system.

In this chapter we discuss the system health dashboard, verifying configuration of objects
configured in Chapter 3, “Step-by-step configuration” on page 67, system security, getting
support from IBM, and data migration.

This chapter has the following sections:


򐂰 “Verifying the configuration” on page 92
򐂰 “Additional settings and basic operations” on page 95

© Copyright IBM Corp. 2024. All rights reserved. 91


8561ch 04.fm Draft Document for Review July 19, 2024 5:41 pm

4.1 Verifying the configuration


In this section we discuss some commons tasks to verify the configuration.

4.1.1 System Health Dashboard


The system dashboard provides a way to quickly assess the overall condition of the system
and view notifications of any critical issues that require immediate action. The bottom third of
the dashboard provides system health details including tiles for hardware components, logical
components, and connectivity components. Each tile has a link to quickly access relevant
information. See Figure 4-1.

Figure 4-1 System Health

4.1.2 Verifying network configuration


The network IPs and WWPNs can be verified by selecting Settings → Network → Portsets.
Selecting any of the portsets will allow viewing details of the ports and IP addresses assigned
to that portset. See Figure 4-2.

Figure 4-2 Portset Mappings

Selecting IP Addresses or Partnerships gives further details. See Figure 4-3 on page 93

92 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:41 pm 8561ch 04.fm

Figure 4-3 Portset IP addresses

4.1.3 Verifying storage configuration


Select Pools → MDisks by pools to view the storage configuration. All pools and Mdisks are
expected to be in an online state. If any are degraded, check Monitoring → Events for any
related events.

There is a column for Usable Capacity and a column for Written Capacity Limit. If utilizing
over provisioned storage (for example, FCMs) the values will be different. The usable capacity
represents the physical capacity available after the data is reduced via compression and
deduplication. The Written Capacity Limit is the effective capacity of data written to the
system before its size is reduced.

4.1.4 Verifying volume configuration


Select Volumes → Volumes to view the volume configuration. All volumes are expected to be
online. If any are degraded, check Monitoring → Events for any related events.

The columns displayed can be modified by right clicking on the column titles bar. There are
some useful capacity related columns that can be displayed.

If a volume is thin provisioned in a standard pool adding the columns Real Capacity and Used
Capacity will provide useful information. Used capacity is the capacity used by the data
written to the volume. Real Capacity is the Used Capacity plus a contingency capacity that is
used for new writes. These values are effective capacity.

Adding the column Compression Savings will display information on how compressible the
data is.

Capacity Savings report


IBM Comprestimator is a utility that estimates the capacity savings that can be achieved
when compression is used for storage volumes. The utility is integrated into the system and
results can be viewed by using the GUI and the CLI. The integrated Comprestimator is always
enabled and running continuously, thus providing up-to-date compression estimation over the
entire cluster, both in GUI and IBM Storage Insights.

Chapter 4. Verifying configuration and basic operations 93


8561ch 04.fm Draft Document for Review July 19, 2024 5:41 pm

IBM Comprestimator provides a quick and accurate estimation of compression and


thin-provisioning benefits. The utility performs read-only operations, so it does not affect the
data that is stored on the volume.

To view the results and the date of the latest estimation cycle, under the volumes view, Right
Click the volume → Capacity Savings → Estimate Compression Savings.

To download a capacity savings reports, under the volumes view, select Actions → Capacity
Savings → Download Savings report.

The report is also useful for determining the physical capacity used by each volume when the
volume is compressed by FCMs or when the volumes are compressed in a Data Reduction
Pool.

A stand-alone comprestimator utility can be installed and used on host systems to estimate
savings before moving data to a Storage Virtualize system. To download the Comprestimator
that can be installed on a server follow this link.

4.1.5 Verifying host configuration


Select Hosts → Hosts to view the hosts status. The status of all hosts is expected to be
online. A host with a degraded status is typically caused by the host being partially connected
to the storage. e.g. both host WWPNs are logged into node1 while one host WWPN is logged
into node2.

To review the host connectivity select Settings → Network → Fibre Channel or NVMe
Connectivity. The results can be filtered by the host.

In the following example Host1 is degraded because each WWPN is logged into node1 twice
and node 2 once. See Figure 4-4.

Figure 4-4 Host with asymmetrical logins

In cases where a host shows as degraded but there is no hardware failure on the host or
storage, check the fibre channel switch zoning and rescan the storage from the host.

Tip: If using Broadcom fibre channel switches the fcping command run via the switch CLI
can be quickly used to verify zoning and WWPN connectivity.

94 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:41 pm 8561ch 04.fm

4.2 Additional settings and basic operations


The following sections discuss additional settings that an administrator should be aware of
when implementing a new systems along with basic operations.

4.2.1 Security settings


Storage Virtualize systems implement various security related features. To configure or
modify security functions select Settings → Security. Some of the more commonly utilized
settings are described below.

For more information see Security documentation.

Also see Storage Virtualize Security Feature Checklist.

Remote authentication
Remote authentication allows users to authenticate to the system using credentials that are
stored on an external authentication service. When you configure remote authentication, you
do not need to configure users on the system or assign more passwords. Instead, you can
use your existing passwords and user groups that are defined on the remote service to
simplify user management and access to enforce password policies more efficiently, and to
separate user management from storage management.

A remote user is authenticated on a remote LDAP server. A remote user does not need to be
added to the list of users on the system, although they can be added to configure optional
SSH keys. For remote users, an equivalent user group must be created on the system with
the same name and role as the group on the remote LDAP server.

For more information see Remote Authentication Documentation.

Ownership groups
An ownership group defines a subset of users and objects within the system. You can create
ownership groups to further restrict access to specific resources that are defined in the
ownership group. Only users with Administrator or Security Administrator roles can configure
and manage ownership groups.

Ownership groups restrict access to only those objects that are defined within that ownership
group. An owned object can belong to one ownership group.

An owner is a user with an ownership group that can view and manipulate objects within that
group.

The system supports the following resources that you assign to ownership groups:
򐂰 Child pools
򐂰 Volumes
򐂰 Volume groups
򐂰 Hosts
򐂰 Host clusters
򐂰 Host mappings
򐂰 FlashCopy mappings
򐂰 FlashCopy consistency groups
򐂰 User groups
򐂰 Portsets

Chapter 4. Verifying configuration and basic operations 95


8561ch 04.fm Draft Document for Review July 19, 2024 5:41 pm

The following basic use cases can be applied to the use of ownership groups on the system:
򐂰 Objects are created within the ownership group. Other objects can be on the system that
are not in the ownership group.
򐂰 On a system where these supported objects are configured, and you want to migrate
these objects to use ownership groups.

When a user group is assigned to an ownership group, the users in that user group retain
their role, but are restricted to only those resources within the same ownership group. User
groups can define the access to operations on the system, and the ownership group can
further limit access to individual resources.

For example, you can configure a user group with the Copy Operator role, which limits access
of the user to Copy Services functions, such as FlashCopy and Remote Copy operations.
Access to individual resources, such as a specific FlashCopy consistency group, can be
further restricted by adding it to an ownership group.

When the user logs on to the management GUI, only resources that they can access through
the ownership group are displayed. Also, only events and commands that are related to the
ownership group in which a user belongs are viewable by those users.

For more information see Ownership groups documentation.

System certificates
SSL certificates are used to establish secure communications for many services. The system
uses a certificate to identify itself when authenticating with other devices. Depending on the
scenario, the system might be acting as either the client or the server.

The system has a root certificate authority (CA) that can be used to create internally signed
system certificates. System setup creates a certificate that is signed by the root CA to secure
connections between the management GUI and the browser. The root certificate can be
exported from the system and added to truststores on other systems, browsers, or devices to
establish trust. Internally signed certificates can be renewed automatically before they expire.
Automatic renewal simplifies the certificate renewal process and prevents security warnings
from expired certificates. Automatic renewal is only supported by using an internally signed
certificate.

Externally signed certificates are issued and signed by a trusted third-party provider of
certificates, called an external certificate authority (CA). This CA can be a public CA or your
own organization's CA. Most web browsers trust well-known public CAs and include the root
certificate for these CAs in the device or application. Externally signed certificates cannot be
renewed automatically because they must be issued by the external CA. Externally signed
certificates must be manually updated before they expire by creating a new certificate signing
request (CSR) on the system and supplying it to the CA. The CA signs the request and issues
a certificate that must be installed on the system. The system raises a warning in the event
log 30 days before the certificate expires.

Note: An externally signed certificate must meet the following requirements:


򐂰 X.509v3 Key Usage Extensions: Include Digital Signature.
򐂰 X.509v3 Extended Key Usage Extensions: Include TLS Web Server Authentication
and TLS Web Client Authentication. Additionally, Any Extended Key Usage (anyEKU)
can be included.

Ensure that the Certificate Authority (CA) used to sign the certificate includes these
extensions.

96 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:41 pm 8561ch 04.fm

Security protocol levels


Security administrators can change the security protocol level for either SSL or SSH
protocols. When you change the security level for either of these security protocols, you can
control which encryption algorithms, ciphers, and version of the protocol are permitted on the
system.

The GUI gives a high level description of each level. For a more detailed description including
the ciphers supported with each level see Security protocol levels documentation.

4.2.2 Audit log


The audit log is useful when analyzing past configuration events, especially when trying to
determine, for example, how a volume ended up being shared by two hosts, or why the
volume was overwritten. The audit log is also included in the svc_snap support data to aid in
problem determination.

The audit log tracks action commands that are issued through an SSH session, management
GUI, or Remote Support Assistance. It provides the following entries:
򐂰 Identity of the user who ran the action command.
򐂰 Name of the actionable command.
򐂰 Timestamp of when the actionable command ran on the configuration node.
򐂰 Parameters that ran with the actionable command.

The following items are not documented in the audit log:


򐂰 Commands that fail are not logged.
򐂰 A result code of 0 (success) or 1 (success in progress) is not logged.
򐂰 Result object ID of node type (for the addnode command) is not logged.
򐂰 View commands are not logged.

The audit log is accessed by selecting Access → Audit Log see Figure 4-5.

Figure 4-5 Audit log

4.2.3 Support settings


Support settings can be configured and modified by selecting Settings → Support.

Call Home
IBM Call Home is a support function embedded in all IBM Storage Virtualize storage
products. By enabling call home, the health and functionality of your system is constantly
monitored by IBM. Should a software or hardware error occur, the call home function notifies
IBM support of the event and then automatically opens a service request. By obtaining

Chapter 4. Verifying configuration and basic operations 97


8561ch 04.fm Draft Document for Review July 19, 2024 5:41 pm

information in this way, IBM support is quickly informed about the issue and can quickly
develop an action plan for problem resolution.

There are two methods available for a system to call home and both can be enabled
simultaneously.
򐂰 Cloud Services uses HTTPS to connect directly to IBM from the management IP assigned
to the lowest physical port id over port 443
򐂰 Email services requires an SMTP server to forward the email to IBM. Email services can
also send alerts to local administrators.

Remote Support Assistance


Remote Support Assistance enables IBM Support to remotely connect to an IBM
FlashSystem system through a secure tunnel to perform analysis, log collection, and software
updates. The tunnel can be enabled as needed by the client or as a permanent connection.

For detailed information on call home or remote support assistance see the following white
paper. IBM Storage Virtualize Products Call Home and Remote Support Overview.

Note: Remote support assistance uses the service IPs to make an outbound connection to
IBM on port 22.

The connections for both Call Home and Remote Support Assistance can be routed via a
client supplied web proxy.

Support package
If you encounter a problem and contact the IBM Support Center, you will be asked to provide
a support package (often referred to as a snap).

You can use two methods to collect and upload the support package from the GUI of your
Storage Virtualize system:
򐂰 Upload Support Package: Use this feature if your system is connected to the internet to
upload the Support Package directly from the storage system.
򐂰 Download Support Package: Use this feature if your system is not connected to the
internet to upload the Support Package manually.

The support agent will advise the type of support package to collect based on the problem.
For general guidelines and the differences between the different support package types see
What Data Should You Collect for a Problem.

98 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

Chapter 5. IBM Storage Insights and IBM


Storage Insights Pro
Managing storage systems can be complex. You need to keep an eye on performance,
capacity, and overall health to ensure your data is always accessible and secure. IBM Storage
Insights offers a solution to help you monitor, manage, and optimize your storage resources.

This chapter has the following sections:


򐂰 “IBM Storage Insights overview” on page 100
򐂰 “IBM Storage Insights monitoring” on page 101

© Copyright IBM Corp. 2024. All rights reserved. 99


8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

5.1 IBM Storage Insights overview


IBM Storage Insights is another part of the monitoring capability of the IBM FlashSystems
and IBM SAN Volume Controller systems running IBM Storage Virtualize software and
supplements the views that are available in the product GUI.

IBM strongly recommends that all customers install and use this no-charge, cloud-based IBM
application because it provides a single dashboard that provides a clear view of all your IBM
block storage. You can make better decisions by seeing trends in performance and capacity.

Note: IBM Storage Insights is available at no cost, to clients who have IBM Storage
Systems on either IBM warranty or maintenance. The more fully featured IBM Storage
Insights Pro is a chargeable product, which can purchased standalone and also be
included in certain levels of IBM Storage Expert Care and IBM Storage Control.

With storage health information, you can focus on areas that need attention. When IBM
support is needed, IBM Storage Insights simplifies uploading logs, speeds resolution with
online configuration data, and provides an overview of open tickets all in one place.

IBM Storage Insights includes the following features:


򐂰 A unified view of IBM systems:
– Provides a single view to see all your system’s characteristics.
– Displays all of your IBM storage inventory.
– Provides a live event feed so that you know in real time what is going on with your
storage so that you can act quickly.
򐂰 IBM Storage Insights collects telemetry data and Call Home data, and provides real-time
system reporting of capacity and performance.
򐂰 Overall storage monitoring, which provides the following information:
– The overall health of the system.
– Monitoring of the configuration to see whether it meets preferred practices.
– System resource management to determine which system is overtaxed and provides
proactive recommendations to fix it.
򐂰 IBM Storage Insights provides advanced customer service with an event filter that you can
use to accomplish the following tasks:
– You and IBM Support can view, open, and close support tickets, and track trends.
– You can use the autolog collection capability to collect the logs and send them to IBM
before IBM Support looks into the problem. This capability can save as much as 50%
of the time to resolve the case.

In addition to the no-charge version of IBM Storage Insights, IBM offers IBM Storage Insights
Pro, which is a subscription service that provides longer historical views of data, more
reporting and optimization options, and supports IBM file and block storage with EMC VNX
and VMAX.

Note: For a comparison of the features in the IBM Storage Insights and Insights Pro,
editions, refer to this page in the IBM Documentation here.

100 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

5.1.1 IBM Storage Insights: Information and registration


For more information about IBM Storage Insights, see the following resources:
򐂰 IBM Storage Insights Fact Sheet.
򐂰 IBM Storage Insights Security Guide, SC27-8774.
򐂰 This IBM Documentation web page.
򐂰 Product registration (used to sign up and register for this no-charge service).

5.2 IBM Storage Insights monitoring


With IBM Storage Insights, you can optimize your storage infrastructure by using this
cloud-based storage management and support platform with predictive analytics.

The monitoring capabilities that IBM Storage Insights provides are useful for things like
capacity planning, workload optimization, and managing support tickets for ongoing issues.

For a live demo of IBM Storage Insights, see Storage Insights Demo (requires login).

Demonstration videos: To view videos about Storage Insights, see What’s New in IBM
Storage Insights. The videos include new features and enhancements of IBM Storage
Insights.

After you add your systems to IBM Storage Insights, you see the Dashboard, where you can
select a system that you want to see the overview for.

There are two versions of the dashboard panel, the classic version and the new Carbon
enhanced version.

Figure 5-1 on page 101 shows the classic version view of the IBM Storage Insights
dashboard.

Figure 5-1 IBM Storage Insights System overview (classic view)

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 101
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-2 shows the newer Carbon enhanced view of the IBM Storage Insights dashboard.

Figure 5-2 IBM Storage Insights System overview (Carbon enhanced view)

For the purposes of the next few examples we will use the classic view screens.

5.2.1 Component health


Component health is shown at the upper center of the window. If there is a problem with one
of the Hardware, Logical or Connectivity components, errors are shown here, as shown in
Figure 5-3.

Figure 5-3 Component Health overview

The error entries can be expanded to obtain more details by selecting the three dots at the
upper right corner of the component that has an error and then selecting View Details. The
relevant part of the more detailed System View opens, and what you see depends on which
component has the error, as shown in Figure 5-4.

102 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

Figure 5-4 Ports in error

From here, it is now obvious which components have the problem and exactly what is wrong
with them, so now you can log a support ticket with IBM if necessary.

5.2.2 Capacity monitoring


You can see key statistics such as Usable and Provisioned Capacity and Capacity Savings as
shown in Figure 5-5. Capacity can be viewed by volume or pool and the View More button
shows a trend curve.

Figure 5-5 Capacity area of the IBM Storage Insights system overview

In the Capacity view, the user can click on the required system. Clicking any of these items
takes the user to the detailed system view for the selection option. From there, you can get a
historical view of how the system capacity changed over time, as shown in Figure 5-6. At any
time, the user can select the timescale, resources, and metrics to be displayed on the graph
by clicking any options around the graph.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 103
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-6 Capacity planning for one system

5.2.3 Performance monitoring


From the system overview, you can scroll down and see the three key performance statistics
for your system, as shown in Figure 5-7. For the Performance overview, these statistics are
aggregated across the whole system, and you cannot drill down by Pool, Volume, or other
items.

Figure 5-7 System overview: Performance

To view more detailed performance statistics, enter the system view again, as described in
5.2.2, “Capacity monitoring” on page 103.

For this performance example, we select View Pools, and then select Performance from the
System View pane, as shown in Figure 5-8.

104 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

Figure 5-8 IBM Storage Insights: Performance view

It is possible to customize what can be seen on the graph by selecting the metrics and
resources. In Figure 5-9, the Overall Response Time for one IBM FlashSystem over a
12-hour period is displayed.

Figure 5-9 Filtered performance graph

Scrolling down the graph, the Performance List view is visible, as shown in Figure 5-10.
Metrics can be selected by clicking the filter button at the right of the column headers. If you
select a row, the graph is filtered for that selection only. Multiple rows can be selected by
holding down the Shift or Ctrl keys.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 105
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-10 Performance List View

5.2.4 Logging support tickets by using IBM Storage Insights


With IBM Storage Insights, you can log existing support tickets that greatly complement the
enhanced monitoring opportunities that the software provides. When an issue is detected and
you want to engage IBM Support, complete the following steps:
1. Select the system to open the System Overview window and click Get Support, as shown
in Figure 5-11.

Figure 5-11 Get Support (see highlighted area)

A window opens where you can create a ticket or update an existing ticket, as shown in
Figure 5-12.

106 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

Figure 5-12 Get Support window

2. Select Create Ticket, and the ticket creation wizard opens. Details of the system are
automatically populated, including the customer number, as shown in Figure 5-13. Select
Next.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 107
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-13 Create Ticket wizard

108 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

3. You can enter relevant details about your problem to the ticket, as shown in Figure 5-14. It
is also possible to attach images or files to the ticket, such as PuTTY logs and screen
captures. Once done, select Next.

Figure 5-14 Add a note or attachment window

4. You can select a severity for the ticket. Examples of what severity you should select are
shown in Figure 5-15. Because in our example there are storage ports offline with no
impact, we select severity 3 because there is only minor impact if any.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 109
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-15 Selecting a Severity Level window

5. Choose whether this is a hardware or a software problem. Select the relevant option (for
this example, the offline ports are likely caused by a physical layer hardware problem).
Once done, click Next.

110 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

6. Review the details of the ticket that will be logged with IBM, as shown in Figure 5-16.
Contact details must be entered so that IBM Support can respond to the correct person.
You must also choose which type of logs should be attached to the ticket. For more
information about the types of snap, see Figure 5-16. Click Create Ticket.

Figure 5-16 Review the ticket window

7. Once done, select Create Ticket. A confirmation window opens, as shown in Figure 5-17
on page 112, and IBM Storage Insights automatically uploads the snap to the ticket when
it is collected. Click Close.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 111
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

Figure 5-17 Update ticket

5.2.5 Managing existing support tickets by using IBM Storage Insights


With IBM Storage Insights, you can track existing support tickets and upload logs to them. To
do so, complete the following steps:
1. From the System Overview window, select Tickets, as shown in Figure 5-18.

Figure 5-18 View tickets

112 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

This window shows the newly created ticket number and a history of support tickets that
were logged through IBM Storage Insights for the system. Tickets that are not currently
open are listed under Closed Tickets, and currently open tickets are listed under Open
Tickets.
2. To quickly add logs to a ticket without having to browse to the system GUI or use
IBM ECuRep, click Get Support and Add Log Package to Ticket. A window opens that
guides you through the process, as shown in Figure 5-19 on page 113. After entering the
support ticket number, you can select which type of log package you want and add a note
to the ticket with the logs.

Figure 5-19 Adding a log package to the ticket

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 113
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

3. The review screen shown in Figure 5-20 allows you confirm what is going to be uploaded.

Figure 5-20 Confirming the log upload

4. After clicking Update Ticket, a confirmation opens, as shown in Figure 5-21 on page 115.
You can exit the wizard. IBM Storage Insights runs in the background to gather the logs
and upload them to the ticket.

114 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:22 am 8561ch05.fm

Figure 5-21 Log upload completed and processing

5.2.6 Enhancements to IBM Storage Insights Pro


In this section we discuss enhancements to IBM Storage Insights Pro in Version 8.7.

Ransomware threat detection


A key part of monitoring your system includes the detection of potential ransomware attacks.
Starting with IBM Storage Virtualize 860 the ransomware threat detection mechanism was
introduce as a Virtualize level only. To ensure that you have the latest storage metadata for
detecting those types of attacks, compression and cyber resiliency statistics for volumes are
collected every 5 minutes. With these statistics, IBM Storage Insights builds a historical model
of a storage system and uses its built-in intelligence and formulas to identify when and where
ransomware attacks might be occurring.

With IBM Storage Virtualize software 870 and FCM’s with FCM firmware 4.1, the ransomware
threat detection is further improved as follows:
򐂰 IBM FlashCore modules collect and analyze detailed ransomware statistics from every I/O
with no performance impact.
򐂰 IBM Storage Virtualize runs an AI engine on every FlashSystem that is fed Machine
Language (ML) models developed by IBM Research® trained on real-world ransomware.
򐂰 The AI engine learns what’s normal for the system and detects threats using data from
FCM.

Chapter 5. IBM Storage Insights and IBM Storage Insights Pro 115
8561ch05.fm Draft Document for Review July 22, 2024 9:22 am

򐂰 IBM Storage Insights Pro collects threat information from connected FlashSystems, alerts
trigger SIEM/SOAR software to initiate a response.
򐂰 Statistics are fed back to IBM to improve ML models.

For more information on the IBM ransomware threat detection solutions including those
mentioned in this book see this link.

Also, refer to this blog post on how to mark volume snapshots those are created after
ransomware threat detection as compromised.

IBM Storage Virtualize 8.7.0 including Flash Grid


Flash Grid – scale out up to 8 FlashSystem or SVC systems to be managed as one, including
non-disruptive workload mobility between members of the grid.

IBM Storage Insights Pro works in conjunction with the new Flash Grid and today provides an
overview of your grid with grouping of your systems and the ability to non-disruptively move
workloads (storage partitions) between systems in the grid. The end goal is to provide a
seamless integration and interaction between the on-premise and cloud based management
portals. Figure 5-22 on page 116 shows the integration on IBM Storage Insights Pro with the
IBM Storage Virtualize software GUI and the linkage to the IBM FlashSystems and IBM SAN
Volume Controllers it monitors.

Figure 5-22 IBM Storage insights Pro and IBM Flash Grid integration

116 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

Chapter 6. Storage Virtualize


troubleshooting and diagnostics
This chapter provides information to troubleshoot common problems that can occur in an IBM
Storage Virtualize 8.7.0 environment. It describes situations that are related to IBM SAN
Volume Controller (SVC), IBM FlashSystems, the storage area network (SAN) environment,
optional external storage subsystems, and hosts. It also explains how to collect necessary
problem determination data.

This chapter includes the following topics:


򐂰 “Troubleshooting” on page 118
򐂰 “Collecting diagnostic data” on page 125
򐂰 “For performance-related issues, it is helpful to have corresponding monitoring. IBM
Storage Insights and IBM Spectrum Control are recommended. If required, you can export
performance data from there for the related period taken from the historical data.” on
page 137

© Copyright IBM Corp. 2024. All rights reserved. 117


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

6.1 Troubleshooting
Troubleshooting should follow a systematic approach to solve a problem. The goal of
troubleshooting or problem determination is to understand, why something does not work as
expected and create a resolution to resolve this. An important step therefore is to make a
proper problem description, which should be as accurately as possible. Then you need to
collect the support data from all involved components of the environment for analysis. This
might include a snap from the IBM Storage Virtualize system, logs from SAN or network
switches and host OS logs.

An effective problem report ideally describes these items:


򐂰 the expected behavior
򐂰 the actual behavior
򐂰 if possible, how to reproduce the behavior
򐂰 a precise timeline

The following questions help define the problem for effective troubleshooting:
򐂰 What are the symptoms of the problem?
– What is reporting the problem?
– Which error codes and messages were observed?
– What is the business impact of the problem?
– Where does the problem occur?
– Which exact component is affected, the whole system or for instance certain hosts,
IBM Storage Virtualize nodes
– Is the environment and configuration supported?
򐂰 When does the problem occur?
– How often does the problem happen?
– Does the problem happen only at a certain time of day or night?
– What kind of activities was ongoing at the time the problem was reported?
– Did the problem happen after a change in the environment, such as a code upgrade or
installing software or hardware?
򐂰 Under which conditions does the problem occur?
– Does the problem always occur when the same task is being performed?
– Does a certain sequence of events need to occur for the problem to surface?
– Do any other applications fail at the same time?
򐂰 Can the problem be reproduced?
– Can the problem be recreated, for example by running a single command, a set of
commands, or a particular application?
– Are multiple users or applications encountering the same type of problem?
– Can the problem be reproduced on any other system?

Note: For effective troubleshooting, it is crucial to collect log files as close to the incident as
possible and provide an accurate problem description with a timeline.

118 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

6.1.1 Storage Insights


As discussed in Chapter 5, “IBM Storage Insights and IBM Storage Insights Pro” on page 99,
IBM Storage Insights is an important part of monitoring and ensuring continued availability of
IBM Storage Virtualize systems.

When IBM Support is needed, IBM Storage Insights simplifies uploading logs, speeds
resolution with online configuration data, and provides an overview of open tickets all in one
place.

IBM strongly recommends that all customers install and use this no-charge, cloud-based IBM
application because it provides a single dashboard that provides a clear view of all your IBM
block storage.

For detailed information install and examples, refer to Chapter 5, “IBM Storage Insights and
IBM Storage Insights Pro” on page 99.

6.1.2 Using the GUI


The IBM Storage Virtualize GUI is a good entry point to start troubleshooting with. There are
two essential icons at the top that are accessible from any GUI panel.

As shown in Figure 6-1, the first icon shows IBM Storage Virtualize events, such as an error
or a warning, and the second icon shows suggested, running or recently completed
background tasks.

Figure 6-1 Events icon in the GUI

Chapter 6. Storage Virtualize troubleshooting and diagnostics 119


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

The GUI dashboard provides an at-a-glance view of the system’s condition and notifies you of
any circumstances that require immediate action. It contains sections for performance,
capacity, and system health that provide an overall understanding of what is going on in the
system.

The System Health section in the bottom part of the dashboard provides information about
the health status of hardware, logical, and connectivity components. If you click Expand in
each of these categories, the status of the individual components is shown (see Figure 6-2).
Clicking on More Details takes you to the GUI panel related to that specific component, or
shows more information about it.

Figure 6-2 System Health expanded section in the dashboard

6.1.3 Recommended actions and fix procedure


A Fix procedure, sometimes also referred to as Directed Maintenance Procedure, assists you
in fixing a problem without doing any harm. Whenever one or multiple unfixed errors need to
be addressed, the management GUI provides the means to run the recommended fix
procedure. Therefore, the first step in troubleshooting is to check the event log for
Recommended Actions in Monitoring → Events. The highest priority event, that is the event

120 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

log entry with the lowest four-digit error code, will be highlighted to be addressed first as
shown in Figure 6-3.

Figure 6-3 Recommended actions

Click on Run Fix to launch the Fix Procedure for this particular event. Fix Procedures help
resolving a problem. In the background, a Fix Procedure analyzes the status of the system
and its components and provides further information about the nature of the problem. This is
to ensure, that the actions taken do not lead to undesirable results, as for instance volumes
becoming inaccessible to the hosts. The Fix Procedure then automatically performs the
actions required to return the system to its optimal state. This may include checking for
dependencies, resetting internal error counters and apply updated to the system
configuration. Whenever user interaction is required, you will be shown suggested actions to
take and guided through the same. If the problem can be fixed, the related error in the event
log eventually will be marked as fixed. Also, an associated alert in the GUI will be cleared.

Error codes along with their detailed properties in the event log provide reference information
when a service action is required. The four-digit Error Code is visible in the event log. They
are accompanied by a six-digit Event ID which provides additional details about this event.

Three-digit Node Error Codes are visible in the node status in the Service Assistant GUI. For
more information about messages and codes, see Messages and Codes.

6.1.4 Storage Virtualize failure recovery


An IBM Storage Virtualize system might encounter various kinds of failure recovery in certain
conditions. These are known as Tier 1 (T1) through Tier 4 (T4) recovery.
򐂰 A T1 or Tier 1 Recovery - Node warmstart (node assert) will be logged with error code
2030 in the event log.
A single node assert is a recovery condition that is deployed by the IBM Storage Virtualize
software, when a single node attempts to run an invalid code path or detects a transient
hardware problem.
A T1 recovery alias single-node warmstart, is performed without suspending I/O. This task
can be accomplished because the cluster is configured into redundant pairs of nodes or
node canisters, and the clustering software ensures the deployment of a “replicated
hardened state” across nodes. A single node can encounter an assert condition, perform
a software restart recovery action (capturing first-time debug data), and return to the
clustered system without the suspension of I/O.

Chapter 6. Storage Virtualize troubleshooting and diagnostics 121


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

On warm restart, the assert condition is cleared and the node rejoins the cluster
automatically. Typically, a single node assert restart takes 1 - 5 minutes. Host data I/O
continues as the host OS multipath software redirects the I/O to the partner node of the
same I/O group.
The event with error id 2030 will be logged upon return of the cluster node to the system.
Right-click the 2030 event and mark it as Fixed to prevent repeated alerts and
notifications for the same event.
򐂰 A T2 or Tier 2 recovery is reported in the event log with error code 1001.
The cluster automatically initiated a warmstart to recover from an issue. This process
(error code 1001) successfully restored the system and resumed I/O operations without
data loss. However, temporary access interruption occurred. Host applications may need
a restart, and it is recommended to perform a sanity check on the hosts' file systems
afterward. Additionally, check the status of remote connections, replications and Snapshot
mappings.
After a T2 recovery all configuration commands are blocked until you re-enable them, so
that the unfixed event log entry with error code 1001 is being marked as fixed. It is
recommended that they are not re-enabled until the recovery dumps and trace files from
all nodes have been collected and were reviewed by IBM Support to confirm that it is safe
to do so.
The Service GUI is the preferred method for collecting logs of each node. Open a browser
session to the Service GUI at https://<cluster_ip>/service. Select the Collect Logs
pane from the left navigation bar, and then select the option to create a support package
with the latest statesave.
򐂰 A Tier 3 or T3 Recovery is required when there is no more active cluster node and all
nodes of the clustered system report node error 550 and/or 578. The Recover System
Procedure recovers the system if the system state is lost from all cluster nodes.
The T3 Recovery procedure restores the system configuration to the state it was in before
the incident that caused this situation. Depending on the type of the IBM Storage
Virtualize system and the configuration, this is achieved by ingesting the configuration and
hardened system data. This data is stored on either a quorum Mdisk, quorum drive or an
IP quorum set up to store metadata. In combination with the information stored in the
configuration backup svc.config.backup.xml the system’s configuration and state will be
restored.

Note: Attempt to run the Recover System Procedure only after a complete and thorough
investigation of the cause of the system failure. Try to resolve those issues by using
other service procedures first.

Selecting Monitoring → Events shows information messages, warnings, and issues about
the IBM Storage Virtualize system. Therefore, this area is a good place to check for problems
in the system.

To display the most important events that must be fixed, use the Recommended Actions
filter.

If an important issue must be fixed, look for the Run Fix button in the upper ribbon with an
error message that indicates which event must be fixed as soon as possible. This fix
procedure helps resolve problems. It analyzes the system, provides more information about
the problem, suggests actions to take with the steps to follow, and finally checks to see
whether the problem is resolved.

122 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

Always use the fix procedures to resolve errors that are reported by the system, such as
system configuration problems or hardware failures.

Note: IBM Storage Virtualize systems detect and report error messages; however, events
may have been triggered by factors external to the system, for example back-end storage
devices or the storage area network (SAN).

You can safely mark events as fixed. If the error persists or reoccurs, a new event will be
logged. To select multiple events in the table, press and hold the CTRL key while clicking the
events you want to fix.

Figure 6-4 on page 123 shows Monitoring → Events window with Recommended Run Fix.

Figure 6-4 Monitoring → Events window

Resolve alerts in a timely manner: When an issue or a potential issue is reported,


resolve it as quickly as possible to minimize its impact, and potentially avoid more serious
problems with your system.

To obtain more information about any event, double-click or select an event in the table, and
select Actions → Properties. You can also select Run Fix Procedure and properties by
right-clicking an event.

The properties and details are displayed in a pane, as shown in Figure 6-5. Sense Data is
available in an embedded tab. You can review and click Run Fix to run the fix procedure.

Chapter 6. Storage Virtualize troubleshooting and diagnostics 123


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

Figure 6-5 Properties and Sense Data for an event

6.1.5 Using the command-line interface


Another option to investigate and resolve issues is to use the IBM Storage Virtualize
command-line interface (CLI). Although the fix procedures automatically perform the
necessary steps, it may be sometimes faster and more convenient to run these commands
directly through the CLI. This is particularly the case, when numerous events of the same kind
need to be fixed, as this can be a strenuous task to click each individual event in the GUI.

Run the commands when you encounter the following issues:


򐂰 You experience a back-end storage or internode issue. For example:
– Error code 1370: A managed disk (MDisk) error recovery procedure (ERP) has
occurred.
– Error code 1630: The number of device logins was reduced.
– Error code: 1230 or 1231 Login Excluded.
򐂰 You performed maintenance on the following items:
– Back-end storage subsystems.
– SAN devices like switches, cables, optical transceivers as SFPs.

Important: Execute these commands when any type of change that is related to the
communication between IBM Storage Virtualize systems and back-end storage subsystem
occurs such as back-end storage is configured or a SAN zoning change occurred). This
process ensures that IBM Storage Virtualize recognizes the changes.

Common error recovery involves the following IBM Storage Virtualize CLI commands:
򐂰 detectmdisk
Discovers changes in the SAN and back-end storage.

124 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

򐂰 lscontroller and lsmdisk


Provides the status of all controllers and MDisks. Pay attention to status values other than
online, for instance offline or degraded.
򐂰 lscontroller <controller_id_or_name>
Checks the controller that was causing the issue and verifies that all the worldwide port
names (WWPNs) are listed as you expect. Also check if the path_counts are distributed
evenly across the WWPNs.
򐂰 lsmdisk
Determines whether all MDisks are online.

Note: When an issue is resolved by using the CLI, verify that the error disappears by
selecting Monitoring → Events. If not, make sure to mark the error as fixed.

6.2 Collecting diagnostic data


Problem source identification and problem determination (PSI/PD) may be a challenging task
in complex and heterogeneous IT environments. It is crucial hence to collect the right
diagnostic data at the right time to enable the IBM Remote Support teams to assist you in
resolving a problem. The following section outlines the steps to enable you to collect
diagnostic data to find and isolate problems in an IBM Storage Virtualize environment.

6.2.1 IBM Storage Virtualize systems data collection


When you encounter a problem with an IBM Storage Virtualize system and you need to open
a case with IBM support, you most likely will be asked to provide a support package from the
system. The support package may interchangeably be referred to as Snap or data collection.

Checking for an automatically opened Call Home case


IBM Storage Virtualize system configured for Call Home automatically report events to IBM. A
support case is automatically opened depending on the type of event. It is a good practice to
check first, whether a case already exists.

To do so, check the My Cases section in the IBM My Support portal.

Storage Virtualize systems configured to be monitored in Chapter 5, “IBM Storage Insights


and IBM Storage Insights Pro” on page 99, will show associated support cases there as well.

Alternatively, you may log in with your IBMid to IBM Call Home Connect Cloud. Call Home
Connect Cloud provides an enhanced live view of your assets, including the status of cases,
warranties, maintenance contracts, service levels, and end of service information.
Additionally, Call Home Connect Cloud offers links to other online tools. For example, IBM
Storage Insights.

What data to collect on IBM Storage Virtualize systems


The data that is needed for analysis depends on the kind of problem to be analyzed. An IBM
Storage Virtualize system stores different kinds of log files, message files, statistics, and
traces. The following terms are commonly used in related publications:

Chapter 6. Storage Virtualize troubleshooting and diagnostics 125


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

򐂰 Dump: A node dump or full dump s collected whenever the software restarts for some
reason. It is similar in nature to a core dump file, and it can be used by IBM Remote
Technical Support and development teams to investigate a problem
򐂰 Livedump: A livedump is a binary data capture of the current state of the software. It
causes only minimal impact to I/O operations. The contents of a livedump are similar to
the contents of a dump with slightly less detailed information.
򐂰 Statesave: The term statesave is interchangeably used for either a dump or a livedump.

Four different types of Snap can be collected, Snap Type 1 through Snap Type 4, colloquially
often referred to as Snap/1, Snap/2, Snap/3 or Snap/4. The Snap types vary in the amount of
diagnostic information that is contained in the package:
򐂰 Snap/1: Standard logs including performance stats.
– fastest, smallest, no node dumps.
򐂰 Snap/2: Same as Snap/1 plus one existing statesave, the most recently created dump or
livedump from the current config node.
– slightly slower than snap/1, large.
򐂰 Snap/3: Same as Snap/1 plus the most recent dump or livedump from each active
member node in the clustered system.
򐂰 Snap/4: Same as Snap/1 plus fresh livedump from each active member node in the
clustered, which are created upon triggering the data collection.

Statesaves, dumps and livedumps


When the Storage Virtualize software stack restarts unexpectedly, a dump file is created and
written to a cluster node's boot drive. Similar to a core dump file, this information can be used
by IBM Remote Technical Support and development teams to diagnose the cause of the
software restart.

A livedump is a binary data capture of the current state of the software. It causes only minimal
impact to I/O operations. Livedumps are preferred when the system is still operational and a
detailed snapshot of the current state is needed. The contents of a livedump are similar to the
contents of a dump with slightly less detailed information. Livedumps can be initiated
manually or automatically based on certain events.

Which Snap to collect


Two major factors play a part in deciding which snap to collect for a specific support case:
򐂰 Speed of collection
– A Snap/1 is generated more rapidly, and it is much smaller.
򐂰 Amount of data
– Collecting a Snap/4 as soon as possible after a problem has occurred increases the
likelihood that the livedump contains the data required to diagnose the problem.

Consider the following points:


򐂰 For issues that are related to interoperability with hosts or storage, collect Snap/4.
򐂰 For critical performance issues, collect Snap/1 and then collect Snap/4.
򐂰 For general performance issues, collect Snap/4.
򐂰 For issues that are related to replication (including 1920 errors), collect Snap/4 from both
systems in the replication partnership.
򐂰 For issues that are related to DRPs, collect Snap/4.
򐂰 For 2030, 1196, or 1195 errors, collect Snap/3.

126 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

򐂰 For all other issues, collect Snap/4.

Tip: For urgent cases, start with collecting and uploading a Snap/1 followed by a Snap/4.
This enables IBM Remote Support to quicker commence the analysis, while the more
detailed Snap/4 is being collected and uploaded.

For more information about the required support package that is most suitable to diagnose
different type of issues and their content, see What data should you collect for a problem on
IBM Storage Virtualize systems?

Recommendation: After an issue is solved, it is a best practice to do some housekeeping


and delete old dumps on each node by running the following command:
cleardumps -prefix /dumps node_id | node_name

Support package collection and upload


The most commonly used method to collect and upload a support package is using the
Storage Virtualize GUI. However, using Storage Insights is even more convenient, as it
automates uploading the collected package to IBM. This method is described in 5.2.5,
“Managing existing support tickets by using IBM Storage Insights” on page 112.

By default, Storage Virtualize offers two options for automatic support package upload:
򐂰 Automatic upload via management interface: You can configure Storage Virtualize to
automatically collect and upload support packages to the IBM Support Center. This can be
done through the GUI or CLI.
򐂰 Download and manual upload: Alternatively, Storage Virtualize allows you to download
the support package locally to your device. You can then manually upload it to the IBM
Support Center if needed.

Collecting data by using the GUI


To collect data by using the GUI, complete the following steps:
1. Select Settings → Support → Support Package. You can choose to download the
support package to your workstation or upload it to the IBM Support Center. The latter
option requires internet connectivity for the Storage Virtualize system, either directly or
through a configured proxy server.
2. To automatically upload the support packages, click Upload Support Package.
3. Select Create New Package and Upload.
4. In the pop-up window, enter the IBM Support case number (TSxxxxxxxxx) and the type of
support package to upload to the IBM Support Center. Press Upload. You can monitor the
progress of the individual sub-tasks by clicking View more details as shown in Figure 6-6.

Chapter 6. Storage Virtualize troubleshooting and diagnostics 127


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

Figure 6-6 Upload Support Package details

Collecting data by using the CLI


Log in to the CLI and run the command that matches the type of snap that is requested:
򐂰 Standard logs (Snap/1):
svc_snap gui1 upload pmr=TSxxxxxxxxx
򐂰 Standard logs plus one existing statesave from current config node (Snap/2):
svc_snap gui2 upload pmr=TSxxxxxxxxx
򐂰 Standard logs plus most recent statesave from each active cluster node (Snap/3):
svc_snap gui3 upload pmr=TSxxxxxxxxx
򐂰 Standard logs plus new statesaves (Snap/4):
svc_livedump -nodes all -yes
svc_snap gui3 upload pmr=TSxxxxxxxxx

To collect a Snap/4 using the CLI, a livedump of each active node must be generated by using
the svc_livedump command. Then, the log files and newly generated dumps are uploaded by
using the svc_snap gui3 command, as shown in Example 6-1. To verify whether the support
package was successfully uploaded, use the sainfo lscmdstatus command (TSXXXXXXX is
the case number).

Note: The use of Service Assistant commands as sainfo or satask requires superuser
privileges.

128 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

Example 6-1 The svc_livedump command


IBM_FlashSystem:FS9110:superuser>svc_livedump -nodes all -yes
Livedump - Fetching Node Configuration
Livedump - Checking for dependent vdisks
Livedump - Check Node status
Livedump - Preparing specified nodes - this may take some time...
Livedump - Prepare node 1
Livedump - Prepare node 2
Livedump - Trigger specified nodes
Livedump - Triggering livedump on node 1
Livedump - Triggering livedump on node 2
Livedump - Waiting for livedumps to complete dumping on nodes 1,2
Livedump - Waiting for livedumps to complete dumping on nodes 1
Livedump - Successfully captured livedumps on nodes 1,2

IBM_IBM FlashSystem:FLASHPFE95:superuser>svc_snap gui3 upload pmr=TSxxxxxxxxx


Collecting data
Packaging files
Snap data collected in /dumps/snap.serial.YYMMDD.HHMMSS.tgz

IBM_FlashSystem:FS9110:superuser>sainfo lscmdstatus
last_command satask supportupload -pmr TSxxxxxxxxx -filename
/dumps/snap.serial.YYMMDD.HHMMSS.tgz
last_command_status CMMVC8044E Command completed successfully.
T3_status
T3_status_data
cpfiles_status Complete
cpfiles_status_data Copied 160 of 160
snap_status Complete
snap_filename /dumps/snap.serial.YYMMDD.HHMMSS.tgz
installcanistersoftware_status
supportupload_status Active
supportupload_status_data Uploaded 267.5 MiB of 550.2 MiB
supportupload_progress_percent 48
supportupload_throughput_KBps 639
supportupload_filename /dumps/snap.serial.YYMMDD.HHMMSS.tgz

If you do not want to automatically upload the snap to IBM, omit the upload pmr=TSxxxxxxxxx
command option. When the snap creation completes, all collected files are packaged into a
gzip-compressed tarball that uses the following format:
/dumps/snap.<panel_id>.YYMMDD.hhmmss.tgz

The creation of the Snap archive takes a few minutes to complete. Depending on the size of
the system and the configuration, it can take considerably longer, particularly if fresh
livedumps are being created.

The generated file can be retrieved from the GUI by selecting Settings → Support →
Manual Upload Instructions → Download Support Package, and then clicking Download
Existing Package. Find the exact name of the snap that was generated by running the
svc_snap command that was run earlier. Select that file, and click Download.

In certain circumstances it may be necessary to collect a livedump from an individual node of


a clustered system at a certain point in time. This can be achieved through CLI commands
preplivedump and triggerlivedump, followed by the targeted node’s numeric id or name. The

Chapter 6. Storage Virtualize troubleshooting and diagnostics 129


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

livedump status of a node can be checked witch command lslivedump. Once the status has
changed from dumping to inactive, the livedump file is ready to be copied off the system
using either the GUI or scp command.

Example 6-2 preplivedump and lslivedump commands


IBM_FlashSystem:FS9110:superuser>preplivedump 2
IBM_FlashSystem:FS9110:superuser>lslivedump 2
status
prepared
IBM_FlashSystem:FS9110:superuser>triggerlivedump 2
IBM_FlashSystem:FS9110:superuser>lslivedump 2
status
dumping
IBM_FlashSystem:FS9110:superuser>lslivedump 2
status
inactive
IBM_FlashSystem:FS9110:superuser>lsdumps 2
[...]
livedump.panel_id.YYMMDD.HHMMSS
[...]

6.2.2 Drive data collection: drivedumps


BM FlashCore Modules (FCMs) are a family of high-performance flash drives. The FCM
design utilizes the NVMe protocol, a Peripheral Component Interconnect® Express (PCIe)
interface, and high-speed NAND memory (e.g., Single-Level Cell) driven by a customizable
Field-Programmable Gate Array (FPGA) to deliver high throughput, inline compression, and
consistent IOPS (Input/Output Operations Per Second) with predictable latency.

For deeper analysis in cases where drives or FCMs are involved, drivedumps are often
useful. Drivedumps are particularly useful for troubleshooting issues with FCMs, as they
capture the low-level state of the drive. Their data can help you understand problems with the
drive, and they do not contain any data that applications write to the drive. In some situations,
drivedumps are automatically triggered by the system.

To collect support data from a disk drive, run the triggerdrivedump drive_id command. The
output is stored in a file in the /dumps/drive directory. This directory is located on one of the
nodes that are connected to the drive.

Example 6-3 shows the usage of the triggerdrivedump command.

Example 6-3 The triggerdrivedump command


IBM_IBM FlashSystem:FS9110:superuser>triggerdrivedump 1
Drive dump on node id [5] successfully created
IBM_IBM FlashSystem:FS9110:superuser>

IBM_IBM FlashSystem:FS9110:superuser>lsdumps -prefix /dumps/drive


id filename
0 drivedump_7812345-1_1_220411_055205
IBM_IBM FlashSystem:FS9110:superuser>

Any snap that is taken after the trigger command contains the stored drivedumps. It is
sufficient to provide Snap Type 1: Standard logs for drivedumps.

130 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

6.2.3 Host multipath software


If a problem occurs that is related to host communication with an IBM Storage Virtualize
system, collecting data from hosts and their multipath software is useful.

Linux using device-mapper-multipath (dmmp)


To troubleshoot by using the multipathd CLI, issue the multipath -ll command, which
shows detailed information about the multipath devices.

Example 6-4 shows the output for the command multipath -ll, including the following
information:
򐂰 Name of the mpath device (mpatha / mpathb).
򐂰 UUID of the mpath device.
򐂰 Discovered paths for each mpath device, including the name of the sd-device, the priority,
and state information.

Example 6-4 Output for the multipath -ll command


root@myServer ~]# multipath -ll
mpatha (3600507680185801aa000000000000b79) dm-3 IBM ,2145
size=100G features='1 queue_if_no_path' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 16:0:0:2 sdl 8:176 active ready running
| `- 18:0:0:2 sdm 8:192 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 16:0:1:2 sdg 8:96 active ready running
`- 18:0:1:2 sdt 65:48 active ready running
mpathb (3600507680185801aa000000000000b78) dm-4 IBM ,2145
size=100G features='1 queue_if_no_path' hwhandler='0' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 16:0:1:1 sde 8:64 active ready running
| `- 18:0:1:1 sds 65:32 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 16:0:0:1 sdj 8:144 active ready running
`- 18:0:0:1 sdk 8:160 active ready running

Expand the command to multipath -ll -v3 to print debug information.

You can also use the multipathd interactive console for troubleshooting. The multipath -k
command opens an interactive interface to the multipathd daemon.

Entering this command opens an interactive multipath console. After running this command, it
is possible to enter help to get a list of available commands, which can be used within the
interactive console. To exit the console, press Ctrl-d.

To display the current configuration, including the defaults, issue show config within the
interactive console.

AIX using multipath I/O


Table 6-1 shows some of the useful AIX® lspath commands.

Chapter 6. Storage Virtualize troubleshooting and diagnostics 131


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

Table 6-1 Useful AIX lspath commands


Command Result

lspath Lists all paths for all hdisks with their status and parent FSCSI
(Fibre Channel SCSI) device information.

lspath -H -l hdisk1 List all paths for the specified hdisk with its status and
corresponding FSCSI device information. The output includes a
column header.

lspath -l hdisk1 -HF "name Lists more detailed information about the specified hdisk the parent
path_id parent connection FSCSI device and its path status.
path_status status"

lspath -s disabled Lists all paths whose operational status is disabled.

lspath -s failed Lists all path whose operational status is failed.

lspath -AHE -l hdisk0 -p Display attributes for a path and connection (-w) (-A is like lsattr
vscsi0 -w "810000000000" for devices. If only one path exists to the parent device, the
connection can be omitted by running:
lspath -AHE -l hdisk0 -p vscsi0)

lspath -l hdisk1 -a Lists the priority for a specific path.


priority -F value -p fscsi0
-w 500507680d7e1264,0

Table 6-2 shows some of the useful AIX lsmpio commands.

Table 6-2 Useful AIX lsmpio commands


Command Result

lsmpio Shows all disks and corresponding paths with state, parent, and
connection information.

lsmpio -q Shows all disks with vendor ID, product ID, size, and volume name.

lsmpio -ql hdisk0 Shows detailed disk information like:


򐂰 Vendor ID
򐂰 Product ID
򐂰 Capacity
򐂰 Machine Type
򐂰 Model Number
򐂰 Host Group
򐂰 Volume Name
򐂰 Volume Serial Number

lsmpo -Sl hdisk0 | grep Shows path statistics.


Path

lsmpio -ar Lists the parent adapter and remote port information (-a: adapter
(local), and -r: remote port).

lsmpio -are Lists the parent adapter and remote port error statistics (-e: error).

lsmpio -z Lists all multipath I/O (MPIO) statistics.

Windows using MPIO


Because IBM Storage Virtualize 8.3.0 is the last version that supports Subsystem Device
Driver Device Specific Module (SDDDSM), you must use native Windows multipathing,
which is provided by the installable feature MPIO.

132 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

Besides managing the multipathing configuration by using the Windows GUI, it is possible to
use the CLI by using the tool mpclaim.exe, which is installed by default.

Table 6-3 shows some of the useful Windows mpclaim.exe commands.

Table 6-3 Useful Windows mpclaim.exe commands


Command Result

mpclaim.exe -e View the storage devices that are discovered by the system.

mpclaim.exe -r -i -d Manages FC devices with MPIO.


_IBM_2145

mpclaim.exe -r -u -d Removes MPIO management of FC devices.


_IBM_2145

mpclaim.exe -r -i Manages internet Small Computer Systems Interface (ISCSI)


-d"MSFT2005iSCSIBusType_0x9 devices with MPIO.
"

mpclaim.exe -r -u Removes MPIO management of iSCSI devices.


-d"MSFT2005iSCSIBusType_0x9
"

mpclaim.exe -r -i -a"" Manages all storage devices with MPIO.

mpclaim.exe -r -u -a"" Removes MPIO management for all devices.

mpclaim.exe -r View storage devices that are managed by Microsoft DSM.

mpclaim.exe -L -M<num> Modifies the load-balancing policy.

mpclaim.exe -s -d Checks the policy that your volumes are currently using.

mpclaim.exe -s -d <number> Checks the policy for a specific disk.

Generic MPIO settings can be listed and modified by using Windows PowerShell cmdlets.
Table 6-4 shows the PowerShell cmdlets, which may be used to list or modify generic
Windows MPIO settings.

Table 6-4 Useful Windows PowerShell cmdlets


Command Result

Get-MSDSMSupportedHW The cmdlet lists hardware IDs in the Microsoft Device Specific
Module (MSDSM) supported hardware list.

Get-MPIOSetting The cmdlet gets Microsoft MPIO settings. The settings are as
follows:
򐂰 PathVerificationState
򐂰 PathVerificationPeriod
򐂰 PDORemovePeriod
򐂰 RetryCount
򐂰 RetryInterval
򐂰 UseCustomPathRecoveryTime
򐂰 CustomPathRecoveryTime
򐂰 DiskTimeoutValue

Chapter 6. Storage Virtualize troubleshooting and diagnostics 133


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

Command Result

Set-MPIOSetting The cmdlet changes Microsoft MPIO settings. The settings are as
follows:
򐂰 PathVerificationState
򐂰 PathVerificationPeriod
򐂰 PDORemovePeriod
򐂰 RetryCount
򐂰 RetryInterval
򐂰 UseCustomPathRecoveryTime
򐂰 CustomPathRecoveryTime
򐂰 DiskTimeoutValue

VMware using VMware native multipathing


There are two methods that are used to obtain the multipath information from the VMware
ESX host:
򐂰 ESXi CLI: Use the CLI to obtain the multipath information when performing
troubleshooting procedures.
򐂰 vSphere Client and vSphere Web Client: Use this option when you are performing system
maintenance.

Command-line interface
To obtain logical unit number (LUN) multipathing information from the ESXi host CLI,
complete the following steps:
1. Log in to the ESXi host console.
2. To get detailed information about the paths, run esxcli storage core path list.
Example 6-5 shows an example for the output of the esxcli storage core path list
command.

Example 6-5 Output of esxcli storage core path list command


fc.5001438028d02923:5001438028d02922-fc.500507680100000a:500507680120000a-naa.6
00507680185801aa000000000000a68
UID:
fc.5001438028d02923:5001438028d02922-fc.500507680100000a:500507680120000a-naa.6
00507680185801aa000000000 000a68
Runtime Name: vmhba2:C0:T1:L54
Device: naa.600507680185801aa000000000000a68
Device Display Name: IBM Fibre Channel Disk
(naa.600507680185801aa000000000000a68)
Adapter: vmhba2
Channel: 0
Target: 1
LUN: 54
Plugin: NMP
State: active
Transport: fc
Adapter Identifier: fc.5001438028d02923:5001438028d02922
Target Identifier: fc.500507680100000a:500507680120000a
Adapter Transport Details: WWNN: 50:01:43:80:28:d0:29:23 WWPN:
50:01:43:80:28:d0:29:22
Target Transport Details: WWNN: 50:05:07:68:01:00:00:0a WWPN:
50:05:07:68:01:20:00:0a

134 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

Maximum I/O Size: 33553920

3. To list detailed information for all the corresponding paths for a specific device, run esxcli
storage core path list -d <naaID>.
Example 6-6 shows the output for the specified device with the ID
naa.600507680185801aa000000000000972, which is attached with eight paths to the ESXi
server. The output was omitted for brevity.

Example 6-6 Output of esxcli storage core path list -d <naaID>


fc.5001438028d02923:5001438028d02922-fc.500507680100037e:500507680120037e-naa.600507680185801aa0
00000000000972
UID:
fc.5001438028d02923:5001438028d02922-fc.500507680100037e:500507680120037e-naa.600507680185801aa0
00000000000972
Runtime Name: vmhba2:C0:T3:L9
Device: naa.600507680185801aa000000000000972
Device Display Name: IBM Fibre Channel Disk (naa.600507680185801aa000000000000972)
Adapter: vmhba2
Channel: 0
Target: 3
LUN: 9
Plugin: NMP
State: active
Transport: fc
Adapter Identifier: fc.5001438028d02923:5001438028d02922
Target Identifier: fc.500507680100037e:500507680120037e
Adapter Transport Details: WWNN: 50:01:43:80:28:d0:29:23 WWPN: 50:01:43:80:28:d0:29:22
Target Transport Details: WWNN: 50:05:07:68:01:00:03:7e WWPN: 50:05:07:68:01:20:03:7e
Maximum I/O Size:
33553920fc.5001438028d02923:5001438028d02922-fc.500507680100037e:500507680130037e-naa.6005076801
85801aa000000000000972
UID:

fc.5001438028d02923:5001438028d02922-fc.500507680100037e:500507680130037e-naa.600507680185801aa0
00000000000972
Runtime Name: vmhba2:C0:T2:L9
Device: naa.600507680185801aa000000000000972
Device Display Name: IBM Fibre Channel Disk (naa.600507680185801aa000000000000972)
Adapter: vmhba2
Channel: 0
Target: 2
LUN: 9
Plugin: NMP
State: active
Transport: fc
Adapter Identifier: fc.5001438028d02923:5001438028d02922
Target Identifier: fc.500507680100037e:500507680130037e
Adapter Transport Details: WWNN: 50:01:43:80:28:d0:29:23 WWPN: 50:01:43:80:28:d0:29:22
Target Transport Details: WWNN: 50:05:07:68:01:00:03:7e WWPN: 50:05:07:68:01:30:03:7e
Maximum I/O Size:
33553920fc.5001438028d02921:5001438028d02920-fc.500507680100037e:500507680110037e-naa.6005076801
85801aa000000000000972
UID:

Chapter 6. Storage Virtualize troubleshooting and diagnostics 135


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

4. The command esxcli storage nmp device list lists the LUN multipathing information for
all attached disks.
Example 6-7 shows the output for one of the attached disks. All other output was omitted
for brevity.

Example 6-7 Output for esxcli storage nmp device list


naa.600507680185801aa000000000000a68
Device Display Name: IBM Fibre Channel Disk
(naa.600507680185801aa000000000000a68)
Storage Array Type: VMW_SATP_ALUA
Storage Array Type Device Config: {implicit_support=on;
explicit_support=off; explicit_allow=on; alua_followover=on;
action_OnRetryErrors=off; {TPG_id=1,TPG_state=ANO}{TPG_id=0,TPG_state=AO}}
Path Selection Policy: VMW_PSP_RR
Path Selection Policy Device Config:
{policy=rr,iops=1000,bytes=10485760,useANO=0; lastPathIndex=1:
NumIOsPending=0,numBytesPending=0}
Path Selection Policy Device Custom Config:
Working Paths: vmhba2:C0:T3:L54, vmhba1:C0:T2:L54, vmhba1:C0:T3:L54,
vmhba2:C0:T2:L54
Is USB: false

vSphere Client HTML5 and Web Client


To obtain multipath settings for your storage in the HTML5 client, complete the following
steps:
1. Select an ESXi host, and click the Configure tab.
2. Click Storage Devices.
3. Select the storage device that you want to verify.
4. Scroll down in the Properties tab and click Edit multipathing....

vSphere Client (Thick Client for 6.x)


To obtain multipath settings for your storage in vSphere Client, complete the following steps:
1. Select an ESXi host, and click the Configuration tab.
2. Click Storage.
3. Select a data store or mapped LUN.
4. Click Properties.
5. In the Properties dialog, select the extent, if necessary.
6. Select Extent Device → Manage Paths and obtain the paths from the Manage Path
dialog.
A more thorough and in-depth discussion of VMware implementation and multipathing can be
found in the following IBM Redbooks publication IBM Storage Virtualize and VMware:
Integrations, Implementation and Best Practices, SG24-8549.

136 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 22, 2024 9:23 am 8561ch06.fm

6.2.4 More data collection


Data collection methods vary by storage platform, SAN switch, and operating system.

For an issue in a SAN environment when it is not clear where the problem is occurring, you
might need to collect data from several devices in the SAN.

The following basic information must be collected for each type of device:
򐂰 Hosts:
– Operating system: Version and level
– Host Bus Adapter (HBA): Driver and firmware level
– Multipathing driver level
򐂰 SAN switches:
– Hardware model
– Software version
򐂰 Storage subsystems:
– Hardware model
– Software version

For performance-related issues, it is helpful to have corresponding monitoring. IBM Storage


Insights and IBM Spectrum Control are recommended. If required, you can export
performance data from there for the related period taken from the historical data.

Chapter 6. Storage Virtualize troubleshooting and diagnostics 137


8561ch06.fm Draft Document for Review July 22, 2024 9:23 am

138 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
Draft Document for Review July 19, 2024 5:41 pm 8561bibl.fm

Related publications

The publications listed in this section are considered particularly suitable for a more detailed
discussion of the topics covered in this book.

IBM Redbooks
The following IBM Redbooks publications provide additional information about the topic in this
document. Note that some publications referenced in this list might be available in softcopy
only.
򐂰 Implementation Guide for IBM Storage FlashSystem and IBM SAN Volume Controller
Updated for IBM Storage Virtualize Version 8.6, SG24-8542
򐂰 IBM Storage Virtualize and VMware: Integrations, Implementation and Best Practices,
SG24-8549
򐂰 Ensuring Business Continuity: A Practical Guide to Policy-Based Replication and
Policy-Based High Availability for IBM Storage Virtualize Systems, SG24-8569
򐂰 Introduction and Implementation of Data Reduction Pools and Deduplication, SG24-8430
򐂰 IBM Storage Insights Security Guide, SC27-8774
򐂰 Data Resiliency Designs: A Deep Dive into IBM Storage Safeguarded Snapshots,
REDP-5737

You can search for, view, download or order these documents and other Redbooks,
Redpapers, Web Docs, draft and additional materials, at the following website:
ibm.com/redbooks

Online resources
These websites are also relevant as further information sources:
򐂰 IBM Storage FlashSystem information:
https://www.ibm.com/flashsystem/
򐂰 IBM SAN Volume Controller information:
https://www.ibm.com/products/san-volume-controller?mhsrc=ibmsearch_a&mhq=SAN%20
Volume%20Controller
򐂰 IBM System Storage Interoperation Center (SSIC):
https://www.ibm.com/systems/support/storage/ssic/interoperability.wss

Help from IBM


IBM Support and downloads
ibm.com/support

© Copyright IBM Corp. 2024. 139


8561bibl.fm Draft Document for Review July 19, 2024 5:41 pm

IBM Global Services


ibm.com/services

140 Unleash the Power of Flash: Getting Started with IBM Storage Virtualize Version 8.7
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth which has a PPI of 526. Divided
250 by 526 which equals a spine width of .4752". In this case, you would use the .5” spine. Now select the Spine width for the book and hide the others: Special>Conditional
Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and File>Import>Formats the
Conditional Text Settings (ONLY!) to the book files.
Draft Document for Review July 19, 2024 5:41 pm 8561spine.fm 141
Unleash the Power of Flash: SG24-8561-00
Getting Started with ISBN
(1.5” spine)
1.5”<-> 1.998”
789 <->1051 pages
Unleash the Power of Flash: SG24-8561-00
Getting Started with ISBN
(1.0” spine)
0.875”<->1.498”
460 <-> 788 pages
SG24-8561-00
Unleash the Power of Flash: Getting Started with IBM Storage ISBN
(0.5” spine)
0.475”<->0.873”
250 <-> 459 pages
Unleash the Power of Flash: Getting Started with IBM Storage Virtualize
(0.2”spine)
0.17”<->0.473”
90<->249 pages
(0.1”spine)
0.1”<->0.169”
53<->89 pages
To determine the spine width of a book, you divide the paper PPI into the number of pages in the book. An example is a 250 page book using Plainfield opaque 50# smooth which has a PPI of 526. Divided
250 by 526 which equals a spine width of .4752". In this case, you would use the .5” spine. Now select the Spine width for the book and hide the others: Special>Conditional
Text>Show/Hide>SpineSize(-->Hide:)>Set . Move the changed Conditional text settings to all files in your book by opening the book file with the spine.fm still open and File>Import>Formats the
Conditional Text Settings (ONLY!) to the book files.
Draft Document for Review July 19, 2024 5:41 pm 8561spine.fm 142
Unleash the Power of SG24-8561-00
Flash: Getting Started with ISBN
(2.5” spine)
2.5”<->nnn.n”
1315<-> nnnn pages
Unleash the Power of Flash: SG24-8561-00
Getting Started with ISBN
IBM Storage Virtualize 8.7 on IBM
(2.0” spine)
2.0” <-> 2.498”
1052 <-> 1314 pages
Back cover
Draft Document for Review July 22, 2024 9:34 am

SG24-8561-00

ISBN

Printed in U.S.A.

®
ibm.com/redbooks

You might also like