Continuous Availability S390 Technology Guide
Continuous Availability S390 Technology Guide
http://www.redbooks.ibm.com
This book was printed at 240 dpi (dots per inch). The final production redbook with the RED cover will
be printed at 1200 dpi and will provide superior graphics resolution. Please see “How to Get ITSO
Redbooks” at the back of this book for ordering instructions.
SG24-2086-00
IBML
SG24-2086-00
International Technical Support Organization
December 1998
Take Note!
Before using this information and the product it supports, be sure to read the general information in
Appendix A, “Special Notices” on page 291.
When you send information to IBM, you grant IBM a non-exclusive right to use or distribute the information in any
way it believes appropriate without incurring any obligation to you.
Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
The Team That Wrote This Redbook . . . . . . . . . . . . . . . . . . . . . . . . . xv
Comments Welcome . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xvi
Chapter 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1 Continuous Availability Systems Design . . . . . . . . . . . . . . . . . . . . 1
Chapter 4. OS/390 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
4.1 OS/390 Availability Attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . 59
4.1.1 Availability Enhancements . . . . . . . . . . . . . . . . . . . . . . . . . . 60
4.1.2 OS/390 - Continuous Availability . . . . . . . . . . . . . . . . . . . . . . 64
4.1.3 SmartBatch . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70
4.1.4 DFSMS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73
4.1.5 JES3 Improved Availability in OS/390 R4 . . . . . . . . . . . . . . . . . 74
4.1.6 Recoverable Resource Management Services (RRMS) . . . . . . . . 75
4.2 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
Contents v
9.8.4 DB2 Attachment Facility . . . . . . . . . . . . . . . . . . . . . . . . . . 181
9.8.5 CICS Interface to MQSeries . . . . . . . . . . . . . . . . . . . . . . . . 182
9.8.6 Internet Access to CICS Applications . . . . . . . . . . . . . . . . . . 182
9.8.7 Backup-While-Open (BWO) . . . . . . . . . . . . . . . . . . . . . . . . 183
Contents vii
14.2.3 Data Piping Transfer . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
14.3 Storage Sharing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
14.4 Real Data Sharing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262
14.5 The Seascape Vision . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 297
Figures xi
xii Continuous Availability S/390 Technology Guide
Tables
The first volume, Continuous Availability Systems Design Guide , discusses the
basic concepts of continuous availability and describes a structured approach to
developing and implementing a continuous availability solution. It is strongly
recommended that Continuous Availability Systems Design Guide , be read as a
prerequisite to the material presented here.
The objective of this book is to list all System/390 related products and product
features that can contribute to a continuous availability solution, and describe
their functional capabilities in relation to continuous availability. However, this
book does not just deal with technology and solution components. It also
discusses key theories, philosophies, and technical challenges that are of
importance to continuous availability systems design.
This publication is aimed at individuals who are involved in the design and
implementation of continuous availability solutions, for the IBM S/390 platform. It
will assist them in:
• Designing an effective continuous availability solution to meet predetermined
business requirements and associated DP requirements.
• Selecting suitable products to match the design.
• Implementing and executing the continuous availability solution.
The information in this publication reflects the present functions and capabilities
of IBM products. However, where possible, future trends and directions have
been taken into account.
Comments Welcome
Your comments are important to us!
However, a continuously available system is not simply built from hardware and
software stock components. More than that, it is the result of an individual
systems design process that researches your business needs, and creates a
solution individually tailored for your individual organization.
As with any design project, a structured approach helps to ensure that all of
these factors are considered and suitably addressed.
Figure 1 on page 3 shows the main activities required in the planning and
implementing of a continuous availability capability. These activities are:
1. Determining what the business requires
During this initial part of the process, you need to determine which of the two
aspects of continuous availability, namely, high availability and continuous
operation, are required to which degree for which business processes.
Many business processes are so dependent upon information technology (IT)
resources that they come to a complete standstill in the event of an outage.
There might be regulatory reasons, too, that require a business process to
be available at all times.
By performing a business impact analysis , the negative impact of such
outages, that is, loss of business and revenue, must be evaluated.
Chapter 1. Introduction 3
1.1.1.1 Iterative Design
Often, the activities we just described do not follow each other strictly in
sequence.
In some instances, implementation work may start even while portions of the
solution design are still being developed.
In other instances, the activities associated with a certain step are achieved in
an earlier step. For example, an organization may decide to use fault-tolerant
disk subsystems even before the full business requirements are established.
More importantly, you will often reach a point in the design process where some
aspect of the solution causes you to rework an earlier stage. In this case, the
process flow may have to loop back and modify the requirements or the design.
Understandably, you may not agree with the proposed design sequence. For
instance, some people may say that in many cases, you have to look at available
products (step 4) before you can come up with the overall design (step 3). We
suggest leaving these arguments to the experts (for instance, availability design
consultants) who do the design in each individual case. This redbook is less an
exercise in system design theory than a comprehensive discussion of planning
topics pertaining to continuous availability.
Table 1 (Page 1 of 2). S/390 Processor Features Continuous Availability Matrix. Legend: X = A p p l i e s ,
Type=Outage Type, HA=Unplanned, CO=Planned, CA=Both
Availability Attribute Reduce Reduce Reduce Type
Outage Outage Outage
Frequency Duration Scope
Processor Design X HA
Fault-Tolerant Design X HA
Concurrent CP Sparing X HA
Dynamic SAP Sparing/Reassignment X HA
Dynamic Coupling Facility Dispatching X X X CO
Error Correction Code X HA
Dynamic Memory Sparing X HA
Sybsystem Storage Protection X X HA
Subspace Group Facility X X HA
Up to 15 LPAR: Software Isolation X X HA
LPAR Dynamic Storage Reconfiguration X CO
LPAR Time Management Reporting CA
ESCON Multiple Image Facility X CO
Automatic Reconfiguration Management X CO
Scrubbing X HA
Partial Memory Restart X HA
Hot-Plugging Channels X CO
Dynamic I/O Reconfiguration Management X CO
Partial I/O Restart X X CO
Concurrent Maintenance X CO
Independent Dual Power Feed X HA
N + 1 Power Supply Technology X HA
Internal Battery Feature X HA
Concurrent Licensed Internal Code Path X CO
Console Integration X CO
Serviceability X HA
Internal Coupling Facility X X X CA
The CMOS chip, at 2,000,000 circuits, contains more than enough circuits to build
a S/390 processing unit. The bipolar 711-based technology, in comparison, takes
somewhere in the neighborhood of 400 chips to build a S/390 PU.
From the point of view &ca, this new packaging lessens the complexity of
installing, monitoring and controlling your data processing environment.
Because of the smaller overall package and fewer parts, maintenance
complexity and cost are also greatly reduced. In addition, the inherent low
power required by CMOS technology implies a significant reduction in
environmental costs.
There are four types of CMOS chips that are packaged onto the multi-chip
module (MCM). Each of these chips has different functions.
There is the Processing Unit (PU) chip, L2 chip (L), the Bus Switching Network
(BSN) chip and the Memory Bus Adapter (MBA) chip. The following sections
explain these chips in more detail. The PU chip contains the Level 1 cache (64
Kbytes), the control store (32 Kbytes) and the execution unit. Part of the
execution unit is the floating point coprocessor.
The PU chips are mounted on the 127x127 mm multi-chip module (MCM). Using
the 10-way RX5 and RY5 models as an example, there are twelve of these chips
mounted on the MCM; ten of the PUs would be S/390 central processors (CPs),
and two PUs would be system assist processors (SAPs).
The SAPs are S/390 PUs with special Licensed Internal Code (LIC) to allow these
processors to perform dedicated I/O function (similar to the IOP in previous
water-cooled processors).
For G4 systems, the multi-chip module (MCM) is the critical package for the
processor. This substrate is mounted on the processor board along with power,
memory, and some I/O capability. A total 10-way model (RX5 or RY5) can be
contained within one or two frames, depending on the I/O configuration. This
includes all power and cooling facilities.
The S/390 G4 server processors have a three-level high speed buffer (HSB)
structure:
• 64KB L1 cache on the PU Chip
• 768KB L2 cache per three processors
• 2MB shared cache
There are four high-speed memory buses that connect processors and I/O
devices to processor storage (PS), so that up to four processing units can
transfer data to/from storage concurrently.
Storage access is interleaved between the storage cards, and this tends to
equalize storage activity across the cards to minimize storage contention.
I/O devices pass data to central storage through the memory bus adapter (MBA).
The physical path from the channel includes the planar board, the channel
adapter card (CHA-IBB not shown) the internal bus buffer (IBB), the self-timed
interface (STI) bus, the bus switching network (BSN), and the storage controller
(STC).
For Basic mode and for dedicated logical partitions in LPAR mode, the following
software is required to support application preservation:
• OS/390 Release 1 or later; MVS/ESA SP Version 5.1 or 5.2
• MVS/ESA SP Version 4.2 or 4.3; MVS ESA SP Version 3.1.3
• VM/ESA Release 2.1 or 2.2
VSE/ESA does not support this function. There are no software corequisites
when running shared CPs in LPAR mode.
All G4 servers, with the exception of the maximum configured models (for
example, the 10+2 model), provide at least one spare PU. The rules for the
allocation of CPs and SAPs are as follows:
• Standard CPs are allocated starting at the lowest PU number and continue in
ascending order.
• SAPs are allocated starting at the highest available PU number and continue
in descending order.
• ICFs follow standard CPs and continue in ascending order.
Note: Model R45 with the Cryptographic Coprocessor function has
non-sequential CP addresses (in order to use both coprocessors); model R45 has
CP 0/1/2/4 (CP 0 and CP4 have the coprocessor attached).
Currently, CICS V3 (and above) is the only user of SSSP. It is supported by the
9021 711-based, 9021 520-based, 9121, 9221, 9672 processors, and the S/390
G3/G4 server. It protects your applications from some CICS outages due to
overwriting CICS code.
2.2.3.13 Up to 15 LPARs
PR/SM supports 15 LPARs in a SI-mode (only 10 LPARs on 9672 R1, R2, R3
models). Fifteen LPARs are supported on S/390 G3/G4 processors only.
2.2.3.17 Scrubbing
Storage background scrubbing provides continuous monitoring of storage for the
correction of detected faults before the storage is used.
Timing is important because, until a reset is issued that resets any outstanding
allegiance and releases outstanding RESERVEs to DASD (for example, such a
reset is issued implicitly when an operator IPLs a system), other systems
sharing that DASD could be locked out. This facility is also known as
system-initiated reset.
The IBF enables between 3.5 to 20 minutes (3 to 15 minutes for the RY5) of full
power capability for the S/390 G3/G4 server models and the coupling facility
model C04/C05.
For the C04/C05 models, there is an additional function called power save mode.
Power save mode suspends all coupling facility operations, but preserves
memory content for up to one hour. With the installation of an IBF, no additional
floor space is required on the raised floor or in facilities areas.
The IBF is fully redundant and complements the N+1 front-end power
technology. The IBF design is consistent with IBM′s CMOS server power
technology and has no energy conversion losses experienced between a
conventional independent UPS support of a server installation. The IBF has
self-testing capability and is integrated into the diagnostics, including remote
service facility (RSF) support. The IBF is an optional feature. It can be used with
a UPS to provide additional protection.
Multiple MVS/ESA systems (running under PR/SM) can share the same console,
which eliminates the need for multiple backup control units and consoles.
2.2.3.30 Reliability
The standard features that provide a high level of reliability include:
• High-reliability technology components
• Parts integration to reduce the number of parts in the machine
2.2.3.31 Serviceability
The standard features that provide a high level of serviceability include:
• Automatic error detection and fault isolation concurrent with system
operation
• Automatic remote support capability
• Multiple channel swap: an enhancement for channel problem determination,
allowing up to four channels to be swapped concurrently with system
operation.
• High degree of concurrent maintenance capability in hardware and code
• A status panel showing the status of N+1 power system
• Enhanced diagnostics for coupling links
The coupling facility is one of the most critical components of a Parallel Sysplex
that has production data-sharing implemented. The coupling facility holds the
shared data and locks required for the systems in the Parallel Sysplex to share
the workload. If the coupling facility fails, and its data can no longer be shared,
the throughput of the Parallel Sysplex may be severely impacted.
You must avoid the use of an active coupling facility function in a LPAR of a
Central Electronics Complex (CEC) where other LPARs contain active MVS
members. A failure of this CEC will cause the loss of two major components,
which means it will be impossible to rebuild structures in another coupling
facility.
On the other hand, it is possible to run with an LPAR coupling facility operating
only as a backup, and carrying no active workload. It is also possible to use
dynamic coupling facility dispatching as a backup to a standalone coupling
facility.
If you only have a unique CEC running two or more OS/390 LPARs and want to
take advantage of the continuous operation software provided by a production
Parallel Sysplex, what should you do?
ICF and ICMF, working together, allows you this opportunity at a low cost.
One or two ICF features can be ordered, utilizing spare PUs as ICMF or as a
coupling facility. This internal coupling facility is defined as a LPAR with spare
PUs dedicated, and it uses either external coupling links or can run as an ICMF
which emulate the coupling links across LPARs. ICF is significantly less
expensive than a separate, entry-level coupling facility.
Another way, with the optional priced ICF feature, is to move the backup
coupling facility inside the S/390 G3 or G4 Server when a standalone coupling
facility is used as a primary, reducing the need for a second standalone coupling
facility in a multi-system Parallel Sysplex configuration. All of the workload must
be carried by the standalone coupling facility during normal operation.
With ICF, you can also get into the Parallel Sysplex environment to take
advantage of its Continuous Operation protection from software outages. You
can use ICF as a production coupling facility even in a single-CEC Parallel
Sysplex configuration running two or more OS/390 logical partitions sharing
non-database data (VTAM, XCF, JES2, RACF and GRS). Individual OS/390
partitions can be taken down for maintenance or release upgrade without
suffering an application outage, through the data-sharing provided by the
remaining LPAR in the CEC. By using the ICF as an ICMF, both the benefits of
reduced software license charges and reduced hardware cost are afforded, while
providing the software availability features of the Parallel Sysplex.
As shown in Figure 5, dynamic coupling facility dispatching is, like the internal
coupling facility, another option to implement a Parallel Sysplex environment
with only one external standalone CF (9674), and to have a backup coupling
facility inside a G3/G4 Server. This function is standard on all G3/G4 Servers
and it does not have processors allocated exclusively for its use.
This feature provides a standby capability for the backup CF. It runs in an LPAR
with shared CPs on a system which also runs production LPARs. The new
coupling facility control code (CFCC) eliminates the need to dedicate an entire
processor for CF use because it is no longer running in active wait mode. CFCC
now uses minimal processor resources in normal operations. If the primary CF
fails, the systems will automatically allocate and re-build structures in the
backup CF LPAR, increasing its processor resources as required (storage is not
dynamically adjusted). When functions are returned to primary CF, the
processor capacity allocated to backup CF will be reallocated to production.
It requires physical coupling links to all system images, including the ones on
the same CEC where the coupling facility LPAR is running (it does not emulate
coupling links between LPARs like ICMF does).
This function is available on the 10-way models, even though the ICF is not
available.
You can install ICMF prior to installing Coupling Facility channel hardware and
continue to use it after adding Coupling Facility channel hardware to the CEC.
Integrated Coupling Facility logical partitions and integrated logical partitions
running MVS can coexist on the same CEC with:
• Coupling Facility logical partitions that use Coupling Facility channel
hardware
• Logical partitions running MVS that use Coupling Facility channel hardware
• Logical partitions running non-coupled workloads
Combining with ICF, you can also get into a production Parallel Sysplex
environment to take advantage of its Continuous Operation protection from
software outages. Refer to 2.2.4, “9672 Coupling Facility Features” on page 16
for more information about ICF.
The IBF enables between 3.5 to 20 minutes of full power capability for Coupling
Facility Model C04/C05. There is also an additional function called Power Save
Mode. Power Save Mode suspends all Coupling Facility operations but
preserves memory content for up to one hour.
From a continuous availability point of view, and without taking into account
performance considerations, a minimum of two links between each system and
the coupling facility is strongly recommended. This potentially removes a single
For more information about how to estimate the number of coupling facility links
you need, see OS/390 MVS Parallel Sysplex Configuration Cookbook .
Many features of the S/390 Multiprise 2000 are the same as those of the CMOS
G4 server, and here we just list high availability function names. To obtain more
details about a particular function, refer to 2.2.3, “9672 High Availability
Features” on page 9.
The following high availability functions are available on the latest S/390
Multiprise 2000:
• Partial memory restart
• Error correction code
• Dynamic memory sparing
• Dynamic SAP reassignment on models having more than one CP
• Independent dual power feed
• N + 1 power supply technology
• Concurrent repair for:
− ESCON channels
− Parallel channels
− HDDs, disk drawers and disk (SCSI) adapters
− S/390 Open Systems adapter 2
− Power element
− Thermal elements
− Service element
− Hardware management console
• Concurrent Licensed Internal Code for some patches. Not all patches are
non-disruptive.
• Concurrent channel upgrade
• Dynamic Reconfiguration Manager
• Subsystem Storage Protection Facility
The internal disk drawer advances the use of industry standard disks, or JBODs
(Just a Bunch of Disks), in a RAID 1 (mirrored) configuration. With internal disk,
you can attach your disks by a SCSI 2 Fast/Wide adaptor card (eliminating
separate external control units previously required for S/390 storage).
PR/SM LPAR sharing for internal disk enables the sharing of logical volumes of
disk storage across multiple partitions. The sharing of internal disk between
Multiprise 2000 systems or other S/390 systems is not possible.
Both machines are ideal platforms for identifying and developing necessary
changes without impacting the host system. Therefore, the availability of the
host production system is increased, dependencies on the host are eliminated,
and risks to the production system are reduced. This way, these systems could
be part of your continuous availability solution.
Those two servers offer a wide variety of possibilities and creative uses for
business needs.
Because OS/390 and MVS/ESA 5.2.2 provide the XPGA base that allows easy
porting of UNIX applications, you can consider one of these machines for
developing and porting your applications to the OS/390 environment.
The time-of-day (TOD) on the IBM PC Server System/390 is easily changed. The
result is a dedicated platform for date-reliant testing of system applications.
With the rapid approach of the year 2000, it is important that you begin to assess
and plan a course of action to address potential challenges during this transition.
Both machines support the year 2000. When used in accordance with their
associated documentation, both machines correctly process, supply, and accept
date information within and between the twentieth and twenty-first centuries.
2.4 OSA-2
OSA-2 incorporates the functions of a channel adapter, a control unit and
network interfaces into an integrated S/390 feature that provides industry
standard connectivity to Ethernet, Token Ring, and FDDI LANs and ATM
networks, simplifying enterprise networks, and reducing the number of physical
boxes that must be managed.
When the S/390 server is running in logical partition (LPAR) mode, OSA-2
resources can be shared among the defined LPARs (OSA/SF required). OSA-2
resources consist of the TCP/IP and SNA/APPN applications (TCP/IP mode and
SNA mode) and the ports on the OSA-2 features. Both modes can share access
to an OSA-2 feature and can also access the same port on an OSA-2 feature.
Both modes can also be running concurrently.
The OSA-2 ENTR (Ethernet/Token Ring) feature has two independent ports which
can be configured as either Ethernet or Token Ring. Thus, you can have two
Ethernet ports, two Token Ring ports, or one Ethernet and one Token Ring port.
In addition, the OSA-2 ENTR feature supports full duplex/switched environments,
offering configuration flexibility with minimal disruption to the infrastructure.
A FDDI feature supports attachment to a 100 Mbps FDDI LAN. One dual-ring or
single-ring attachment is supported, as well as attachment to an optical bypass
switch.
For information on the specified operating environment and the hardware and
software requirements, refer to Planning for the S/390 Open Systems Adapter
Feature .
TCP/IP protocol processing is performed on the host (TCP/IP for MVS and VM).
With the addition of OSA-2 to the IBM family of products that provide network
connectivity to distributed systems, connectivity can now be accomplished in a
more integrated manner.
This assumes you are also using RouteD which provides automatic and
transparent recovery from controller, interface, and network failures.
S/390 applications can now take advantage of native, high speed, Asynchronous
Transfer Mode (ATM) connections into the ACF/VTAM network without
application change. The Communications Server Release 2, in conjunction with
OSA-2 and OSA/SF provide an ATM Forum user-to-network interface
(UNI)-compliant, native ATM capability for the S/390 server. This support is
available for the APPN environment only.
Native ATM coupled with VTAM′s High Performance Data Transfer (HPDT) and
APPN High Performance Routing (HPR) gives you the ability to utilize advanced
networking functions in the S/390 server. APPN HPR class of service (COS) is
mapped to ATM virtual circuit connection characteristics. Existing applications
require no changes and can fully utilize the capabilities of the ATM network.
2.4.4 OSA/SF
The S/390 Open Systems Adapter Support Facility (OSA/SF) is an application for
customizing and managing OSA-2 features to support different modes of
operation for host-to-LAN services:
• TCP/IP providing communications between TCP/IP clients and host TCP/IP
applications
• SNA/APPN providing communications between SNA/APPN clients and host
SNA/APPN applications
OSA/SF allows you to change port parameters and obtain status about an OSA-2
feature. OSA/SF can also provide direct management to an OSA-2 defined to an
LPAR.
Access to OSA/SF is also available via a command line interface through a REXX
exec.
A call from a REXX exec can be made to the application programming interface
(API) of an OSA/SF application. You can, therefore, automate OSA/SF
procedures.
OSA/SF is required when using the 155 ATM OSA-2, when using any OSA-2
mode other than TCP/IP, or when sharing OSA-2 LAN ports among multiple
logical partitions (LPARs).
A single copy of OSA/SF running in one LPAR is all that is required to configure
all OSA-2 features associated with all of the defined LPARs running on the S/390
Server. For continuous availability purposes, it may be advantageous to have a
second copy of OSA/SF in a second LPAR, in the event the first LPAR is
unavailable.
The support element (SE) provides the supporting hardware interface controls to
the CPC functional element. It is also the interface used by the hardware
management console for the control, monitoring and service functions on the
9672 and 9674 products. This unit is physically mounted within the 9672 or 9674
A-Frame.
The HMC uses a Token-Ring LAN connection to communicate with the SE. When
tasks are performed at the HMC, the commands are sent to one or more SEs
through the LAN. The SEs then issue commands to their associated CPCs.
CPCs can be grouped at the HMC, so that a single command can be passed
along to all the CPCs in the LAN, if desired. This can help you to reduce the
number of HMCs, and to consolidate operations. For example, if you have a
configuration with four 9672s and two 9674s, then there will be six CPCs and SEs
that can be operated across the LAN by one HMC.
Figure 9 on page 28 shows the recommended HMC configuration for a local site.
Four HMCs are used to ensure that the hardware management console functions
are available when needed.
An HMC is required in each machine room for use by service personnel for
installation and maintenance.
In the operations area, two HMCs are recommended for awareness of status
changes and management of the configuration. One is the primary HMC and the
other is used for backup.
Systems support personnel also require access to some HMC functions for
customizing profiles. This access can be provided through DCAF. Note that the
systems support DCAF should access to HMC4, the backup HMC, so that the
operator access to the HMC is not disrupted.
Figure 9 also shows ESCON Director 9032-3 and Sysplex Timer 9037-2 connected
to the HMC LAN. However, note that their console applications are not placed
on the PC that supports the HMC in the machine room. IBM recommends that
you do not place the extra applications on the same PC that supports the HMC
application because when the machine room HMC is unavailable for service
reasons, access to the extra applications′ functions is lost.
An HMC is required in each machine room for use by service personnel. In the
operations area, two HMCs are recommended. One is the primary HMC and the
other is used for backup.
Several different connected LANs are required to support full connectivity to the
HMCs in this configuration.
Table 2 (Page 1 of 4). Functional Differences between I B M 9672-, 9021-, and 9121-Based CPCs
ES/9000 ES/9672
Multiprise
Function 9021 9121 R2,
R1 G3 G4 2000
711 511 R3
Parallel Sysplex:
ICMF S S S S S S S
Dynamic CF Dispatching - - O O O S -
CF Links O O*N O O O O -
Reconfigurable CFR Channel Paths - - - - S S -
HiPerLinks - - - - O*Q O -
CFCC in LP O - S S S S -
CF Shared CP O - S S S S -
CF Level 0, 1, 2, or 3 O - O O O - -
CF Level 4 O O O O O O ICMF
The traditional S/360 channel cables contained a separate path for each bit, and
therefore transmitted all nine bits of a data byte in parallel. ESCON channels
have only a single transmit fiber, so all data bits are transmitted serially. The
terms parallel and serial are often used to distinguish between these two types
of architecture.
Fiber optic transport has several advantages over the existing metal cable used
in processor-to-I/O communications. Among these are:
• Fiber optic transport is much smaller, lighter and more easily handled.
• It is more secure; there is no electromagnetive field around the cable, and
therefore no pickup of the radiated signal from the cable.
• It is not affected by other electromagnetic sources (electrical noise), so no
special cable routing is required to avoid such sources.
The ESCON architecture implemented in S/390 also enhances the level of error
checking and recoverability for channel connections over the parallel
environment.
On replacing the cable, the ESCON architecture ensures that the environment is
not restarted without full checking. The host is forced to check that the
environment it is now reconnected to is the same environment that existed
before the path error event. The ESCON control unit prevents all operation on
this path until this check has been completed successfully by the host. This
offers full protection against accidental wrong connection, and was not possible
under the parallel channel architecture.
Architectural design points like these ensure that the ESCON channel
environment offers the best possible support for continuous availability.
More detail on the general function and specification of ESCON can be found in
the redbook entitled Enterprise Systems Connection (ESCON) Implementation
Guide . This chapter will focus on ESCON′s contribution to Continuous
Availability, although some background is given to assist understanding.
3.2.1 Topology
The parallel channel architecture was capable of supporting the connection of
several control units to each channel, sometimes called daisy-chaining or
multi-dropping.
Consider the three types of connection shown in Figure 11. For the parallel
channel the maximum distance is around 120 metres (400 feet). For ESCON
point-to-point, the maximum distance is three km. For the switched
point-to-point configuration shown, the ESCON Director can be placed at up to
three km from the processor, and each control unit can then also be placed at
up to three km. This gives a total separation between processor and control unit
of six km.
The normal light-transmitting component used is the light emitting diode (LED).
Another transmitter using a laser and operating over nine micron fiber is
This compares very favorably with the traditional parallel channel′s maximum of
400 feet. In summary, the distance and speed attributes alone of ESCON open a
completely new range of possibilities when designing for both continuous
availability and also disaster recovery.
A second definition for slow speed parallel devices exists for connection through
an ESCON converter. The byte multiplexor (CBY) definition supports devices that
can only work in byte multiplexer or burst mode. This permits the connection of
several slow speed devices to operate at the same time.
The aim is continuous availability, or to be more precise, in this case the aim is
continuous operation - the reduction of scheduled outages. It is possible to
replace or upgrade the subsystem attached to the converter without a
closedown. Any recabling work can be performed without the need for a service
interruption. This work has been possible for several years, but the final step of
introducing the change to the channel subsystem microcode and the operating
system has always required a service closedown.
HCD has a graphics capability to display the new configuration from several
viewpoints. It creates a block diagram of the newly defined configuration paths
and allows a simple comparison to be made. We strongly recommend that you
use this feature after each change in the channel subsystem definition. It helps
to eliminate potential human error or misunderstandings between departments
before activation of the new definitions.
Figure 12 shows a connection to a control unit. EMIF also supports the sharing
of an ESCON CTC channel between LPARs. Figure 13 shows the same four
LPAR configuration, and the number of CTC links required to support an
any-to-any connection.
Figure 13 shows that 12 channels are required to form this configuration. For
the mathematically inclined, the formula is n(n-1), where n is the number of
High availability design requires redundancy for each path. Without EMIF, 24
paths would be required for a high availability design. EMIF reduces this to only
four.
The small increments for channel upgrades and the variation of I/O cards
available tend to make each 9672 configuration unique, but by looking at the
hierarchy within the channel subsystem, it is possible to offer some rules for
configuring for higher availability.
The channels for the 9672 processors are arranged in groups called I/O cages.
Figure 14 on page 42 shows a 9672 with two I/O cages labelled I/O Cage 1 and
I/O Cage 2.
These labels have been chosen for this example only; refer to the documentation
for your machine for the labeling associated with your hardware. The number of
I/O cages and the number and type of cards installed in your machine depends
on which features were ordered.
ESCON channels, parallel channels, Open Systems Adapter (OSA), and coupling
links are all connected through this arrangement.
The fine detail of the internal connections is not important at this point. You can
see that the channel subsystem has a hierarchy within each I/O cage.
Each cage consists of one to three Fast Internal Bus (FIB) Cards, which are
connected to the processors. This connection is also known as the Self Timed
Interface (STI).
Each FIB card supports one or two channel driver cards (CHA), and each CHA
card supports one to four channel cards.
This grouping of an FIB card with its CHA cards and channel cards is called an
I/O domain.
The number of domains and cages varies according to the features installed on
your machine; the configuration shown here is for guidance only.
Each parallel channel card supports three channels. An OSA-2 card supports
one OSA channel. A coupling link card supports two coupling channels.
(Coupling links do not connect through the CHA card, but connect directly to the
FIB card.)
The STI connections are presented to the processors through the memory bus
adapter components. There are two MBAs, numbered MBA 0 and MBA 1.
There is an affinity between MBA 0 and the even-numbered STIs, and an affinity
between MBA 1 and the odd STIs.
This relationship is important to fully use the Partial I/O Restart feature of the G4
CMOS processor. Partial I/O Restart allows the system to restart and process
following the loss of an MBA. This means that multiple paths to the same
control units must be balanced between MBAs, or more accurately odd and even
STIs, for best connectivity to critical devices.
Figure 15 on page 44 shows more detail about the internal connections from the
FIB card down to the channel card for an eight card group. The internal
connections between the channel cards and their CHA “mother” cards is not a
simple pattern. Care must be taken to avoid placing multiple paths within the
same failure boundary.
When your machine order is processed, a report called the CHPID Report maps
out the exact layout of your machine. This report describes the relationship
between the FIB, CHA, and channel cards in your machine.
Note: Channel cards 1, 2, and 3 are labelled CC instead of CH. This is to
indicate that this position can be used for either a channel card or a coupling
link.
• In the case of a hardware failure, any spare port can be re-assigned as the
failing port.
Simply swap the cable from the failing port to the new port and the path can be
recovered by reassigning port numbers. This temporary reassignment of ports
is performed through the ESCON Director console. The port hardware can then
also be maintained concurrently, and when convenient, the cable can be
returned and the temporary reassignment of ports can be removed.
If the affected path is one of a group connecting to the same control unit, then
performance may be affected during re-assignment actions, but no break in
service should be seen by the users.
Each path to a control unit can be totally open for all to access, or it can be
restricted to a subset of channels only. This subset could be only a single
channel on a single processor.
The paths to the control unit are very effectively shared by the dynamic
switching action of the ESCON Director, but they can also be manipulated and
regulated in a variety of ways. Figure 16 shows a typical connection scenario.
The figure shows two processors, each with a potential connection to both
control units. The ports used by the connections are labelled A, B, C, and D.
Although each machine is physically connected into the switch, a connection
from a processor to a control unit across the switch must be allowed within the
switch port map (or matrix).
Where a port is allowed to switch is determined by the attributes set on both the
input and the output port. The attributes are set either by HCD definition, or from
the ESCON Director console.
The default is to allow any port to switch to any other (except itself). This means
that the channel from Processor 1 (ESCD Port A) is allowed to connect to either
Control Unit 1 (ESCD Port D) or Control Unit 2 (ESCD Port C). Processor 2 (ESCD
Port B) has the same access unless the defaults are overwritten.
The attribute set on any port can be changed dynamically when required.
This means that by connecting all processors to all directors, a fully redundant
physical configuration can be prepared, yet this configuration can be controlled
at a very fine level. For example, it is possible to prevent backup processors
from accessing critical devices by simply changing the attribute set on the
ESCON Director ports. Full connectivity can be granted by simply reversing this.
This allows very fast reconfiguration when trying to recover from a failure, or
when scheduling a workload move to allow a processor to be closed down.
Note: Reducing this interruption time while changing configuration improves
service recovery time and overall availability time.
The use of standard LED port cards allows up to three km between the ESCON
Directors; this may be suitable for campus environments.
The use of an XDF card allows much larger distance between processor and its
peripherals.
This now allows the establishment of a second processing facility for backup and
disaster recovery, at a reasonably safe distance, while still allowing all control
units to be connected to all processors.
The foundation can now be laid for a good continuous availability design that can
tolerate the loss of a complete site. Figure 18 on page 50 shows a very simple
example of such a design.
Both types of converters can be used in conjunction with the 9032 or 9033
ESCON Director, but some limitations are placed on this use. For example, an
ESCD port connected to a 9034 must be dedicated to a channel port in order to
function. No dynamic switching of the channel connection is possible. Also, the
maximum distance at which devices can be situated when using converters may
Not all models may be available in all countries; check the status accordingly.
Multiple 9036 units can be placed in a single path, but the minimum is two model
001s. The 9036 can also be used in conjunction with a 9032 or 9033. A maximum
of three 9036 ESCON Remote Channel Extenders may be present in any
end-to-end link, in addition to the two ESCON Directors currently available in a
link.
The 9036 can also extend the PPRC distance to 43 km. Refer to the Remote
Copy Administrators Guide and Reference for further information in this area.
This would appear to be an extreme single point of failure to introduce into the
configuration compared with individual links, but a backup fiber can be installed
and dynamic switching can occur if the first link fails. This enables some degree
of backup in the interests of high availability.
The ports within an ESCON Director are installed on a card with four ports on
each card. These ports can be either LED or XDF, but not mixed on the same
card.
Let us call a port attached to an ESCON channel an input port, and a port
attached to a control unit an output port. Any director port card could contain:
• Four input ports
• Four output ports
• A mixture of both
But which arrangement is best? The following sections answer this question.
For high availability, paths to the same subsystem are normally distributed
across as many failure boundaries as possible. With the ESCON Director′ s
switching capability however, a single input port may be allowed to connect to
many output ports. Also, a single output port may be allowed to connect to
many input ports. Mapping the extents of a failure boundary is complex, and
designing a configuration to minimize impact for concurrent maintenance is now
also made more complicated.
If a port card is configured to carry four input ports, then replacement of the card
requires all four channels to be stopped while the replacement is carried out.
This may affect many paths to many different control units. These paths may
also be spread across up to four different processors, each with several LPARs
sharing each path. Scheduling the closedown of all paths is not trivial. If a port
card is configured to carry four output ports, then replacement of the card means
a similar exercise.
If the port card is configured to carry both input and output ports from the same
path, then the impact of replacing this card is minimized. Figure 19 on page 53
shows this situation.
This configuration shows two processors each sharing two control units with four
paths. Each input path from each processor is connected to a different port card,
and each connection to each control unit also sits on a different port card. This
is good separation from the view of either the processor or the control unit.
However, each input and its associated outputs are grouped on the same port
card. If any path is broken, its associated inputs or outputs are affected anyway,
so now the number of additional paths required to be closed for the repair action
is reduced.
There is no hard rule that says that the dedicated link must be placed on the first
or second director. The choice is best made based on what connectivity is
required from the channel.
Placing the dedicated link in the first director allows devices attached to the
second director to be shared. This is an effective way of connecting devices in a
second site to a single processor in the first site. No connection to the second
site can access any device at the first site, so good separation of the
environments is maintained.
Placing the dedicated link in the second director allows the link to the second
director to be shared. This is good for sharing links to a second site between
multiple processors at the first site (reduces the number of inter-site fibers), or
for sharing a channel path between both sites.
The channel connection in this case could also be shared with 3990 Peer-to-Peer
Remote Copy (PPRC) between compatible 3990s at each site. In the lower
diagram of Figure 21, if control units 1 and 2 are both 3990 model 6s, then by
This is a very useful facility to build into your operation. It makes it much easier
to co-ordinate system definition changes and any switch matrix changes that are
associated with them. It is strongly recommended to have this facility on all
attached processors in addition to the ESCON Director Console. This
combination makes a loss of ESCON Director management very unlikely.
When chained directors are used, the only way that a processor can control the
second director in the chain is to have a dedicated link in the first director. A
dedicated link in the second director cannot dynamically switch and connect to
the internal control unit.
3.8 Management
When considering a hardware configuration to support continuous availability, it
soon becomes obvious that it probably becomes more complex than those built
simply for service delivery from a single processor.
It has been shown earlier in this chapter that HCD is the main tool to build and
dynamically change the ESCON environment, but in day-to-day operation,
another tool is required. ESCON Manager was developed to perform this task.
3.8.1.1 History
ESCON Manager was originally introduced as a product in its own right
(5688-008) and developed through three releases. It has now been incorporated
into the System Automation for OS/390 product (5645-005). The function
described applies to both the final release of &escm, and to the component of
System Automation for OS/390 .
3.8.1.2 Function
ESCON Manager works dynamically with all OS/390 systems, HCD, the channel
subsystem microcode, and the installed hardware to build and maintain a picture
of the current configuration and other legitimate connections. It also works with
other ESCON Manager components on other OS/390 images to obtain their view
of the same hardware components. This view can be filtered and manipulated
by system operators using menus supplied with the product. ESCON Manager
performs these tasks, regardless of whether the component is attached to a
parallel channel or to an ESCON channel. It can also discover and display
information on other hardware items associated with a Parallel Sysplex (coupling
facility and links), and the Open Systems Adapter (OSA) used on S/390 systems.
When the status of an individual item of hardware has been determined, ESCON
Manager can also perform any change (vary offline or online) decided by the
operator, and then schedule this change across all sharing systems.
The scope of this vary command can also be limited, if required. It contains a
very sophisticated checking system to ensure that the command is valid and
legal for the device, and also a backout facility to restore the status of the device
if anything prevents the command from completing successfully. In addition, it
can also be used to check the switch port connections and attributes in an
ESCON Director, and then change them as required.
With its (multi-system) dynamic search and control ability, ESCON Manager is
uniquely placed as a tool to aid operator knowledge and decision-making.
Better operator tools and aids help achieve the target of high availability, by
avoiding unscheduled outages. ESCON Manager can also influence continuous
operation by managing the efficient reconfiguration of hardware paths for
workload switches between processors, or similar activity. ESCON Manager
therefore has an extremely important role to play in the goal of continuous
availability.
S/390 Parallel Sysplex continues to provide the basis for OS/390 availability
improvements. OS/390 builds on the high availability inherent in a sysplex
because its functions have been tested in an integrated fashion.
XCF Coupling Facility Failure Policy: This policy is enhanced to optimize the
selection of a coupling facility during structure allocation. It also gives
applications a way to define their connectivity requirements during an initial
connect. The rebuild process is also enhanced to guarantee that the rebuilt
structure has better connectivity to the systems in the sysplex than the old
structure.
XES lock support: Locking is enhanced to support the GRS star topology.
ICKDSF R16: Release 16 Refresh with Peer-to-Peer Remote Copy adds a new
command, PPRCOPY, to ICKDSF. On those subsystems that have Peer-to-Peer
Remote Copy, PPRCOPY supports synchronous copying of a DASD volume from
one subsystem to another subsystem volume, without host system
dependencies. Use the PPRCOPY command to:
• Establish ESCON paths between two subsystems
• Establish remote copy device pairs
UNIX 95 Certified: In order to achieve full UNIX 95 branding from the X/Open
Company, OS/390 has added new programming and shell interfaces. In addition,
changes have been made to improve performance.
DCE Distributed File Service (DFS): With the DFS client, DFS becomes an
integral part of OS/390 OpenEdition DCE, offering a seamless data repository.
Distributed data with DFS confers such benefits as consistency among
distributed files, improved response time and data availability, access to the
OpenEdition Hierarchical interrupting end users, and directory and security
services that are integrated with OpenEdition DCE.
The DFS client communicates with file server machines to access files for
application programs and to provide local data storage. Also called the cache
manager, the DFS client translates file requests made by application programs
on a client machine into remote procedure calls (RPCs) to file exporter
processes on file server machines. The cache manager caches all data it
requests by storing it on disk or memory before passing it on to the requestor.
In addition, DFS ensures that the cache manager always has access to the most
current copy of the data. If the data changes, the cache manager retrieves the
most current version, automatically and totally transparently to the user.
DCE Network File System (NFS): In addition, for OS/390 Release 2, OpenEdition
DFS provides an NFS-to-DFS gateway. The gateway will allow NFS clients
access to the MVS system by using the MVSLOGIN command and RACF security
services. Users of NFS clients, once authenticated by enrolling on an MVS
machine, can access data in the DFS filespace.
DFSMS/MVS Network File System V1R3 (NFS): A new Network File System client
is provided by the DFSMS/MVS Network File System on the OS/390 platform.
Chapter 4. OS/390 61
The NFS client program on the client system maps local file system calls into
network calls and sends a request for data or control information to the NFS
server. In this way, NFS clients and servers can deliver remote data to an
application without special interfaces in the application. The client program
supports the SUN NFS Version 2 protocols. It uses RPC requests to
communicate with remote NFS servers over a TCP/IP network.
Access Method Support For OpenEdition MVS Files: Many applications written to
use the BSAM/QSAM access methods, and some applications written to use the
VSAM ESDS access method, now have transparent access to files associated
with OpenEdition. With this support, the conventional MVS application, using the
normal record or block interfaces provided by the MVS access methods, can
access files residing on remote NFS server platforms.
TSO sysplex support for VTAM Generic Resource, which distributes TSO logons
to available capacity in a coupled systems environment, is provided.
DB2 TCP/IP support is available, which provides the ability to spread work
across a DB2 datasharing group using the WLM recommendation given in the
sysplex routing selection service.
UNIX services: The OpenEdition kernel starts at IPL so that UNIX C services are
always available. Users no longer need to write code to deal with kernel
termination and restart, simplifying server application programming. Additional
enhancements include increased NLS support. Also included is Web thread
support, allowing a server to dynamically change its environment without having
to be restarted.
Chapter 4. OS/390 63
• DFS for OS/390 can now be configured using ISPF dialogs. This improvement
will provide ISPF panel-driven configuration and unconfiguration that users
are accustomed to from DCE base addition services.
Internet Connection Secure Server (ICSS) MVS dataset support: With this
feature, World Wide Web documents may be retrieved from native MVS datasets
as well as from the MVS/Open Edition Hierarchical File System.
Communication server:
Security server:
The security server includes RACF and the OpenEdition DCE security server.
• It allows a single sign-on to the RACF and DCE domain
• It gives enterprise-wide security
• It enables porting of applications that use DCE security services to OS/390
• It allows DCE application services to associate a DCE user with a
corresponding RACF-defined user
Businesses increasingly wish to use the Internet to market products and conduct
business with suppliers and customers. The IBM Internet Connection Secure
Server for OS/390 enables the use of S/390 as a Web server.
The benefits of the IBM Internet Connection Secure Server for OS/390 are that
you have S/390 security, the utilization of large storage capacity on S/390, the
use of centralized skills, a single point of entry and control, consolidation of
multiple Web sites, and secure Internet transactions.
Chapter 4. OS/390 65
OS/390 is the base for future enhancements to S/390 software, and in
combination with IBM′s new CMOS parallel enterprise servers, provides you with
the opportunity to improve your continuous availability.
The environment simulates between 8500 and 12000 end users and performs
between 400-to-520 transactions per second with an internal response time goal
of 80% of CICS transactions completed in less than 0.6 seconds. There are 108
CICS regions, 27 IMS regions and nine DB2 regions.
Each of the integrated elements and optional features of OS/390 have been put
through a traditional product development test cycle. This ensures that each
element and feature, when considered on its own, provides the function defined
in its specification. In addition, the elements and features are combined
(integrated) at one time and run in the simulated production environment. This
test ensures that the elements and features can be installed and run together
and have the robustness that is characteristic of the MVS platform. Particular
attention is paid to the following in a Parallel Sysplex environment:
• Continuous operation
• Recovery from hardware and software failures
• High transaction rates
• High CPU utilization
The focus of this testing is Parallel Sysplex datasharing with workloads of the
following type:
• CICS with DL/I
• CICS with DB2
• CICS VSAM RLS (record level sharing)
• DB2 batch
• IMS BMP
• IMS DL/I batch
• IMS TM with DL/I
• IMS TM with DB2
All of the functions necessary for Parallel Sysplex have been tested in a Parallel
Sysplex environment.
Smaller enterprises, with a single system and just a few programmers, will find
that they too can keep current with new functions delivered in OS/390 releases
with far less work than before because of OS/390 systems integration testing of
each release.
Additionally, OS/390 Release 3, like Release 2, offers customers two new ways to
test applications on the system, particularly applications changed for the year
2000.
Customer feedback has indicated that the systems integration testing can save
them several months in moving to a parallel environment.
Chapter 4. OS/390 67
Figure 22. Operating System Co-Existence within a Sysplex, N, N + 1 , N + 2
Subsystems such as DB2, CICS, and IMS will allow two consecutive releases to
coexist.
While it is expected that most installations will migrate to all elements of OS/390
at the same time, the migration of JES2 and JES3 can be staged. The OS/390
levels of JES2 and JES3 are contained in their own libraries and, beginning with
Release 3, in their own SMP/E zones. This makes it possible for you to use your
current copies of JES2 and JES3 more easily, if you so choose. For full details,
refer to OS/390 Planning for Installation Release 3 .
OS/390 provides the base for the terms and conditions of a Parallel Sysplex
environment. Its functions have been tested in a sysplex environment. It is
estimated that with OS/390, moving to a parallel environment can be done
significantly sooner than it can be done using a piecemeal approach. You also
have the ability to clone, allowing your applications to remain available to your
end users. Changing a release causes an IPL of only one system; the others
remain available to your users and applications.
In this example, we have a Parallel Sysplex that consists of four MVS images,
SYS1, SYS2, SYS3, and SYS4, using a coupling facility.
• SYS1 is at operating system Level N
• SYS2 is at operating system Level N
• SYS3 is at operating system Level N+1
• SYS4 is the test system at operating system Level N+2
SYS4 has been removed from the sysplex for a system software upgrade. For
the purposes of our example, it is at OS/390 V1R3 (or N+2, as relative to the
lowest operating system level installed in the sysplex).
Figure 23 shows the steps you need to follow when you want to propigate
production to this newer level of OS/390.
1. Add the new system to the Parallel Sysplex as a participating system and
not a test system.
2. Drop SYS1 from the sysplex, keeping SYS4 (the former test OS/390 system)
as a production system.
Chapter 4. OS/390 69
3. IPL SYS1 with the OS/390 N + 1 or N + 2 level of code and re-connect to
sysplex.
This roll-out process is conceptual. Whether you are running OS/390 or current
levels of the stand-alone products, you might have to perform certain migration
actions as you upgrade to OS/390 R3. For migration information, refer to OS/390
Planning for Installation V1R3, GC28-1726 .
4.1.3 SmartBatch
IBM SmartBatch for OS/390 is a major enhancement to, and a replacement for,
BatchPipes/MVS. It maintains the single system data piping capabilities of
BatchPipes/MVS and provides the following enhancements.
4.1.3.2 BatchPipePlex
BatchPipePlex extends the data piping capabilities of BatchPipes/MVS to batch
jobs operating in a Parallel Sysplex environment. By using the coupling facility,
BatchPipePlex can enable parallelism across OS/390 and MVS/ESA images and
provide an opportunity to distribute a batch workload to improve resource
utilization.
Use of BatchPipePlex support will help some installations to move from the large
water-cooled mainframe to the less expensive CMOS mainframes. Since
BatchPipePlex support provides installations with the option of spreading their
batch pipelines across a Parallel Sysplex, they can use the smaller, less
expensive CMOS mainframes to meet their data processing needs.
In cases where an installation has a large pipeline of jobs that could execute
concurrently, but the installation is unable to schedule the sequence of jobs all
together on a single system because of resource constraints, the installation can
establish a BatchPipePlex and spread the pipeline of jobs across more than one
system.
Data Accelerator provides two levels of I/O performance processing, basic and
advanced.
At the basic level, Data Accelerator automatically determines and sets the
optimum buffering values for the utility or application.
Chapter 4. OS/390 71
• Providing larger data transfers per I/O operation
• Learning an application′s access patterns to restructure I/O requests
Data Accelerator also provides I/O performance options to control how I/O
performance processing occurs. The Resource Usage Bias option controls how
Data Accelerator obtains and uses resources to provide performance benefits.
The Dynamic Region Adjustment option allows Data Accelerator to automatically
modify region specifications to compensate for additional I/O-related storage
areas obtained on the application′s behalf. These and other performance
options can be used selectively to fine-tune application processing.
These Data Accelerator functions do not require any application or JCL changes.
These enhancements add function to allow IBM SmartBatch for OS/390 users to
optionally synchronize the allocation and termination of their jobs, to ensure that
IBM SmartBatch for OS/390 propagates errors to all jobsteps that execute as a
unit of work. Users can specify the new subparameters, AllocSync, AllocNow,
and TermSync, on the JCL SUBSYS DD statement in their input jobstreams.
IBM SmartBatch for OS/390 is one of several software and hardware functions
that IBM offers to improve the elapsed time of batch jobs that run on OS/390 and
MVS/ESA. Others include:
• VIO to expanded storage
• Hiperbatch
• Sequential data striping
Both VIO to expanded storage and the IBM SmartBatch for OS/390 BatchPipes
component eliminate the I/O wait times that occur when a job waits for I/O
transfers to and from storage devices. In addition, BatchPipes allows the writer
However, VIO supports both random and sequential access to data, while
BatchPipes supports only sequential access.
4.1.3.6 Hiperbatch
Hiperbatch targets a different type of batch application from the IBM SmartBatch
for OS/390 BatchPipes component. With Hiperbatch, the target application has
many readers simultaneously accessing data from an already existing dataset.
The shared data all resides in processor storage and on DASD; the data persists
after being read. In contrast, for BatchPipes, the target application has one
writer passing data to one reader. BatchPipes holds very little of the data in
processor storage at one time, and the output from the writer job does not
reside on tape or DASD.
With Hiperbatch, readers run simultaneously with each other, but not with the
writer; with BatchPipes, writer and reader jobs run simultaneously.
Sequential data striping substantially improves I/O access times to DASD. For
sequential data sets, BatchPipes can eliminate I/O and enable concurrent
processing.
Data accelerator can be used for data sets being accessed by sequential data
striping for possible additional performance improvements.
For further information, refer to IBM SmartBatch for OS/390 User ′ s Guide ,
GC28-1640. This book is softcopy only. You can find it on the OS/390 V2R6
online library, SK2T-6700.
4.1.4 DFSMS
DFSMSdss is a component of DFSMS/MVS which works with DFSMShsm to
provide the following availability management functions. Applications that
demand total availability need advanced capabilities like these:
Chapter 4. OS/390 73
4.1.4.2 S/390 Remote Copy
S/390 has two remote copy functions that will write data to a remote secondary
disk as it is written to a local primary disk. If a failure occurs on the primary
disk, up-to-the-minute data can be recovered from the copy. That copy can be
many miles away, allowing you to have a remote site for disaster recovery
purposes. These solutions provide backup copies with no loss of service.
• HSM CDS in CF
• RMM/BTLS
The following describes products and functions that could utilize RRMS to extend
transaction processing on OS/390.
Where existing distributed processing capabilities do not exist, these new system
services enable OS/390′s existing OLTP information systems, IMS and CICS, to
extend their distributed processing capabilities such that clients and distributed
servers have transactional access to the OS/390 IMS/ESA Transaction Server
and the OS/390 CICS Transaction Server through the protocol that is appropriate
Chapter 4. OS/390 75
for the distributed environment. Multiple client/server platforms have distributed
transactional access to new or existing OS/390 transaction and database servers
through one of the following distributed communication protocols:
• APPC/MVS support for SNA′s LU6.2 syncpoint architecture (introduced in
OS/390 Release 3)
• Transactional RPC support on OS/390
• Object Transaction Service based on OMG′s CORBA specification
• Message and Queuing based on MQSeries for MVS/ESA
Resource managers like DB2, MQSeries for MVS/ESA, transaction monitors like
IMS and CICS, and multiple communications protocols like APPC_PC,
Transactional RPC, and OTS can use these services to provide new transactional
applications, as well as combinations of new and existing transactional
applications. These capabilities allow new applications to take advantage of
existing investments in business applications, new application technologies,
distributed protocols of choice, the Parallel Sysplex, and the strengths of OS/390
for scalability, reliability, availability, security, integrity and the system
management of a distributed enterprise environment.
In the OS/390 Release 3 timeframe, DB2 provided a new OS/390 Attach Facility
that uses the Recoverable Resource Management Services. This enables DB2
to participate in commit scopes with other participating resource managers for
most OS/390 application environments. DB2 Stored Procedures also uses the
Recoverable Resource Management Services to participate in commit scopes
with other participating resource managers, as well as the commit scope of
other transaction servers.
Transactional RPC for OS/390 IMS/ESA Transaction Server can also be provided
for distributed transactions by an OS/390 Encina Toolkit Executive using
Recoverable Resource Management Services. Through similar extensions, DCE
servers and any OS/390 participating resource manager or transaction monitor
are able to participate in TRPC-initiated transactions. The distributed transaction
processing capabilities of OS/390 CICS Transaction Server can also be extended
to support transactional RPCs. This allows clients on heterogeneous platforms
and workstations access to the existing CICS and IMS transaction processing
environments for new and existing applications using TRPC. Since IMS and
CICS can participate as resource managers, this extends the transactional RPC
scope to existing transaction applications and potentially any of their databases
or resource managers currently accessible by them.
Object Transaction Service uses these same base system services to provide
distributed transaction capabilities for OS/390 object servers. The object servers
can include both OO resource managers as well as procedural resource
The integration of WLM and the Parallel Sysplex with DB2, OS/390 IMS/ESA
Transaction Server, OS/390 CICS Transaction Server, OS/390 object servers,
OS/390 application servers and RRMS, together with the work routing
capabilities of APPC/MVS, DCE, TCP/IP and VTAM, allows installations more
control and management of their business applications and the total resources of
the sysplex, consistent with the priorities and policies of their business
objectives.
In summary, these capabilities will, over time, allow new and existing
applications to take advantage of existing investments in business applications,
new application technologies, distributed protocols of choice, the Parallel
Sysplex, and the strengths of OS/390 for scalability, reliability, availability,
security, integrity and the system management of a distributed enterprise
environment for their transaction processing requirements.
4.2 Summary
The future for OS/390 is continued expoitation of Parallel Sysplex and ever
increasing improvements in continuous availability. The future for OS/390 is
more integration, more function, and more synergy between your applications.
OS/390 is a growing platform. It gives you continuous availability, with capacity
you cannot outgrow, at low incremental cost. Strategic initiatives include base
business enhancements in systems management and security, more and more
application enablement across secure networks, whether Internet or intranet,
and server consolidation, which means integrating smaller and multiple servers
such as file serving and print serving into OS/390. These initiatives ensure that
the goals you have set for your business are directly supported by the
technology you have installed.
Chapter 4. OS/390 77
78 Continuous Availability S/390 Technology Guide
Chapter 5. OS/390 Open Edition
S/390 is well known as having the right attributes to serve the enterprise, but it
is perceived to cost much more than alternative (UNIX) platforms. However, a
single alternate platform UNIX server for a single department running one
application is a low-cost platform, but to perform as an alternative to a single
S/390, a UNIX system must be able to:
• Handle multiple diverse workloads
• Provide high availability (99.9%)
• Give fast consistent response times (<3 seconds)
• Handle large amounts of data (GB to multi-TB databases)
With S/390 and OS/390, these attributes are builtin. In contrast, alternate
platform UNIX systems must supplement their basic configuration significantly,
which results in a more complex and expensive configuration that is difficult to
manage.
Most alternative platforms perform well with decision support workloads which
are reading data intensively. However, for mission-critical workloads like online
transaction processing (OLTP), fast processing, fast I/O, large buffers, and the
ability to maintain data integrity when multiple users request the same data
concurrently, a single S/390 performs best. To provide the same service,
alternative platforms must duplicate processors and DASD, increasing the
complexity and cost of their solution.
You can use TSO/E commands or shell commands. to work with the HFS. If you
have access to ISPF, you can use the panel interface of the OpenMVS ISPF shell.
For example, you can create a directory with the TSO/E command mkdir, or with
the shell
mkdir
.
Furthermore, you can issue TSO/E commands from the shell command line, from
a shell script, or from a program. Reference OS/390 OpenEdition User ′ s Guide ,
SC28-1891 for a list of TSO/E commands you can use to work with the file
system.
You can write MVS job control language (JCL) that includes shell commands.
You can use the OSHELL REXX exec to run a non-interactive shell command or
shell script from the TSO/E READY prompt and display the output to your
terminal. This exec is shipped with OS/390 OpenEdition MVS.
Programs access the information in HFS files through OpenEdition system calls
instead of through standard MVS access methods and macro instructions. Data
files are identified and accessed by specifying the path leading to them.
HFS data sets appear to the MVS system much as a PDSE does, but the internal
structure is entirely different. DFSMS/MVS manages HFS data sets, providing
access to the files of data within them as well as backup/recovery and
migration/recall capability.
You can back up the datasets while open. You do not have to shut down the
OpenEdition address space. However, the same rules apply to backing up HFS
files as for backing up anything else; if they are open, they could be becoming
updated just before or just after they are copied for backup.
When DSS is the data mover in HSM, HSM can also back up HFS data sets;
however it backs them up successfully only if they are not open.
To determine when the HFS file is open, since the interface is through
OpenEdition, start an OpenEdition shell session and issue the df command. It
displays information about all mounted files and the relative MVS dataset names.
For example:
#df
/u/gronenb (OMVS.GRONENB.HFS) 14336/14400
/u/mmcdavi (OMVS.MMCDAVI.HFS) 14288/14400
/u/linetti (OMVS.LINETTI.HFS) 14360/14400
/u/vanaers (OMVS.VANAERS.HFS) 14336/14400
/u/newling (OMVS.NEWLING.HFS) 14320/14400
/ (OMVS.ROOT) 184/37440
The OMVS.** datasets are the MVS files that are HFS datasets. HFS files can be
mounted dynamically in OpenEdition by any authorized user, so you need to
check in a running system. This information is documented in more detail in the
OpenEdition manuals.
5.1.3.2 Exchangability
With OS/390 OpenEdition services, workstation users can copy data from MVS
data sets into hierarchical files for use at the host or, through Network File
System server support, at the workstation. Users can also upload
ASCII-encoded hierarchical files to the host, where they can be translated into
EBCDIC encoding for the host. These files can be stored or processed either as
hierarchical files, or copied to an MVS sequential data set or partitioned data set
member.
In OS/390, a hierarchical file system (HFS) is treated as a new kind of data set.
Like MVS data sets, HFS data sets can be allocated, backed up, restored, and
deleted.
In the event that the HFS is exposed to a power failure, it is designed to maintain
its data structure. You can lose recent data that is still buffered, but the file
system structures, directories, inodes and such will not be damaged. There is a
shadow writing technique that is used to ensure structural changes are always
committed atomically. The HFS does its own repair, as needed, on each mount
of a file system. This is based on records it keeps of changes in progress.
The same application program can access both traditional MVS data sets and
the hierarchical files within HFS data sets. In addition to the C language
interface, a callable service interface is provided for invocation from Assembler
H applications and subroutines.
Record Data Access: In addition to the OE HFS and the DCE Local File System
(DCE LFS) byte file data that DFS for OS/390 previously supported, DFS for
OS/390 Release 3 now supports DFS Client Access (TM) to the Sequential,
VSAM, PDS, and PDSE data sets that are allocated to direct access storage and
cataloged using the Integrated Catalog Facility (ICF).
DFS support refers to the Sequential, VSAM, PDS, and PDSE data sets as record
data sets, and to the data in these data sets as record data . DFS for OS/390
Release 3 can export a record data set for access through DFS client support on
workstations or other S/390 systems.
Network access to record data sets through DFS for OS/390 has the benefit of
Distributed Computing Environment (DCE) security based on Kerberos, and DFS
client caching for improved performance. The record data is presented to the
DFS client as byte stream data. EBCDIC to/from ASCII translation is also
supported for single byte code page (SBCS) file data.
DFS server machines run processes that provide services such as making data
available and, on an administrative level, monitoring and controlling other
processes. Server machines are defined by the processes/daemons they run.
Among the processes DFS server machines run are these:
• The File Exporter (DFSKERN) fills requests for data from client machines
anywhere in the network.
• DFSXPORT determines the data to be made available to DFS clients
anywhere in the network.
• The Basic OverSeer Server (BOSERVER) monitors the DFS processes that
run on a server and restarts them when needed.
• The Backup Tape Coordinator (BUTCnn) controls a backup tape drive and
accepts service requests for backup.
• The Replication Server (RPSERVER) copies read-only replicas of read/write
data onto other file server machines.
• The System Control Server (UPSERVER) updates other server machines with
identical versions of system configuration files.
• The System Control Client (UPCLIENT) retrieves system configuration files
from the system control server.
• The Fileset Server (FTSERVER) provides an interface to the DFS commands
and components used to manipulate filesets.
• The Backup Database Server (BAKSERVER) houses the master and replica
versions of the Backup Database, where information used to back up and
restore system and user files resides.
• The Fileset Location Server (FLSERVER) houses the master and replica
versions of the Fileset Location Database (FLDB), where information about
the location of systems and user files is maintained.
When the Cache Manager receives a packet of data for an HFS or DCE LFS file
from a file server machine, it stores copies of the data in a local storage area
called the cache. The cache is an area reserved for data storage on the local
disk or in memory on the client machine. The Cache Manager uses the local
copy of the cached file; it does not continue to send network requests to file
server machines for data it has stored locally. Access to locally stored data is
much quicker than access to data stored across the network.
As you save the changes you make to data, the Cache Manager periodically
sends the changed data back to the appropriate file server machine, where your
changed version replaces the data stored on the server. When the central
version of the file stored on the file server machine changes, DFS advises all
other Cache Managers with copies of the file that their versions are no longer
An OS/390 user can get access to DCE from either the TSO or the UNIX
environments in OS/390. Access to the UNIX environment is provided either
directly by using a Telnet session, or through TSO.
Since the UNIX environment in OS/390 provides all the native DCE tools and
utility programs, the UNIX environment is the best place to function for DCE
application developers and DCE administrators.
You can do the same thing with the hierarchical file system (HFS). However,
since the individual HFS data sets that make up the active file hierarchy must be
on SMS-managed volumes, there are some special considerations for making a
transportable copy:
1. Dump each HFS data set to be transported into individual sequential data
sets using the DFSMSdss dump utility. These sequential data sets contain
all the necessary information about the files and can also exist with other
product libraries that need to be transported.
2. After the system image has been transported to the target system, restore
individual product libraries or full volumes using the DFSMSdss restore
utility.
3. After the data sets have been unloaded on the target system, use the DFDSS
restore utility to restore the sequential data sets into individual HFS data
sets. These HFS data sets will make up the active file hierarchy on the
target system.
Refer to the following documentation for information on how to dump and restore
a hierarchical data set:
• DFSMS/MVS Planning for Installation
• DFSMS/MVS DFSMShsm Storage Administration Guide
• DFSMS/MVS DFSMSdss Storage Administration Reference
5.1.4 ADSM
ADSTAR Distributed Storage Manager (ADSM) offers two types of backup for
OpenEdition MVS clients:
1. Selective, in which the user backups up specific files.
2. Incremental, in which all new or changed files are backed up.
For a business that needs an extra measure of security, the Internet Connection
Secure Server for OS/390 can be run in a PR/SM logical partition environment to
separate public network connections from a private customer environment.
For example, a user who wants to connect to the Internet could run the Internet
Connection Server, their Web server and their corporate information systems on
the same physical system with assurance that the workloads will remain
separate. PR/SM on ES/9000 processors (9021 and 9121) achieved an E4 rating
against the Information Technology Security Evaluation Criteria (ITSEC), the
highest achieved by any vendor. There is no concept like PR/SM in any
RISC-based server. Therefore there is no RISC platform that can be used for
combining workloads with different security classifications.
It also means that your CICS applications, even when used internally (no Internet
connection), can have a much improved user interface, using the Web browser
as an alternative to the 3270 interface. No CICS specific software is needed on
the end user′s system. Further information on this capability is available on the
Internet at the CICS Web page:
http://www.hursley.ibm.com/cics/internet/index.html
The CICS Gateway for Java enables Java applications to communicate with CICS
servers through a CICS Client. It is itself a Java application which provides an
API that enables conversations between an application or applet in Java and a
transactional application running in a CICS server. Further information on this
capability is available on the Internet at the CICS Web page:
http://www.hursley.ibm.com/cics/internet/index.html
The IMS World Wide Web Connection provides a connection from Web browsers
into IMS/ESA through ICSS for OS/390 or workstation web servers. Further
information on this capability is available on the Internet at the IMS home page:
http://www.software.ibm.com/data/ims
.
AIF applications do not have to be aware of HTML format and vice versa. All
HTML features, including any kind of graphics and animation, can easily be
combined with AIF-based S/390 business applications. Further information on
this capability is available on the Internet at the AIF home page:
http://www.s390.ibm.com/products/aif/aif.html
You can also customize it to your local system and run it whenever you suspect
problems.
Businesses can rest assured that as they evolve their network computing
strategies and seek to revitalize existing applications and deploy new ones
Additionally, IBM has announced the intention to enable ICSS for OS/390 to
participate in the S/390 Parallel Sysplex environment. Given this degree of
flexibility, Internet projects can start with minimal configurations and can be
scaled to handle thousands of users with demanding e-business workloads.
An organization can serve up one or many Web sites from a single copy of ICSS
for OS/390. This foundation product provides a high performance Web/proxy
serving solution that continues to be enhanced with added security,
management, monitoring, and application development features, along with
state-of-the-art Internet protocol support.
ICSS for OS/390 may be configured to operate as either a Web server or proxy
server. A proxy server, which makes requests on behalf of clients, can simplify
network design, provide additional privacy, and improve responsiveness by
caching frequently accessed information.
ICSS is also an essential base product when used in conjunction with IBM′ s
Net.Commerce for OS/390.
The Internet Connection Secure Server for OS/390 supports the latest industry
standard RSA public key cryptography algorithm in the form of Secure Sockets
Layer (SSL). This means that security can be applied to transactions to prevent
tampering and keep dialogs private when working over the Internet or intranets.
In addition, each party in the dialog can be authenticated, thus removing the fear
of conducting business with impostors. However, cryptography does have an
associated system cost. Each outgoing and in coming transaction that requires
this level of security must be encrypted and decrypted. This operation is
generally performed in software that will have an impact on processor
performance.
The S/390 Parallel Enterprise Server′s G3 & G4 and S/390 Multiprise 2000
servers address this challenge by performing cryptographic functions in an
exclusive specialized hardware function called the Cryptographic Coprocessor
feature. As a result, the main processor(s) is freed to concentrate on application
functions, and is not burdened with the encryption/decryption process. This new
capability, coupled with resource managers such as RACF or the OS/390 security
server, ensures that system resources are protected from unauthorized access,
making S/390 one the most secure environments in which to conduct electronic
commerce over the Internet.
System security starts with the integrity of the operating system. Resource
protection managers like RACF are used to protect the OS/390 system
authorization libraries from unauthorized access.
The Internet Connection Secure Server for OS/390 has extensive access control
support when the option to use Resource Access Control Facility (RACF) is
chosen to perform validation of user IDs and passwords and access control
through surrogate user ID support. RACF has proven to be the method of choice
over the use of basic server security.
However, it is possible for a hacker coming from a network that you do not
control (for example, the Internet), to forge the header information so that it
falsely identifies a trusted client. It is virtually impossible to detect when
someone impersonates another user unless you are using a resource protection
facility like RACF.
With the introduction of Internet Connection Secure Server for OS/390, Version 2,
Release 2, IBM is continuing to deliver on its promise and commitment to
provide customers with industry leading network computing technology and
solutions for the S/390 environment.
OS/390 Dataset Support for World Wide Web documents is provided as well as
OS/390 OpenEdition Hierarchical File System document support.
Additional OS/390 Console Support allows the server administrator to shut down
the ICSS in an orderly manner and use the OS/390 Modify Command to pass
requests to the server. An example of this is turning debug on or off.
HTTP support now includes full HTTP 1.1 compliance. Improvements in this
version include persistent connections and virtual hosts. Persistent connections
allow the server to accept multiple requests and to send responses over the
same TCP/IP connection. Virtual hosts allow the use of one IP address to serve
multiple files instead of having different IP addresses for different files.
Automatic browser detection allows the server to respond to requests with the
version of a Web page or document appropriate for the requesting browser.
Support for Year 2000 operations and content delivery is provided. Page counter
and date/time information can be displayed on a page as graphical images.
Logging and reporting are enhanced.
The previous version of ICSS for OS/390 provided standard reporting such as
number of hits and bytes transferred. The new Java Graphical User Interface
(GUI) version supports URLs, host names, method, or return code reports.
Additionally, associations between these report types can be generated, for
example, URL hits by host, host hits by URL, return codes by URL, and methods
by URL.
The Common Gateway Interface (CGI) support has been expanded to include the
Java programming language (in addition to the other languages such as C,
REXX, and Perl which are already supported).
Platform for Internet Content Selection (PICS) provides a way for webmasters to
provide ratings for the documents they own, allowing Internet users to filter
material based on those ratings.
Client Authentication provides the server with the capability to determine, with a
high degree of confidence, that the client is who it says it is. An example of how
this might be used is a medical clinic where a patient′s files could be viewed
only by physicians containing private patient data.
ICSS for OS/390 will now have a Simple Network Management Protocol (SNMP)
subagent which will maintain ICSS information and performance data in an
SNMP Management Information Base (MIB). Now, from any SNMP-capable
network manager, like NetView/AIX, you can display, monitor, and threshold on
your server performance.
SOCKS support provides the capability to use ICSS for OS/390, when located
behind a firewall, as a proxy server to the destination host via a SOCKS server
on the other side of the firewall. SSL tunneling allows the server to act as a
proxy server for secure transactions. Using a browser, you can observe the
status of the server through the Server Activity Monitor. Examples of the
categories of data you can view are the Thread Counts, Server Request
Statistics, Server Throughput Statistics and Connection Counts.
For more information about IBM′s Internet Connection Secure Server for OS/390,
visit the ICS web site at:
http://www.ics.raleigh.ibm.com/icserver
. You can also visit the S/390 web site at:
http://www.s390.ibm.com/nc/
.
S/390 has the technology and products to meet your business requirements.
Some of these capabilities include:
• A network-ready family of servers with direct ATM connectivity
• An integrated server-capable operating system with X/Open XPG4 UNIX
branding
• World-class hardware and software security
• Secure and scaleable Web serving
• Object technology and native Java support
This relationship will enable C and C++ programs written to the Windows NT
APIs (application programming interfaces) to be ported to OS/390. Customers
and independent solution developers will be able to leverage OS/390′s unique
scalable and robust strengths for their Windows NT applications.
The Berkeley paper hypothesized that a collection of small inexpensive (of low
$/MB, low capacity, low reliability, and low performance) devices could provide
improved performance relative to high-end devices (of high reliability, high
capacity, and high performance) by aggregating the data rates of the several
devices in an array. Today, “inexpensive” has been dropped from the RAID
acronym. The Berkeley authors changed it to “independent” in acknowledgment
of the need for highly reliable devices to meet the requirements of today′ s
customer environments.
RAID-1 delivers high availability because there is a second copy of the data that
can be accessed if the first copy cannot. The disadvantage of RAID-1 is the cost
of providing two copies of data, thus doubling the required disk space.
The access arms can move independently of one another. This array type is
intended to support multiple concurrent accesses to the array devices to satisfy
multiple concurrent I/O requests and thus provide high transaction throughput.
Data is striped in record increments across the array of devices. This enables
an access by one actuator to complete a request from the host. Depending on
the striping increment, one full record or block can be accessed by any of the
array device actuators. Thus the arrays can support concurrent transactions or
I/Os, with each actuator capable of satisfying a user request.
Parity is distributed across multiple drives in the array, rather than having a
dedicated parity drive. Each device then contains both data and parity. With
parity distributed, the potential for contention of the parity drive is minimized
when writing multiple blocks. The chances that the parity device for a data
device will be available are much improved because the parity has been
distributed n ways across an array of n devices.
High data availability is provided through the use of parity devices. In the event
that data becomes unavailable from a disk drive in the RAID-5 array, the
subsystem can recreate the missing data through the use of stored parity
information. For data to become unavailable in a RAID-5 system, two drives in
the array must fail.
Independent access, striping, and distributed parity permit a very high bandwidth
for large block transfers, while at the same time providing high I/O rates for
short records.
When writing using RAID-5, the system must update both the data and parity (to
maintain recoverability). The new parity is calculated by the subsystem. To
update parity, the subsystem must perform four I/O operations: two reads (read
old data and old parity so that the new parity can be calculated) and two writes
(new data and new parity). Highly sophisticated destage algorithms limit the
impact of the RAID-5 write penalty.
RAID-6 provides independent access to the disk with floating parity. Multiple
concurrent accesses to the array devices are supported to satisfy multiple
concurrent I/O requests. Data is striped in blocks across the disk drives.
Therefore, single records or tracks can be read from one disk drive without
accessing other disk drives in the array.
To provide more continuous availability to your data, you need to find ways to
reduce the impact of backup, while still providing your end users with data
recoverability.
IBM has developed two methods of concurrent backup to help you reach this
goal:
• Concurrent Copy
• SnapShot Copy
The time between the start of the backup operation and the issuance of the
Logical Complete message is required to prepare an extent table in the 3990
cache. Except for this brief period of time required to initiate the Concurrent
Copy session, Concurrent Copy enables applications to continue running while
data set backups are performed concurrently. The extend table maintains
information about which tracks have been backed up.
In regard to Figure 26, when the application attempts to change a track that has
not yet been backed up (1), the original track is copied into a side file (2) and
sent to data mover in DFSMS/MVS (3). The track is held in dataspace until it is
merged into the correct place in the backup (4).
As soon as the original track is copied to the side file, the application update is
allowed to proceed. Concurrent Copy preserves the point-in-time state logically,
while the physical disk always contains the current state of the file.
DB/2, IMS and CICS have some ability to use the Concurrent Copy function
provided by DFSMS.
SnapShot works in conjunction with IBM Extended Facilities Product and the IBM
RAMAC Virtual Array Storage.
SnapShot can duplicate individual data sets, entire volumes, and VM minidisks
within the same RAMAC Virtual Array storage subsystem.
SnapShot supports sequential (PS), partitioned (PO), direct access (DA), and
partitioned extended (PDSE) data sets.
SnapShot has always been capable of snapping VSAM data sets, but previously
was restricted to the volume level and required manual actions to use the VSAM
data sets. It recently has been improved to support VSAM files at the individual
data set level. This new release also fully supports multi-volume data sets,
which previously had the same restrictions as VSAM file, as mentioned earlier.
This function can be delivered by both XRC and PPRC, although there are some
differences in both functionality and usability.
Using PPRC
Using PPRC, the main problem is to avoid obtaining fuzzy copies of data,
such as the one that results if you do not freeze your primary applications′
updates.
After being sure that the application stopped the updates, you can suspend
one or more PPRC pair by issuing the CSUSPEND command. You can
suspend in one command all PPRC active pairs attached to a primary control
unit by entering a wrong link path with CESTPATH. This will lead the
primary 3990 to truncate all the communication with the secondary 3990,
suspending at the same instant all the PPRC pairs. From this moment you
can resume your production activities, and all the updates of the suspended
pairs will be kept in a bitmap format in the 3990 nonvolatile storage (NVS).
You have to reset the secondary volume into a simplex state (CRECOVER) to
copy or dump it from another MVS partition.
When the copy processing ends, you can re-estabilish the links and
re-synchronize PPRC pairs. All the updates kept in the primary′s NVS will
be transferred and applied on the secondary volumes, thus re-syncronizing
the PPRC pairs.
Using XRC
Figure 27. IBM ′ s S/390 DASD Offerings. The broadest range of S/390 Disk Arrays
The RAMAC Array family offers you a broad range of products. Each product
has many continuous availability features.
• RAMAC 2 Subsystem (9394) offers 181GB in a compact frame. It continues to
be an entry-level RAMAC solution for those requiring smaller GB storage
with outstanding availability.
• RAMAC 3 and 3990-6 are examples of IBM protecting your previous
investment dollars. It allows your current 3990-6 to be upgradable, via a only
Based on the success of the industry-leading RAMAC Array Family and 3990
Model 6 storage control unit, RAMAC 3 Array Storage is an evolutionary step
into the third generation of RAID-5 Disk Arrays. RAMAC 3 incorporates the
experience of more than 11,000 installed RAMAC and RAMAC 2 disk arrays
worldwide. Since the RAMAC and RAMAC 2 disk arrays proved their high
availability design and leading performance, RAMAC 3 and 9390 Storage Control
benefit from this experience and knowledge from thousands of various customer
environments. The maturity of the microcode to properly handle drive failures
and much more, together with complete hardware redundancy and fast, reliable
electronic components gives RAMAC the best data availability in the industry.
IBM has integrated these RAID-5 drawers into the RAMAC 3 extension frame.
New adapter cards with higher throughput complement the capabilities of the
RAMAC 3 drawers.
The frame can be attached to either the new integrated packaged 9390 control
unit or to installed 3990-6 storage controls. Both solutions deliver great
improvements in performance, capacity and cost of operating, as well as
footprint savings. The new Licensed Internal Code (LIC) running the RAMAC 3
Array contains several enhancements as well, and delivers a RAID-5 storage
system solution of proven robustness and reliability.
The 9390 storage control is available in two models: the 9390-001 and 9390-002.
The name “model 001” stands for a single control unit and the name “002”
stands for a dual controller, each independent of the other. The model 001
controller can be upgraded to an model 002 controller. Model 002 is equivalent
to two 3990-6s. Highly integrated CMOS IV technology has been packaged into
the 9390 controller, significantly improving performance of the 9390 internal
processing as compared to the 3990-6.
Since one of the most important objectives of IBM is to protect your investment,
the broad range of all 3990-6 functions (especially Remote Copy for disaster
recovery) is fully supported by the new 9390 Storage Control. This makes data
migration nearly effortless when integrating the new 9390/RAMAC 3 Disk
Subsystem into your existing installations.
9390 and 3990 both support 3380 and 3390 format. The standard maximum
ESCON distance is 43 km. The 3990/9390 storage control unit is designed to
meet continuous availability requirements. Four storage paths between the
control unit and the disk are available. Two independent storage clusters per
frame provide separate power and service regions. Each control unit supports
128 logical paths to the host.
Additionally, 3990/9390 have no components whose failure will cause a total loss
of data availability. The 3990/9390′s fault-tolerant design supports two
independent, point- to-point storage clusters which each contain two storage
path (SP) microprocessors, a support facility (SF) microprocessor, and a disk
drive. This no-single-point-of-failure design continues the evolution toward fully
fault-tolerant DASD subsystems.
A service information message (SIM) is issued in the event of a fan failure, and
the fan can be replaced without impact to the normal operations of the 3990.
Control unit LIC, subsystem status, and diagnostic information are stored on two
hard disks (one per cluster). Two diskette drives (one per cluster) are also
available as a data transport vehicle.
The hard disks contribute to improved control unit availability in two significant
ways:
1. Two different levels of Licensed Internal Code can be stored on the disks,
thus enabling the control unit to switch from one level of LIC to a new level
with minimal disruption.
Because two levels of LIC are stored on the hard disks, the control unit can
rapidly return to the previous level of LIC in the unlikely event that a problem
is discovered with the new level.
2. The space available hard disks enable the control unit to store substantially
more diagnostic information should a hardware or LIC error occur. This
ability supports the control unit′s First Failure Support Technology, and is
designed to enable the vast majority of problems to be diagnosed and
corrected based on information gathered at the time of failure, minimizing
subsystem downtime for problem resolution.
Refer to “Concurrent Copy” on page 102 for further details about concurrent
copy.
Using dual copy may result in fewer application failures, fewer IPLs, and a
reduced frequency of DASD recovery activity. The implementation of dual copy
is straightforward since normally no application changes are necessary.
Normally, using the dual copy function requires no special attention from the
installation beyond having procedures in place to establish the dual copy pairs
and perform recovery actions when required. DFSMS management classes may
be used to direct the allocation of desired data sets automatically to duplex
pairs.
If you have 3390 DASD, it could be interesting to reserve a “spare” HDA for each
control unit to use concurrently with dual copy in case the microcode detects the
beginning of a production HDA failure. The 3990/3390 subsystem incorporates
extensive self-diagnosis features to detect errors as they occur and analyze the
error information to determine if action is required. A SIM delivers the results of
the subsystem problem determination, providing you to start a dual copy pair.
If ESCON directors are used between the two sites, PPRC links and ESCON
channels can share the same fiber links. Either way, no additional investment is
required on the 3990-6, as PPRC uses the standard ESCON interfaces.
PPRC provides synchronous data copying capability from one 3990 model 006 to
another 3990 model 006. This ensures full DASD data currency in the event of an
outage of the primary copy. Updates are sent from the primary 3990 model 006
directly to the secondary 3990 model 006 through ESCON links between the two
3990 model 006s (the host is not involved in the primary-to-secondary transfer).
The supported distance between DASD and processor, and between control units
for PPRC attachments is up to 43 km (26.7 miles).
To ensure consistency between the two copies of the data, any update is not
considered complete by the application until the data is written to both 3990s.
The unavoidable performance impact of this requirement is mitigated by the
ability of the 3990s to consider a write as complete once the update is in the
controller cache.
P/DAS is initiated by operator command, which enables the I/O supervisor (IOS)
to queue application I/O to the primary volume. IOS performs the switching
function, and then another MVS console command is issued to resume I/O to the
secondary volume. During the command execution and switching phases of the
process, IOS queues application I/O, thus avoiding the need to stop and restart
the application.
Figure 31. P/DAS and MVS Control Blocks before the Swap
The application system has connectivity to both application site DASD and
recovery site DASD. Four-path connectivity to application site DASD is through
CHP01, CHP02, CHP03, and CHP04, while four-path connectivity to recovery site
DASD is through CHP17, CHP18, CHP19, and CHP20.
The application site 3990-6 and the recovery site 3990-6 are connected through
four ESCON links that support the logical paths used for PPRC real-time data
shadowing.
Defined in the application′s code is the data control block (DCB), which contains
a description of the characteristics of the data, such as record format, data set
organization, and logical record length. When the application opens a data set,
a data extent block (DEB) is created, and its address is stored in the DCB. The
DEB contains details about the data set, such as access method type, data set
status, and extent information.
To access the data set on a DASD volume, the DEB also contains the address of
the unit control block (UCB), which is the software representation of the device.
The UCBs for all devices are created at IPL time when MVS reads the IODF data
set that is created by the HCD program. The UCBs are stored in the system
queue area of the MVS storage map. They contain details such as subchannel
number, device number, VOLSER, and CHPD array used to help define the
channel paths to access the device. So, the DCB points to the DEB, and the DEB
points to the UCB.
As shown in Figure 31 on page 112, only the primary volume of the PPRC pair is
online, while the secondary volume is normally offline. The primary device
number 100 uses channel paths CHP01, CHP02, CHP03, and CHP04, while the
secondary volume on device number 200 uses CHP17, CHP18, CHP19, and
CHP20.
Figure 32. P/DAS and MVS Control Blocks after the Swap
While the I/O is queued, IOS swaps the contents of the PPRC primary and
secondary volumes UCBs without interfering with the application control blocks,
which still point to the same UCBs address as before.
Figure 32 on page 113 shows that after the swap, the online device is device
number 200, which uses channel paths defines by CHP17, CHP18, CHP19, and
CHP20.
The application still points to the same UCB address, but now all I/O takes place
to the remote secondary volume as determined by UCB contents.
The entire process is triggered by only three MVS operator console commands,
assuming that PPRC volume pairs have already been established, and that the
primary and secondary volumes are in duplex status. These commands are:
1. IOACTION STOP,DEV=0100
Quiesce I/O to primary device
2. SWAP 0100,0200
Swap primary and secondary UCBs
3. IOACTION RESUME,DEV=0200
Resume I/O to secondary device
This function is implemented through a combination of the 3990 model 006 and
DFSMS/MVS software. Updates to data are copied asynchronously by
DFSMS/MVS from a 3990 model 006 at the primary site to DASD at a recovery
site.
The drawer offers third generation RAID-5 protection and the most reliable
microcode for its disk drives. The other hardware components, including the
microprocessors, are fully redundant and were developed to deliver 100% fault
tolerance. All electronic components are fully battery-backed up, and the battery
itself is checked by the control unit every 24 hours. If required, components can
be exchanged concurrently during normal system operation. The drawer can
perform data reconstruction on a failed disk drive when dynamic sparing has not
been performed.
The RAMAC 3 drawer incorporates the new IBM Ultrastar 2XP 3.5″ 9.1GB
high-performance disk drives which double the capacity of the previous disk
drives.
Each RAMAC 3 drawer contains eight 3390-3 or 3380-K logical volumes for a total
of 22.7GB of storage. The previous RAMAC 2 drawer offered a data capacity of
11.35GB for four logical volumes.
RAMAC 3 Array Storage and the 3990-6 attaching to the RAMAC 3 drawer retain
this strong focus on concurrent drawer maintenance by supporting both sparing
and dynamic disk reconstruction.
• RAMAC 3 sparing
There are two sparing options: dynamic sparing and manual sparing.
Dynamic sparing has the advantage of initiating automatically on fault
detection, thus providing maximum data loss protection by copying the data
to the spare drawer and restoring RAID-5 redundancy as soon as possible.
Manual sparing offers you the flexibility of invoking sparing for specific
drawers, and provides the choice of not reserving a spare drawer, but
instead, of quiescing a drawer when appropriate and initiating the sparing
process with the help of a service representative. Sparing with a RAMAC 3
configuration is similar to the process we are familiar with on RAMAC and
RAMAC 2 subsystems. The default for RAMAC, RAMAC 2, and RAMAC 3 is
to automatically start the sparing process, assuming that a spare has been
defined to the subsystem.
During the dynamic sparing process, SIMs are generated and sent to the
host so that the operations staff is aware of ongoing status. Up to two
drawers can be defined as spare drawers, but only one can be the target of
the sparing operation at any time.
Sparing is the only drawer maintenance function that protects against all
drawer failures. Dynamic disk reconstruction should be used only for disk
drive module (DDM) failures.
We recommend one spare drawer per RAMAC 3 storage frame.
• Dynamic disk reconstruction
Dynamic disk reconstruction offers an alternative method of concurrent DDM
maintenance. It applies to the DDM component of the drawer only and
enables disks to be replaced concurrently with drawer I/O activity. While the
DDM is being replaced, the enhanced RAID-5 function regenerates the data
for the failed disk. After the DDM has been replaced, the data is
reconstructed and written to the new disk.
Dynamic disk reconstruction does not offer the global “all-component” repair
philosophy of dynamic sparing because the technique applies only to DDM
components, but it does enable installations to consider a “no-spare” policy
for DDM repair scenarios. Dynamic disk reconstruction also avoids the time
taken to copy data to the spare drawer, but remember that dynamic sparing
paces the copy to minimize the impact on application performance.
Depending on the microcode level, you should perform all or some of the
following tasks:
1. 3990-6 LIC download
The 3990-6 enables LIC download to take place independently of LIC
activation. Thus an installation has more flexibility in planning the cut-over
to a new level of 3990-6 LIC by “readying” the subsystem in advance. The
3990-6′s fault-tolerant Each disk drive (one per cluster) can hold two levels of
3990-6 LIC (the current level being executed, and a second level, ready to be
activated). When 3990-6 LIC is downloaded, one copy is stored on each disk
in each cluster.
The 3990-6 LIC can be downloaded concurrently with I/O activity. During the
download operation, SPs, cache, NVS, and all other components continue
operating with no impact on application performance.
The SF performs download independently of SP application I/O processing.
The SF reads the LIC from the diskette and writes it to the hard disk in the
cluster.
Note: The download operation must be executed for each cluster.
2. 3990-6 LIC activation
For concurrent LIC activation, the activation process must be performed one
cluster at a time. As the initial step, the 3990-6 LIC activation process
invokes the control unit-initiated reconfiguration (CUIR) function, if enabled,
to request that all host channel path connections to this storage cluster be
varied offline by the host system. CUIR is a function of the 3990-6 only when
configured in extended mode. MVS/ESA, VM/ESA, and VSE/ESA support the
CUIR process.
The RAMAC Virtual Array offers an integrated storage control and disk array
storage capable of using both Parallel and/or ESCON channels. With 128 ESCON
logical paths capability, the RVA 2 can be fully exploited in a Parallel Sysplex
environment.
Figure 37 on page 122 shows you a high level overview of the internal structure
of the RVA T82.
Shared memory contains status information for the functional and physical
devices in the RVA subsystem. This status information is maintained by the
microprocessors. It is used to keep track of I/O operations in progress and to
manage the subsystem. Shared memory also provides a means of
communication between the microprocessors. Shared memory is duplicated for
redundancy.
To the right are the data handling functions of the RVA subsystem. The RVA can
provide host connection through either parallel or ESCON channels. The design
of the RVA allows for up to 32 parallel channels, or up to 16 ESCON channels.
The RVA′s log structured file architecture enables the use of data compression,
which facilitates the most efficient use of subsystem cache, nonvolatile storage
(NVS), internal data paths, and array disk space. Write data arriving over host
channels is compressed before it is stored in subsystem cache and NVS. Read
data coming from cache is decompressed before it is sent to a host channel.
The RVA cache can have a physical capacity from 512MB to 1536MB, increasing
in 128MB increments. Because of the RVA data compression, the effective
capacity of this cache is greatly increased when compared with a traditional
DASD subsystem with similar physical cache size. IBM conservatively rates the
effective cache capacity of the RVA family at twice the physical capacity, to
provide a meaningful comparison with traditional DASD subsystems. The RVA 2
and the RVA 2 Turbo can have an effective cache capacity from 1024MB to
3072MB, increasing in 256MB increments. All internal data transfers in the RVA
go through cache. All write operations are handled as DASD fast writes. Front
end (channel transfer) operations occur independently of back end (device
transfer) operations.
The RVA subsystem contains an NVS of 8MB physical capacity. All write data is
copied to NVS, in addition to being written to cache. Thus two copies of write
RVA provides four or eight internal data paths between the channels and cache.
There is one compression engine for each of the data paths.
When destaging write data from cache, the RVA appends dual redundant parity
before writing the data and parity to the disk arrays. Generation of the parity is
done by hardware. The parity enables the regeneration of data when one disk,
or even two disks in the same array, fails.
The RVA provides 16 separate data paths from cache to the disk arrays. Of
these, 14 are used to transfer read and write data. This arrangement makes
possible a high aggregate data transfer rate between the cache and the disk
arrays. Because the data is compressed, the effective transfer rate is
approximately 3.6 times the physical data transfer rate.
Disks can be defined as an eight-disk array group with 80GB capacity (5 data, 2
parity and a spare). This definition supports a minimum RVA 2 two-disk array
configuration of 160GB.
RVA disks can also be defined as a 16-disk array group with 210GB capacity (13
data, 2 parity and a spare). This definition supports a larger two array
configuration of 290GB (one 80GB array and one 210GB array). Capacity
increments of 80GB or 210GB support capacities up to the maximum
addressable 726GB (256 3390 model 3).
Since dual parity is provided, the rebuild is done in the background during
periods of low activity.
Data written to disks is compressed and compacted using full track blocking
without gaps. Since block lengths will change when updates occur, Log
Structured Array (LSA) architecture enables compression and compaction
because updates are never done in place.
Best volume utilization is achieved because the RVA does not update data in
place. This is a radical departure from other disk array architectures and
enables usage of unique and powerful storage subsystem functions. Unallocated
space does not use physical disk array space. Allocated but unused space does
not use physical disk array space either. This capability allows you to
over-allocate data sets in order to prevent out-of-space abends, allowing your
batch production to run more efficiently.
No RAID-5 or RAID-6 write penalty is incurred with this architecture because the
updated data and dual parity is always written to a new location. Dual
redundancy is not feasible without LSA because of the prohibitive write
performance penalty. As data is updated, it is moved within the subsystem and
placed on the disks with the least activity, thereby balancing the load of the disk.
The RVA is thus a self-tuning subsystem.
The IXFP Deleted Data Space Release (DDSR) function makes sure that RVA
knows about potential free space which is created when MVS deletes data sets
or VM deletes minidisks. Data set pointers are removed when data sets or
minidisks are deleted. This process can operate either as dynamic DDSR when
the deletion occurs, or as interval DDSR on a scheduled basis.
When the valid tracks have been properly relocated, the array cylinders are
added to a collected free space pool. This self-management is transparent to
the users.
Automatic load balancing is achieved by spreading the data across all of the
disks in the virtual disk system. This eliminates “hot spots” or disk access skew
which occurs in traditional disk array systems.
IBM RAMAC SnapShot for MVS and VM, or simply SnapShot, is a duplication
solution only possible with LSA disk array systems. As we discuss later, this
unique, advanced duplication product offers the potential to revolutionize the
batch processing operation.
The total amount of unused RVA disk free space can provide emergency
capacity for your peak workloads. This feature of the RVA LSA architecture is
unique and cannot be exploited by traditional disk.
Hot global disk sparing is another RVA exclusive. RVA Hot Global Disk Sparing
is truly global with a minimal performance impact because of the dual
redundancy design. The global disk sparing allows spare drives be shared
among different disk arrays.
Redundant pointer table protection: Logical disk volume tracks are dynamically
mapped with pointer tables to the physical disk locations. Over time, all logical
volume tracks will be spread across all of the physical disks in the virtual array
Pointer tables are located in separate electronic base memory and do not reside
in user cache. Since they are stored electronically, they are exposed to loss in
case of a power failure. They are continuously backed up to two separate disk
arrays, each with dual redundancy.
Because there is a delay in updating the disk copies, changes are journaled in
nonvolatile storage (NVS) to enable recovery from a power outage.
While the initial release of SnapShot enabled the duplication of VSAM data sets,
it implied some catalog manipulation, which made it difficult to use. This
difficulty was eliminated by the new release announced on April 8, 1997.
SnapShot operates on the data set, minidisk, and volume levels, and provides
MVS/ESA and VM/ESA users with a means to rapidly duplicate views to their
data.
No physical movement of data is required because only pointer tables. and not
the data, are actually copied. The original volume or data set is mapped to the
physical disks with a set of pointers. A volume snap (or data set snap) is
performed by duplicating the pointers (not by physically duplicating the data),
and then both pointers point to the same physical data.
Easy creation of test data without duplicated space means that your current test
processes which include dedicated disk arrays, host channels and processing
resources may now be reassigned to standard application processing. As an
additional benefit, the availability of complete, current test data may improve the
overall testing process for both Year 2000 testing and application testing which
may bring new applications online faster with fewer quality issues.
SnapShot permits the deferral of physical copies to off-peak periods. This has a
direct cost savings by reducing the tape processing resource requirement.
Additionally, by shifting peak sequential processing to non-peak periods, the four
or eight paths host channel constraint can be greatly reduced and additional
capacity added to RVA without compromising performance, thus reducing the
overall RVA system cost.
Only SnapShot can duplicate data at electronic speeds, thus creating major
competitive advantages for customers exploiting its revolutionary possibilities.
Figure 40. RSA-2: Control Unit Design. Scalable performance and capacity
Using an innovative, multiple bus architecture that permits very high scalability
of capacity and performance, the RAMAC Scalable Array 2 focuses on delivering
high-to-extremely high levels of performance in a disk storage subsystem.
The RSA-2 emulates 3380 J and K volumes as well as 3390 models 1, 2, 3 and 9.
You can also define FlexVolumes, which are allocated with a minimum of five
cylinders and can be increased in nine-cylinder increments.
Memory card failure will not affect performance because of the 100% mirrored
electronic memory. A channel adapter failure would cause the load from this
channel to be spread across other resources. The net result is a performance
system that is fully fault-tolerant and sets new standards in the industry.
The RAID-5 architecture uses nine data disks plus a parity disk for a 9+1 parity
system. The use of a very large cache allows the RAID-5 write operations to be
done asynchronously and thus eliminates any RAID-5 write penalty seen by the
host. The RSA 2 has combined the advanced RAID-5 disk array technology and
the IBM Ultrastar 2XP 9.1GB Disks.
RAID-5 function provides protection against a disk failure. The RAMAC Scalable
Array 2 also has a sophisticated approach to achieving high availability by
factoring other potential failures.
The RAMAC Scalable Array 2 uses a true dual port architecture that ensures full
data access in case of a device controller or SCSI bus failure. Full function dual
port is not offered by any SCSI disk drive manufacturer, so the RAMAC Scalable
Array 2 uses a custom-designed card that provides two SCSI ports to the single
port SCSI drives.
All power functions are duplexed (two power cords, separate AC/DC converters,
separate battery systems, and separate voltage converters). Apart from the AC
input control, all power system components are hot-pluggable. If an AC input
control should fail, the other AC input control will take over.
The RAMAC Scalable Array 2 keeps two complete copies of all data in separate
cache cards that are serviced by separate memory controller cards, each
powered by separate power supplies and separate battery systems.
All reads occur from both sides of the mirrored cache and the data is compared
bit for bit, giving essentially an infinite bit error detection. If a transient
uncorrectable error does occur during reading from the mirrored cache, there
will be a miscompare at the bit level and the failing side will report an ECC
uncorrectable error. The controller can then tell which data is good and satisfy
the read request with good data. The controller will then overwrite the bad data
with the correct data from the other cache card.
If a cache card should fail, the system will be able to continue at full
performance. The error will be recognized by the controller and the controller
will phone home to report the error. The system will switch off mirrored mode
and just read and write to the good card. After the defective card is replaced,
the controller switches to recovery mode. All reads will go to the good side and
all writes will go to both sides. As a background operation, the system will read
from the good side and copy the data to the replaced memory card. This
recovery operation continues until fully mirrored data resides in both memory
cards and the system switches to full mirrored operation.
The device controllers (DC) are responsible for all media management tasks.
The DC performs the disk-scrubbing and error detection and correction for the
disk device. If a media error occurs, the DC maps around the defective media,
assigns an alternate sector or track, rewrites the data, verifies the write and
increments the soft error counter. When the soft error count exceeds the
threshold, the support controller is notified and a SIM is generated.
But the REA 2 also offers new levels of availability by duplexing hardware
components including the solid state memory. In the event of a power failure
batteries take over, providing 48 hours of power for the largest 4GB duplexed
system and greater than 48 hours for smaller memory configurations.
The basic architecture of the REA 2 is the same as the architecture of the IBM
RAMAC Scalable Array 2. The REA 2 is essentially an RSA 2 without the
attached disk frame and a different Licensed Internal Code (LIC)load. Your
investment in an REA 2 provides you with the ability to upgrade to an RSA 2
simply by adding a disk frame and changing the LIC. This provides superior
investment protection.
All reads occur from both sides of the mirrored memory and the data is
compared bit for bit giving essentially an infinite bit error detection. If a
transient uncorrectable error does occur during reading from the mirrored
memory, there will be a miscompare at the bit level and the failing side will
report an ECC uncorrectable error. The controller can then tell which data is
good and satisfy the read request with good data. The controller will then
overwrite the bad data with the correct data from the other memory card.
If a memory card should fail, the system will be able to continue at full
performance. The error will be recognized by the controller, the system will
switch off mirrored mode and just read and write to the good card. After the
defective card is replaced, the controller switches to recovery mode. All reads
will go to the good side and all writes go to both sides. As a background
operation, the system will read from the good side and copy the data to the
replaced memory card. This recovery operation continues until fully mirrored
data resides in both memory cards and the system switches to full mirrored
operation.
All power function are duplexed (two power cords, separate AC/DC converters,
separate battery systems, and separate voltage converters). Apart from the AC
input control, all power system components are hot-pluggable. If an AC input
control should fail, the other AC input control will take over.
The support facility performs such functions as configuration changes, initial LIC
load, error analysis, logging and reporting. The support facility is a critical
element for high data availability by continuous monitoring and reporting of all
key components. The support facility is also fully redundant.
The internal disk drawer advances the use of industry standard disks, or JBODs
(Just a Bunch of Disks), in a RAID 1 configuration. With internal disk, customers
Refer to 2.2.3, “9672 High Availability Features” on page 9 for more detailed
information.
As we seen, many IBM DASD functions can help you achieve your data
migration requirements in term of continuous availability Let us have a look at
each of them.
• Software Physical Copy
Physical volume copy is the traditional method for migrating data from
similar device geometry. If you are moving volumes to like devices of larger
capacity, generally you need a larger VTOC because the larger device can
hold more data sets. This type of migration is disruptive in most cases
because you have to “freeze” all updates on the migrating volume.
• Software Logical Copy
Logical copy allows you to move data between unlike devices. As a physical
copy, this is a disruptive migration.
• Using SMS
Using SMS to move data provides an easy non-disruptive migration path.
The following procedure could be use to migrate data either between like or
unlike devices with SMS.
1. Define a new storage group.
In this branch, four bank tellers are working at their positions. They each have
access to all the money and other facilities needed to support the services the
This simple example shows what the Parallel Sysplex approach to continuous
availability is about. The services provided by the bank are viewed from the
outside. The door of the bank is never closed as a result of a single teller
position being unavailable; instead, as long as any position is open, customers
are still allowed to enter and their requests for service will still be handled.
save You can view Parallel Sysplex in the same way. If multiple members are
sharing the workload then the loss of any single member will not cause a loss of
services from the sysplex as a whole. The “door” of the sysplex remains open
to customers, and work is simply re-routed to other active members. In this
way, Parallel Sysplex offers a platform for both high availability and continuous
operation.
One final example demonstrates another important part of the Parallel Sysplex
design.
In the same bank branch, if the workload starts to increase, an additional teller
is required to meet the extra demand. It would be unacceptable to close the
bank to allow this additional teller to help with the workload. Instead, by
ensuring that the additional teller has both the correct skills and access to all the
same facilities as the other tellers, a new position is easily opened to deal with
the load. This occurs without interrupting the service to customers.
• (This is another aspect of continuous operation: non-disruptive change.)
Important
The fact that a Parallel Sysplex can meet all of these challenges makes it the
ideal platform for delivering continuously available services.
There are many books on the subject of Parallel Sysplex which describe the
technical details of exactly how data sharing and dynamic workload distribution
are achieved. A short description of each is included later in this chapter after
the main components of a Parallel Sysplex have been described. This chapter
concentrates on examining the sysplex components and the way they combine.
The aim is to review alternative Parallel Sysplex configurations to understand
limitations or strengths when designing for continuous availability. This chapter
also discusses the topic of non-disruptive software upgrades and what must be
done to perform them successfully.
The processors shown in this figure are S/390 models with multiple processing
units to provide recovery in the event of an error. This configuration may allow
a maintenance action to be deferred to a more convenient time, when full
capacity is not required from the sysplex. Several types of S/390 processors can
participate, from the final generation of the water- and air-cooled ES/9000 range,
through all generations of the 9672 CMOS range. The environment supporting
the sysplex is also assumed to be free of single points of failure (power line, air
conditioning, and so on).
The objective of the coupling facility is to form a shared storage device for use
by the active members of the sysplex. The information stored in the coupling
facility can be active data, locking information for active data, or
informational-type records and messages required by all sysplex members.
These records in the coupling facility are called structures . By placing the
relevant data and control information in this shared storage, recovery and
7.3.2.1 Description
The hardware which forms the communication path between the coupling facility
and the S/390 processor is known as the coupling link. Fiber optic cable (either
multi-mode or single mode) is used to connect the two devices. This is a high
speed, high bandwidth design and although it uses the same standards of fiber
optic cable used by ESCON channels, it does not use normal ESCON channel
protocols. Links can be added individually from 0 up to a maximum of 16 on a
9672, and from 2 to a maximum of 32 on a 9674 standalone coupling facility.
Each coupling link can be considered as directional, that is, the end connected to
an OS/390 image is defined differently from the end connected to a coupling
facility. These ends are known as CF Send (CFS) and CF Receive (CFR)
respectively. The ESCON Multi-Image Facility (EMIF) allows the CFS hardware of
a link to be shared between OS/390 LPARs.
Note: The CFR link hardware cannot be shared at this time.
A link attached to a normal 9672 can be defined as either a CFS or a CFR if ICF
is being used. This is defined and changed using HCD.
Note: The link attached to a 9674 standalone coupling facility can only be a CFR
type.
The standard maximum distance between a 9037 and a processor is 3 km, but an
RPQ is available to extend this distance using the 9036 model 003 Extender. The
Distributed Console Access Facility (DCAF) can also be used to provide a remote
console access to the 9037 timer. This can be over a Token Ring or via a dial-in
connection.
The 9037 can also use the Coordinated Universal Time (UTC) signal as a further
check on its accuracy and status.
You should ensure that primary and alternate devices are separated between
different DASD subsystems, and are placed on low activity volumes. Avoid
mixing these datasets with other data that may be used by an application using
a Reserve on the volume, for example some HSM functions.
The Sysplex Couple Dataset is used by all members to check that each member
is alive. Each is required to read and write to this dataset within a set time
period. A reserve may prevent this from happening and cause a member
system to be forced out of the sysplex because it has not been able to complete
this operation. Other conditions can cause similar contention. Refer to S/390
MVS Parallel Sysplex Continuous Availability SE Guide for further information in
this area.
ESCON Directors reduce the need for permanent connections between each
processor and all devices, by allowing a dynamic connection only when
required. They also make reconfiguration and upgrade much easier as they
7.4 Datasharing
Datasharing is the ability of the Parallel Sysplex to allow its members read/write
access to the same databases, while preserving the sequence of updates and
ensuring the integrity of each database. This must also be performed with
minimum performance overheads. The coupling facility is used extensively to
solve each of these challenges by using the different types of structures to lock a
record during update. It is also used to make the updated record available to all
without waiting for normal I/O write operations.
Chapter 10, “IMS” on page 185 and Chapter 11, “DB2” on page 199 explain
how IMS and DB2 make use of the coupling facility to achieve datasharing, and
no further technical explanation is included here.
Chapter 9, “CICS” on page 163 and Chapter 10, “IMS” on page 185 describe
how each of these transaction management systems exploits workload balancing
in a Parallel Sysplex, so no further technical information is included here.
The figure shows that this now requires four coupling links on each processor
(or eight, for redundancy), and the cables between the two machines can be left
permanently connected,
You may ask, is a power-on reset required to change the LPAR definitions? This
is a situation that the Dynamic coupling facility Dispatching (DCFD) feature can
assist with. Without the DCFD, an ICF backup like the one shown in Figure 44
would take CP cycles away from the OS/390 LPARs sharing the CP resource.
With DCFD however, the ICF only runs when there is a request for it to process.
It now offers a solution to providing a test and development facility that can
tolerate planned outages.
Note: This configuration is not recommended to support production workloads,
because using only an an ICF creates a single point of failure that can affect two
major components.
In this configuration, EMIF is still used to share CFS ends of each coupling link,
but again no link redundancy has been shown. Full redundancy would require
all numbers to double again, needing eight links on Processor A, and four links
each on Processor B and the 9674.
Again, consider the possibility that the function of Processor A may need to be
switched to Processor B. In a Parallel Sysplex, the workload can be seamlessly
moved if Processor has enough capacity to carry it all.
Many installations, however, may plan configurations that allow the full function
of Processor A to be swapped with Processor B and vice versa. This means that
the full hardware configuration of Processor A must be repeated on Processor B,
including all coupling links. This approach requires careful planning and
assessment, to see if the additional hardware and reconfiguration effort required
to make the change is justifiable. If the configuration shown was intended to be
built across two different sites, then placing the 9674 and Processor A on the
same site may give enough resilience to meet the service requirements.
Also, the maximum number of coupling links that can be installed on a 9672 is
16, which can be rapidly used up in a configuration like this.
In this configuration, the workload can be split not only between the two
processors, but also can be split between the two coupling facilities. This offers
the best combination of resilience and performance available to support a full
data-sharing, workload balancing Parallel Sysplex. It also uses the minimum
number of links on each S/390 9672 processor.
New releases of OS/390 are currently scheduled for April and October of each
year. This probably places the biggest burden on systems programming staff to
keep within the thresholds. Releases of the major subsystems are more
variable, and are generally not superceded for at least two years. Service will
still be released for all products and should not be ignored.
The same constraints must apply when redesigning old or developing new
applications to suit the Parallel Sysplex environment. If multiple versions of the
same application are intended to be active concurrently in the sysplex, their
ability to tolerate each other must have been considered and proved.
Again the subject of effective local testing is important to ensure that changes to
the software inventory are fully checked before introducing them into the
production sysplex. It is impractical for the full combination of IBM and OEM
The major hardware purchase is probably the additional coupling links that are
needed on all machines to build this configuration. The alternative is to add
additional processors, which increases the software licensing costs more than
this approach does.
7.11 Summary
The Parallel Sysplex is IBM′s most extensive solution to address the challenge
of continuous availability. It offers an environment that fully supports all the
demands of eliminating both unscheduled outages and scheduled outages, and
yet this environment can be easier to operate and run than most multiple
processor configurations in use today.
8.1 Introduction
To this point, most approaches to continuous availability, high availability, and
continuous operation have concentrated on multiple systems. The best solution
for any platform to attain continuous availability appears to be having a Parallel
Sysplex built around multiple processors. However, multiple processor
footprints and LPARs, with multiple coupling facilities for redundancy and high
availability, cannot be cost-justified by many single processor installations.
Therefore, you need to ask: Are there are any benefits associated with Parallel
Sysplex to consider its use on a single processor?
8.2 Configuration 1
Figure 48 on page 156 shows a basic arrangement of LPARs within a S/390
processor to support this approach.
The configuration shown makes use of the Integrated Coupling Migration Facility
(ICMF), which is a standard feature of the S/390 PR/SM microcode. As well as
allowing a number of LPARs to be built, PR/SM allows up to two coupling facility
LPARs to be created within the same machine. It also emulates the Sysplex
Timer function and coupling links between OS/390 and the coupling facility. This
means that a sysplex can be built without acquiring additional hardware,
although capacity (storage size and the number of CPs) may need to be
reviewed.
The targets or goals set for WLM can be altered without closing and reloading
the system, which may also eliminate one or two scheduled outages each year.
WLM can be used in a single system without migrating to a full Parallel Sysplex,
that is, it can be activated without the use of a coupling facility. Refer to OS/390
MVS Parallel Sysplex Configuration Cookbook for further information on WLM.
When all testing is complete, the OS/390 test system can be defined to allow it to
become a member of the production sysplex. The coupling link connection can
remain associated with the test OS/390, because until the system is defined to
become part of a sysplex it will not use the link. This means that no redefinition
is required to switch the Test OS/390 image into and out of the Parallel Sysplex.
The two OS/390 LPARs will now be sharing the workload. This should continue
until the test LPAR has been accepted as a solid base for supporting the
production workload. The length of this period depends on the installation, but a
full month should cover most business cycles to prove the new system.
When the new system is accepted as stable, the closedown procedure for the old
production LPAR (OS/390 1) can be started. When drained, it can be removed
from the sysplex. Now the upgrade/maintenance process can begin again using
this drained system as the new test facility. Its libraries should be dumped to
tape before making any new changes as a precaution.
When combining the two images in the Parallel Sysplex, some form of resource
serialization is needed to safely share the production I/O subsystems. In a
single processor installation, this may not have been required before. The
OS/390 component, Global Resource Serialisation (GRS), allows this I/O sharing
to be performed correctly, and is also a pre-requisite for sysplex activation. It
Network control takeover depends on the machine type and features used for
connection to the network. If a native ESCON channel interface is available, then
this can simply be shared between the LPARs and EMIF. If the network front end
processor (FEP) is parallel channel-attached, then several options are available.
• The simplest option is to use a second channel adapter on the FEP. This
allows the network to be shared easily between the LPARs.
• If using a second channel adapter is not an option for you, then
non-disruptive change becomes impossible. Instead, you can perform
takeover by using CTCs between the LPARs, so that the test LPAR can use
the CTC for VTAM when joining the sysplex. This would be fine until the
time arrives to close the production LPAR. If the parallel channel to the FEP
is defined as reconfigurable, then it can be switched from the production to
the test system, but this means an interruption to services through the FEP.
Your choice of solution really depends on what hardware you are using.
A similar switching problem probably exists for all printers, and 3x74 local
terminal device controllers.
CTC links can be used between the LPARs to overcome this response problem
for XCF traffic, and are probably useful for other traffic (for example, VTAM).
DB2 in particular makes extensive use of the coupling facility when working in
datasharing mode, so heavy users of DB2 may find this performance overhead
unacceptable. This penalty can be overcome with the use of configuration 2.
8.3 Configuration 2
The next step up the evolutionary scale is shown in Figure 49 on page 160. It
uses four coupling links to form the communication path between the OS/390
LPARs and coupling facility LPAR. This overcomes the performance penalty
associated with the emulated links. All coupling facility requests that can be
handled synchronously will now be processed in this way. Another advantage of
using the external links is that they can be dynamically redefined, if required.
The EMIF feature of the S/390 allows the CFS end of the link to be shared
between OS/390 LPARs. This means that, for the period that both LPARs need
to access the CF, both links can be shared for both resilience and performance.
Alternatively, only one link can be defined to the test LPAR, and then the second
added after testing in the sysplex has shown the new test image to be stable.
IPL′ing an LPAR with no sysplex definitions will stop the LPAR from using the
coupling link.
An extra step is now possible after testing the new system in standalone mode.
If no devices are shared between the two LPARs, it is possible to operate the
test OS/390 in a Monoplex. (See section 8.3.2, “Multiple Sysplex Considerations”
on page 161 for restrictions on sharing devices outside of a sysplex.) This is not
a full Parallel Sysplex configuration, but simply the connection of the test LPAR
to a different set of sysplex couple datasets. This could allow testing with WLM,
which is an important aspect of the workload balancing and takeover between
images when joining the production Parallel Sysplex.
The sysplex couple datasets that form a non-volatile source of configuration and
control information for each sysplex are also non-sharable between sysplexes.
A group of sysplex couple datasets (primaries and alternates) is needed for each
sysplex. Care needs to be taken when placing these control datasets, to avoid
possible interaction between the two sysplexes. Guidance is printed in many
sysplex manuals on where to place these datsets to best advantage; you can
also refer to S/390 MVS Parallel Sysplex Continuous Availability SE Guide .
8.4 Summary
As the examples have shown, management of the LPARs in this way makes it
possible to introduce software changes without impacting business services.
Other variations, with different number of LPARs and coupling links, are also
possible.
You should also note that, although simple examples are provided, planning and
constructing a Parallel Sysplex is not a trivial exercise. Project durations of six
to 12 months are normal. The success of the project depends upon the effort put
into it by your technical support team, especially at the planning stage. The
more effective the planning is in anticipating and understanding obstacles, the
greater and earlier the success will be.
IBM now has staff with a high level of experience in planning and constructing
Parallel Sysplexes in different situations. Their skills and experiences are
available at every stage of your project. Refer to section 7.10, “Parallel Sysplex
Implementation Service Offering (EPSO)” on page 153 for a sample of services
available.
With each new CICS, version the number of operations required to shut down the
CICS system is reduced.
MRO is used to connect CICS regions that are within the same MVS image or
that are using the cross-system coupling facility (XCF) of MVS/ESA 5.1 within the
same Parallel Sysplex (MRO/XCF).
The following five CICS regions (building blocks) can be combined to implement
a CICS architecture for a specific installation:
• Terminal owning region (TOR)
• Application owning region (AOR)
• File owning region (FOR)
• Data owning region (DOR)
• Queue owning region (QOR)
ISC allows CICS to communicate with other CICS or IMS systems in a different
MVS system. You normally require an SNA access method, such as ACF/VTAM,
to provide the necessary communication protocols. ISC allows the
implementation of the same kind of regions that can be used with MRO.
Figure 50 shows an example of connected CICS regions using MRO, MRO/XCF,
and ISC.
When MRO, MRO/XCF, and ISC are not exploited, terminal management, data
management and application belong to the same CICS region. A functional split
is a partitioning of these functions into separate CICS regions. For example, the
terminal management function can be isolated in a TOR, file management can
be isolated in a FOR, and application can be isolated in an AOR.
Distributed program link (DPL) enables a CICS program (the client) to call
another CICS program (the server program) in a remote CICS region. Using
DPL, you can separate the end-user interface from the application business
logic, such as accessing and processing data, to enable parts of the applications
to be ported from host to workstation more readily. In many cases, DPL offers a
simple alternative to writing distribute transaction processing (DTP) applications.
CICS support for DCE remote procedure calls (RPCs) enables a non-CICS client
program running in an open systems Distributed Computing Environment (DCE)
to call a server program running in a CICS Transaction Server for OS/390
Release 1 system, and to pass and receive data using a communications area.
The CICS program is invoked as if linked to by another CICS program.
The functional split, supported by TR, FS and AP allows the isolation of functions
into TOR, AOR, FOR, QOR, and DOR, and provides:
• Stable TORs, FORs, QORs and DORs containing no user code
• Very fast restartable AORs without logs and independent of the TOR, which
takes much longer to restart
• Stable DORs, FORs, QORs handling DASD, buffers, and logs for increased
data integrity ans availability
With static transaction routing , the transaction is predefined with the name of a
remote CICS region to which it is to be routed for execution.
With dynamic transaction routing , one of the remote attributes on the transaction
resource definition specifies DYNAMIC(YES), indicating that the transaction to be
Improved availability can be achieved in the form of both high availability and
continuous operations:
With multiple AORs available to choose from, the dynamic transaction
routing program bypass any failed or “sick” AOR. Faster restart is possible
for smaller AORs, and fewer end users are affected by any single failure. All
these factors contribute to:
• Reducing the number of failures seen by end users
• Reducing the length of outage as seen by end users
• Reducing the incidence of “sympathy sickness” between CICS regions
Dynamic transaction routing helps you to achieve continuous operations by
allowing you to:
• Quiesce an AOR and direct transactions to an alternate AOR
• Stop and restart an AOR to apply service changes to the AOR when no
transactions are active
• Add an AOR clone and notify the dynamic transaction routing program
that a new target AOR is available
CICSPlex SM′s DTR program supports both workload separation and workload
balancing:
Workload separation The routing of particular transactions to an particular
group of AORs based on any combination of user ID,
terminal ID, and transaction name. For example, using
CICSPlex SM′s workload separation function, you can
specify that transactions beginning where the characters
SAL and initiated by members of your sales department
must be routed to the group of AORs, SALESGRP,
allocated to that department.
Workload balancing The routing of transactions among a group of AORs
according to the availability and activity levels of those
AORs. Workload balancing can be used in addition to,
or in place of, workload separation. For example,
CICSPlex SM can balance the transaction workload
among the SALESGRP AORs by selecting, as each
transaction is initiated, the AOR that is likely to deliver
the best performance.
DFSMS 1.3 supports a new data-sharing subsystem, SMSVSAM, which runs in its
address space to provide the RLS support required by batch jobs and AORs or
cloned FORs within each MVS image in a Parallel Sysplex environment. An
SMSVSAM server, used by the resident AORs, is shown in each of the MVS
images in Figure 53 on page 172. The SMSVSAM server, which is generally
In earlier CICS releases, VSAM recovery attributes for a file are defined on the
file′s resource definition in the CSD. To support RLS access, VSAM supports
some new data set attributes like “recoverable” and “backup-while-open
(BWO).” These attributes are stored in the ICF catalog.
The main benefits accrue when VTAM generic resource registration is applied to
CICS terminal owning regions. This enables VTAM to perform dynamic workload
balancing of the terminal sessions across the available TORs. If a TOR is not
available for any reasons, VTAM does not try to establish a session to that TOR.
CICS support of persistent sessions includes the support of all LU-LU sessions
except LU0 pipeline and LU6.1 sessions. CICS determines for how long the
sessions should be retained from the PSDINT system initialization parameter.
This is a user-defined time interval. If a failed CICS is restarted within this time,
it can use the retained sessions immediately; there is no need for network flows
to rebind them.
The end user of a terminal sees different symptoms of a CICS failure following a
restart, depending on whether VTAM persistent sessions, or XRF, are in use:
• If CICS is running without VTAM persistent sessions or XRF and it fails the
user sees the VTAM logon panel followed by the “good morning” message
(if AUTOCONNECT(YES) is specified for the TYPETERM resource definition).
• If CICS does have persistent session support and fails, the user perception is
that CICS is hanging and what is on the screen at the time of the failure
remains until persistent session recovery is complete.
While the active CICS is working, the alternate system is monitoring it for
failures and can detect failures more quickly than an operator. When the
alternate has detected a failure, the automatic takeover time is faster than an
ordinary restart, because the system is already partly initialized.
If you do not use dynamic transaction routing, you cannot get all the availability
benefits of that facility. Implementing XRF may be an alternative if you
nevertheless want high availability.
CICS has always allowed you to create transactions with ACID properties (4)
(atomicity, consistency, isolation, and durability). In particular, the atomic
property of a CICS transaction means that any changes the transaction
makes to recoverable resources (such as files, temporary storage, or DB2)
are carried out in such a way that either all changes happen or none
happen.
However, in previous CICS releases, restrictions were imposed on
distributed units of work; that is, on units of work that update resources on
two or more systems. These restrictions were necessary because of the
difficulties of coordinating updates to distributed resources.
In CICS Transaction Server for OS/390, the CICS recovery manager allows
you to distribute recoverable resources in a network while ensuring that all
updates made by a distributed unit of work are coordinated, thus preserving
the transaction′s atomicity.
• Log manager domain
The CICS log manager is a new domain that replaces the journal control
management function of earlier releases. It works with the MVS system
logger, which is introduced in MVS/ESA 5.2, to provide a focal point for all
system log, forward recovery log, and user journal output within an single
MVS sysplex. It provides flexible mapping of CICS journals onto MVS log
streams, which you can use, for example, to merge all the forward recovery
logs for a given VSAM data set from many CICS systems on several MVS
images.
9.7.2 Autoinstall
Before the introduction of autoinstall, all resources had to be defined individually
to CICS before they could be used. With the introduction of autoinstall for
terminals in CICS/OS/VS 1.7, VTAM terminals could be defined dynamically on
their first usage, thereby saving on storage for terminal entries and saving on
time spent creating the definitions (and possibily making definiton failures).
To provide protection for the transaction data of application programs, the CICS
transaction isolation facility requires hardware and MVS facilities that support
subspace groups. For more information about subspace groups, see 2.2.3.12,
“Subspace Group Facility” on page 12.
You cannot use MVS automatic restart for CICS regions running with XRF; it is
available only to non-XRF CICS regions. If you specify XRF=YES, CICS
deregisters from ARM and continues initialization with XRF support.
The new adapter must be used with CICS Version 4 and works with all current
releases of DB2. The CICS attachment facility from previous releases is still
delivered with DB2.
CICS TS improves the CICS-DB2 attachment. It can now remain active if DB2
fails, and automatically reconnect when DB2 recovers. Previously, if DB2 failed,
the CICS-DB2 attachment was stopped and had to be manually restarted when
recovered.
If the CICS-DB2 attachment is active but DB2 is not active (STRTWT=AUTO), the
attachment can return a SQL return code to an application that tried to issue a
SQL request, rather than to the application abending with abend code AEY9.
The CICS-DB2 attachment facility has been enhanced to support the new
INDOUBTWAIT RMI protocol. The CICS-DB2 task-related user exit (TRUE) is
The CICS-DB2 attachment has been enhanced to issue calls to the MVS
workload manager if a severe error occurs, or if the CICS region is active but not
connected to DB2, so that CICS-DB2 work is not routed to this CICS region.
The benefit from an availability point of view is that you make your business
applications available through a new and additional access path.
The BWO facility, together with other system facilities and products, allows you
to take a backup copy of a VSAM data set while it remains open for update.
Then, only the updates that have been made since the last backup copy was
taken need to be recovered. This could considerably reduce the amount of
forward recovery that is needed.
IMS has always been known for high availability. As the product has evolved, it
has become more reliable and more available for your end users. The time
needed to run batch jobs (the “batch window”) has been reduced because more
of these tasks can be done online.
The Database Image Copy 2 utility (DFSUDMT0) performs the tasks required to
take image copies of IMS databases. The databases must therefore registered
in the DBRC.
10.3.2 Logging
In the online environment, IMS keeps a log called the online log data set (OLDS).
The OLDS is on a direct access storage device (DASD) and is later written to
another data set called the system log data set (SLDS). This SLDS can be on
DASD, tape, or mass storage. In the batch environment, log records are written
directly to an SLDS. IMS also keeps another log called the restart data set
(RDS). The RDS contains information needed for restart. The writing of logs is
handled totally and automatically by IMS.
10.3.3 Recovery
There are two major forms of recovery: one involves reconstruction of
information (called forward recovery) and the other involves removal of bad or
incomplete information (called backout or backward recovery). Wit forward
recovery, you reconstruct information or work that has been lost. With backward
recovery or backout, you remove incorrect or unwanted changes from
information.
In IMS, you perform the task of forward recovery (although IMS does supply
utilities to help).
In order to get maximium availability, you need to develop a plan for backup and
recovery that minimizes the duration of the outage because data are not
available. You must decide how to use the tools that IMS provides for operating
your system. This includes choosing, for example, whether to use dual logging
IMS DB Version 5 was the first to support sysplex data sharing. This means that
multiple IMS systems (up to 32) can now share data across multiple MVS images
(also up to 32). Before this enhancement, you could only share data across two
MVS images.
Block-level sharing of VSO DEDB areas in IMS Version 6 allows multiple IMS
subsystems to concurrently read and update VSO DEDB data. The three main
participants are the coupling facility hardware, the coupling facility policy
software, and the XES and MVS services. Figure 63 on page 192 shows the
major components of this environment.
CQS uses the MVS coupling facility as a repository for the data object. The
coupling facility stores and arranges the data according to list structures. The
shared queues are kept in a primary list structure and optionally, in an overflow
list structure. The two list structures are known as a structure pair .
CQS maintains the following components for each structure pair it manages:
• A primary structure that contains the shared message queue.
• An overflow structure (optionally).
• An MVS log stream that contains all CQS log records from all CQSs
connected to a structure pair. This log stream is important for recovery of
shared queues, if necessary. There is one log stream for each structure
pair.
• A checkpoint data set that contains CQS system checkpoint information.
Using shared queues enables automatic workload balancing across all IMS
subsystems in the sysplex. Also, new IMS subsystems can be added to the
sysplex, and share the queues as workload increases, thus resulting in
incremental growth. And using shared queues provides increased reliability and
failure isolation: if one IMS subsystem in the sysplex fails, any of the remaining
IMS subsystems can process the work that is waiting in the shared queues.
Distributing sessions among the IMS systems in a generic resource group has
the following advantages:
• It avoids bottlenecks within any one IMS system,
• It uses the available capacity in other systems to share the work.
• It provides increased availability by allowing sessions to be re-established
with another IMS, upon the failure of an IMS image.
Previously, in order to balance sessions across IMS systems, you had to devise
a means of assigning terminals to specific systems. Now VTAM performs that
task for you.
The IMS Partition DB, and the additional utilities supporting operations on
partition level provide the following benefits:
• Capacity for growth
• Manageability
• Higher availability
• Potential for better performance
• Ease of use
• Transparency to applications
• Simple conversion
DB2 was designed so that the operator should not have to take DB2 down in
order to perform traditional database activities. Every new version of DB2
delivers new functions that are designed to ensure high availability.
In this chapter we discuss the main procedures and features of DB2 that are
introduced in Version 4.1 or 5.1 to improve both high availability and continuous
operation. The main areas we cover are:
• Summary of DB2 continuous availability attributes
• DB2 recovery features
• DB2 Parallel Sysplex exploitation
• Other DB2 continuous availability features
The CHANGELIMIT option on the COPY utility lets DB2 determine when an image
copy should be performed and whether a full or incremental image copy should
If you run REORG or LOAD REPLACE with the COPYDDN option, a full image
copy is produced during the execution of the utility and DB2 does not place the
table space in copy-pending status. This eliminates the period of time when the
table space is in copy-pending status and a separate COPY step is not required.
Therefore the data is available sooner.
If you use LOG(YES) and log all updates, then an image copy is not required for
data integrity. However, the utility job elapsed time increases because of the
additional logging activity, and in a recovery case, the recovery process using
the image copy is much more efficient than applying a large number of log
records.
11.2.3 Logging
DB2 provides a common log to record data changes for all activities running in
the DB2 subsystem. Since the information in the log is extremely critical for
preserving data integrity, provision is made for writing two copies of the log, for
either the active log and/or the archive log.
DB2 also provides an automatic archiving facility. The active logs are
maintained on disk and as an active log fills up, it is automatically archived to
tape or disk. An active log space cannot be reused until the existing log data has
been archived. DB2 will stop processing if all active logs become full before this
archiving can take place. It is also very important, therefore, to have enough
active logs specified and to make sure that the tape or DASD used for archiving
is available promptly. If dual archive logs are specified, the active log is written
to two archive log data sets.
11.2.4 Recovery
To recover a table space, the RECOVER utility uses these items:
• A full image copy, that is, a complete copy of the table space
• Any later incremental image copies, each ow which summarizes all changes
made to the table space from the time the previous image copy was made
• All log records created since the most recent image copy
In order to get maximium availability, you need to develop a plan for backup and
recovery that minimizes the duration of the outage because data is not available.
Some factors to consider are:
• Backup frequency
Use your recovery criteria to decide how often to make copies of your
database. You should make copies after jobs that make large numbers of
changes. Recovering from an image copy is faster than to applying a large
amount of log records.
• The right characteristics for your logs
For example, if you have enough DASD space, use more and larger active
logs. Recovery from active logs is faster than from archive logs.
For more information about recovery consideration, refer to DB2 for OS/390 V5.1
Administration Guide .
There are several levels of preparation for disaster recovery, such as:
• Dump the whole DB2 environment.
This is a very cheap and easy way to prepare for disaster recovery, but it
can only be used to recover to the point in time when the dump was taken.
In the case when your disaster backup belongs to a point in time long before
the point of failure, you may lose a large amount of workload that had been
done. In order to obtain consistent data, you have to shut down DB2 during
the back.
• Create a system-wide point of consistency for more important table spaces
by using the QUIESCE utility.
For more information about DB2 Disaster Recovery, see DB2 Data Sharing
Recovery and Disaster Recovery Library - S/390 Technology Guide .
Figure 71 on page 204 shows the major components of a DB2 data sharing
group. The two DB2 subsystems accessing the same data are called members
of a data sharing group . They can be running on the same MVS image or
different MVS images in the same Parallel Sysplex.
All members of the DB2 data sharing group can read and update the same
tables at the same time. Therefore all data must reside on shared DASD. The
members of a DB2 data sharing group uses coupling facility services to
communicate and move data between the individual members.
The individual members of a data sharing group need the connection to the
coupling facility (CF) to exchange locking information, system and data object
status information, and to cache shared data pages with the other members of
the group.
During planned outages for maintenance or other work, your users can keep
working. You can shut down a subsystem while switching your users to another
subsystem in the group.
The group attachment name acts as a generic name for the DB2 subsystem in a
data sharing group. It substitutes for the DB2 subsystem name running on the
MVS the job was submitted from. Utility jobs can also use the group attachment
name. By using the group attachment name, the TSO user or batch job does not
You cannot use the group attachment for CICS and IMS applications because
those transaction managers must be aware of the particular DB2 subsystem to
which they are attached so they can resolve in-doubt units of recovery in the
event of a failure.
If you use IMS as transaction manager, you should define all DB2 subsystems in
the control region of IMS. Then you are able to connect message regions to
different DB2 subsystems. Before you stop the DB2 subsystem, you should stop
the message regions that the DB2 subsystem is connected to. Message regions
connected to another DB2 system will care of the new incoming transactions.
Type 2 indexes eliminate locks on index pages. Because there are usually fewer
rows to a data page than there are index entries on a index page, locking only
the data when you lock pages is likely to cause less contention than locking the
index. This way you can avoid most of the deadlock and timeout problems on
the index that often caused application abends because data was not available.
Another advantage of type 2 indexes is that they let you use other function such
as processing multiple parallel tasks, improved partition independence, row
locking, and the read-through locks. All of these functions also improve
concurrency in accessing data, and therefore avoid outages due to data not
being available to applications.
There are different kinds of applications that, in the past, caused or met data not
available condition during online processing. These applications, as listed, can
tolerate the previously described condition:
• A reference table, such as a table of descriptions of parts by part number,
because it is rarely updated.
• When, for security reasons, the updater is also the only reader of the table.
The access can be serialized within the application.
• All statistical kinds of applications; for example, if someone wants to do
statistical analysis on their employee data and reading an accasional
uncommited record cannot affect the averages much. A typical question
might be “What is the average salary by sex within education level?”
With the SHRLEVEL CHANGE option, the applications have read and write access
to the data being reorganized through almost the entire process. The process
involves these activities:
1. The utility unloads the data from the original tablespace, sorts the data by
clustering key, and reloads the data into a shadow copy. Concurrently,
applications have read and write access to the original tablespace and
changes are recorded in the log.
2. The utility reads the log records and applies them to the shadow copy
iteratively. During the last iteration, they have only read access.
3. The utility switches the application access to the shadow copy by renaming
the data sets for the table and the indexes. During the actual switch,
applications have no access to the data.
4. The applications read and write to the renamed shadow copy.
Figure 74 gives you a view of this process.
The utility has additional options that helps you to improve runtime. The
keyword SORTKEYS allows you to specify that sorting is to be done in memory,
rather than in writing and reading workfiles. You also may now have an image
copy taken at the time of loading or reorganizing the tablespace.
For example, the RECOVER utility can run against a single partition instead of
recovering the whole table space. Figure 75 shows an example of how a
partition tablespace can be accessed. The first partition is not accessed at all,
the second partition is stopped for any reason, the third partition is running an
SQL statement and the fourth partition is in recovery pending mode.
The OS/390 RRS can be used for an application that connected to DB2 using
RRSAF, and for DB2 stored procedures that execute in an MVS WLM-managed
stored procedures address space.
All applications designed so far have been “conscious” of this platform since
normally it is hardcoded in the application, for example, as in the 3270
communication protocol of an IMS application.
This new design methodology allows you to build applications that are
continuously available, thus you need to understand what Message Queueing is
and how it works.
12.2.1 Messages
A message is a string of bits and bytes that has meaning to one or more
application programs. It comes into existence when it is passed to the local
queue manager by the sending application ( putting the message ) and it ceases
to exist when the receiving application removes the message from the queue
( getting the message ).
The message-queueing service does nothing with the application data other than
to deliver it safely to the receiving application.
Control information: The data defines various properties of the message, and is
used by the message-queueing service to decide how the message should be
processed and it is returned by the queue manager to the receiving application.
Each queue manager is known by its name which, generally, must be unique
within the network of interconnected queue managers, so that one queue
manager can unambiguously identify the target queue manager to which any
given message should be sent.
12.2.3 Queues
A queue is a type of list that is used to store messages until they are retrieved
by an application. Its physical representation depends on the environment, but
can be as follows:
• A buffer or buffers in main storage
• A file or files on disk or other permanent storage device
• Both of the above
12.2.5.1 Triggers
Some applications run continuously, and so are always available to read a
message as soon as it arrives on the application′s input queue. However,
having the application active consumes system resources, even when the
application is waiting for a message to arrive. This additional load on the
system is not desirable when the volume of message traffic fluctuates wildly.
Instead of the application running continuously, it is better for the application to
run only when there are messages to be processed. The queue-manager′ s
triggering facility can be used to help make this happen.
The queue manager defines some conditions as trigger events. When trigger
events occur, the queue manager sends a trigger message to an initiation
queue, which will be processed by a trigger-monitor application in order to take
appropriate action. Normally this action would be to start some other application
to process the queue that caused the trigger message to be generated.
Triggering can be also controlled by the minimum priority that a message must
have in order to be able to cause triggering. If a message arrives with a lower
priority, the message is ignored for the purposes of deciding whether a trigger
message should be generated.
The message descriptor in all messages contains the coded character set
identifier (CCSID) and encoding information, which identifies the representation
used for the character and numeric data that is in the data portion of the
message.
Some MQSeries products can convert the data within messages from one
representation to another, and this operation is performed at the following times:
• By a queue manager during receipt of a message, if the data conversion
option is included in the request.
• By a message channel as it transmits a message to a remote queue
manager, if the data conversion option is included in the channel definition.
• Remote communication
With this type, if Programs A and B reside on different systems (that is, if one
is remote from the other), then they interact with different queue managers.
From the point of view of Programs A and B, there is no difference between
the situations described in the previous two points, because each program
interacts with only its local queue manager.
• Client-server applications
The previous points illustrate the simple communication of messages
between two programs. However, message queueing is also ideal for the
client-server type of application, where many client programs send
messages to a single server program. It is also possible for several
programs to get messages from the same queue, that is, to have several
servers processing the queue. This is useful in situations where the volume
of message traffic exceeds the processing capacity of one server; additional
servers can be started during peak hours to cope with the extra traffic, but
without the client programs needing to be made aware of this.
Figure 80 on page 218 shows a client/server structure.
These functions use data sets to keep track of ongoing activities, such as:
Log data sets also contain checkpoint records , which are used to reduce restart
time.
At a checkpoint, MQSeries logs its current status and registers the RBA of the
checkpoint in the bootstrap data set (BSDS).
For more information about CICS recovery, refer to CICS for MVS/ESA Recovery
and Restart Guide . For information about the CICS resource manager interface,
refer to CICS for MVS/ESA Customization Guide .
Queues are no exception to this; applications need to be able to get and put
messages (and possibly update other resources, such as databases), and know
that either all of the operations take effect, or that none of them does. The set of
operations involved in this is called a unit of work (UOW).
A unit of work starts when the first recoverable resource is affected. For
message queueing, this means when a message is put or retrieved as part of a
unit of work.
The unit of work ends either when the application ends, or when the application
declares a syncpoint , at which time any party that has an interest in the unit of
work has to commit about the work they did so far, and the changes that were
made as part of the unit of work become permanent. If any part that is involved
in the unit of work does not commit, the unit of work is backed out; that is, all
changes to data are discarded.
If the unit of work is ended by the application declaring a syncpoint, another unit
of work can then start, so that one instance of an application can be involved
with several sequential units of work. Figure 81 shows an example of UOWs and
syncpoints.
Triggering messages within a unit of work: If the message that caused the
trigger condition is put as part of a unit of work, the trigger message does not
become available for retrieval by the trigger monitor application until the unit of
work completes, whether the unit of work is committed or backed out.
Many applications that use units of work will also want to use persistent
messages. However, there are some situations in which it may be beneficial to
use one without the other.
For example, if the application contains logic to recover after a failure of the
queue manager, using units of work with nonpersistent messages gives a
performance benefit in the normal, nonfailure case. This combination can also
be used to ensure that a final (nonpersistent) message is not sent if the unit of
work is backed out before it reaches the syncpoint.
Syncpointing cannot occur directly across remote links, but if you use the
assured message delivery feature of MQSeries, a commit request is sent in a
message to a remote system, where it remains until it can be delivered. When
delivery eventually occurs, any desired commitments can then be made.
An in-doubt message is one where the point of consistency prior to the message
is known, but it is not known with certainty whether the new point of consistency
that the message creates has been reached, or that a backout to a previous
point of consistency has been completed.
The application must decide what action to take to resolve the in-doubt situation.
With message queueing, the exchange of messages between the sending and
receiving programs is also time-independent.
This means that the sending and receiving applications are decoupled so that
the sender can continue processing without having to wait for the receiver to
acknowledge the receipt of the message. The receiving application may be busy
or, indeed, it does not even need to be running. MQSeries holds the message in
the queue until it can be processed.
You can now design your applications such that they can improve the total
system availability by exploiting this new way of communication, as described in
the following section.
When the host is again available, and the accounting function of the application
is started up, the host will reconnect with the end-user workstation queue
manager to gather the data.
From the user perspective, the total application was continuously available
because it kept working even when the host site was unavailable. Indeed, the
user might be completely unaware of what happened.
One way to segregate message traffic is to use different transmission queues for
the different classes of message by using queue-manager aliases. The
applications themselves need not be aware that message traffic is being
Thus it is very important, in order to achieve a better availability, that all the
required data sets are duplicated and that the backup copies are also available
in case a restart is needed.
MQSeries does not provide a duplication of page data sets by itself, but these
sets can be duplicated using other techniques such as IBM 3990-6 Remote Copy.
This chapter describes how good systems management can help you achieve
continuous availability. It shows that availability cannot be bought as an
off-the-shelf technology solution, but that instead you need good processes and
tools to increase availability.
13.2 Introduction
Let us assume you have bought the latest generation S/390 processors and have
attached RAID DASD through ESCON and ESCON Directors in a fully
fault-tolerant configuration. You have also started to use techniques like Remote
Copy to ensure access to the latest copies of your data. You have introduced a
Parallel Sysplex to support critical business services. So, will these steps
automatically deliver continuous availability?
The answer is no !
The diagram indicates that continuous availability is built from two main
elements, high availability and continuous operation. These elements are then
Over the last few years, the growth of IT resource outside of the traditional
mainframe has multiplied the problem of effective systems management. It is no
longer acceptable to consider managing all of these resources independently.
They each support an important facet of your business and are increasingly
being combined for a service that needs two or more of these resources for its
delivery. In this distributed environment, an effective management solution must
also span the same resources. An integrated toolset that makes you as
comfortable controlling a S/390 Parallel Sysplex as an OS/2 Server from the
same terminal with the same guidance is no longer a luxury; indeed, continuous
availability cannot be achieved without it.
This chapter does not describe how to select and implement systems
management products or projects. It concentrates mainly on the S/390
13.3 OS/390
The foundation for good systems management is designed into OS/390 and its
optional elements. These in turn are built on the 20-year experience of MVS as
a commercial platform supporting many different businesses around the world.
OS/390 contains several key components that provide facilities for the additional
systems management products to connect to and interact with. These
components and elements include:
• System Services
JES
DFSMS
TSO/E
ISPF
Automatic Restart Manager (ARM)
Workload Manager (WLM)
Sysplex Failure Manager (SFM)
Coupling Facility Resource Manager (CFRM)
• Systems Management Services
Resource Measurement Facility (RMF)
System Modification Program (SMP/E)
System Display and Search Facility (SDSF)
Hardware Configuration Dialogs (HCD)
Recoverable Resource Management Services (RRMS)
IMS, CICS, and DB2 provide their own equivalents of these services, but can
make use of this OS/390 facility to extend commit processing management
across a distributed environment. It also means that batch programs can make
use of this and take regular syncpoints. This may make each job run slightly
longer, but in the event of a failure, the recovery is much faster. It may also
mean fewer tape copies, as the updates are being more carefully managed now
during the batch process, and regular incremental tape copies are no longer
needed.
HCM can also store information with each connection and unit that is not related
to the system definition. For example, this means that information like asset
Many products have and will be merged and then supported as a single
application. The offering keeps the TME name due to the fact that the products
will continue to be built around the object-oriented (based on CORBA standards)
Tivoli Management Platform. Some products still hold the SystemView reference
in their name, but this will disappear with future releases.
The Tivoli framework breaks systems management into four distinct areas:
• Security management
• Operations and administration
• Deployment management
• Availability management
For S/390 mainframe management, the products that fit in these categories are
shown in Table 13
NetView has been the core product of IBM′s mainframe and network
management for over 10 years, and its position has been kept in the Tivoli
portfolio. It provides the TME 10 automation services for S/390, and through the
TME 10 Global Enterprise Manager it can exchange events, commands, and
topology data with the TME 10 Enterprise Console and other TME 10 distributed
products. Many of the other S/390 management products build on the NetView
platform to extend its scope and capability as a manager. It delivers the function
to perform tasks like events management, topology management, and network
management on a single mainframe, a Parallel Sysplex, or a complex network.
AOC, along with the set of Automation Towers, aims to deliver a complete
automation solution for the mainframe environment. Instead of providing an
environment for you to develop automation routines, it delivers drop-in
automation for starting, shutting down, and re-starting MVS address spaces and
many other operator tasks. It also delivers an automation framework which
ensures that defined critical resources receive the correct levels of service.
These resources include things like systems, DASD volumes, or tapes.
Information is collected from RMF for them and used to drive automation. This
means that both performance-driven automation and event-driven automation
can be used to improve the management of your system.
ESCON Manager works dynamically with all OS/390 systems, HCD, the channel
subsystem microcode, and the installed hardware to build and maintain a picture
of any I/O device selected. It also works with other ESCON Manager
components on other OS390 images to obtain their view of the same hardware
components. For example, by selecting a DASD device, the display shows
whether the device is online or offline, shared between multiple systems,
whether it is attached to an ESCON channel or parallel channel, whether it is
attached through an ESCON Director, whether or not the ESCON channel is
shared between multiple LPARs, and so on. This information is displayed in a
set of colors and shaped icons to represent the different components. Selecting
individual icons allows more detailed information to be displayed. ESCON
Manager performs these tasks regardless of whether the component is attached
to a parallel channel or to an ESCON channel. It can also discover and display
information on other hardware items associated with a Parallel Sysplex (coupling
facility and links), and also the Open Systems Adapter (OSA) used on S/390
systems.
When the status of a device has been determined, ESCON Manager can manage
any change (vary offline or online) decided by the operator and then schedule
this change across all sharing systems. The scope of this vary command can
also be limited if required. It contains a very sophisticated checking system to
The Target System Control Facility (TSCF) extends the control, monitoring, and
automation capabilities of NetView. NetView provides tools to manage systems
that are up and running. With TSCF, it then has the ability to initialize, configure,
recover, and shut down systems, and to monitor multiple systems and respond
to a variety of detected conditions.
You designate one processor called the focal-point system to control your
complex. Each system that the TSCF focal-point system manages is called a
target system. Target systems can be placed at dial-up line distances, so the
extent of control is potentially large. Some examples of remote control provided
by TSCF are:
• Perform a power-on reset of a target system
• Perform an initial program load (IPL) of a target system
• Set the target system′s time-of-day (TOD) clock
• Synchronize clocks on all target systems
• Configure the target system
• Detect and respond to target system messages
• Detect disabled console communications facility (DCCF) situations
• Detect wait states at the target system
• Use commands to simulate console keys
• Shut down the target system
These functions require a connection to the hardware management console of
the processor being controlled, and also a connection to any LPARs that need to
be monitored and controlled. Many different types of connection are possible to
perform this remote connection including SDLC links and LANs. TSCF Planning
and Installation gives detailed information of the different configurations that
TSCF can use.
Service Level Reporter Version 3 (SLR) will be maintained for managing and
reporting of systems measurements data in a non-DB2 environment.
NPM can perform the data collection task from VTAM on the mainframe out
across the network, and with PR/MVS this can be interpreted and displayed to
help you keep track of resources.
From a single point of control, OPC analyzes the status of the production work
and drives the processing of the workload according to installation business
policies. OPC also includes a graphical user interface for application description
that can significantly improve productivity. It supports a multiple-end-user
environment, enabling distributed processing and control across sites and
departments within the enterprise.
ADSM also offers help in the area of disaster recovery. When preparing for the
possibility of severe service interruptions, you need to maintain a backup copy of
your vital business data offsite (offsite recovery media). When disaster strikes,
you need to know where your backup data is and what to do with it (recovery
plan) and what your vital business applications and their requirements are
(client recovery information).
TME 10 GEM is the first software specifically designed to manage the strategic
business systems that run your company and make it competitive. It provides
management for business-critical applications that span both the S/390 and
distributed environments. The management and control is presented in a
business context and features:
• Application Policy Management, a Java-based application for managing
business systems that does the following:
Provides a business system view of all resources required to deliver
specific business services
Allows control of components of each business system with
cross-platform, single-action management
Automatically discovers and builds graphical views of the business
system components and their relationships
Allows a view of the business system from the standpoint of a specific
management discipline
Provides pre-built business views of TME 10 systems management
products and of Lotus Notes E-mail systems when used in conjunction
with TME 10 Module for Lotus Domino/Notes
Works with existing applications and management tools instead of
replacing them
• TME 10 GEM Management Integration Services, a bi-directional
communication between management centers of both S/390 and distributed
environments
In these cases, a product such as IBM′s Smartbatch for OS/390, which reduces
the elapsed time of batch jobs without requiring a major redesign or rewrite, is
of real benefit in improving availability.
Even an installation with single processor may be able to find some advantage
from Smartbatch. The ability to run job steps or jobs at the same time, and I/O
reduction and optimization can be performed on a single processor. The overall
gain, however, may be dependent on CPU stress and storage use at the time of
the job run.
BatchPipeWorks
Yes Yes Yes Yes Yes Yes Yes Yes
Commands
• This function is not automatic; your installation must manually modify JCL
and job scheduling to some extent, such as converting job steps to jobs.
Note: The source of this data is IBM Smartbatch for OS/390 Overview , GC28-1627
There is no conflict here with TME 10 OPC, since OPC controls when a job is
being presented for processing, and Smartbatch influences the way that a job
processes when released. The products are complementary and are both of
benefit for higher availability.
For unplanned outages, WLM can redistribute the arriving work around the
remaining members of the sysplex. Then the dispatching priorities, storage
allocations, and other parameters affecting individual workload performance can
be rearranged to compensate for this extra work. The net result is continued
service for the end user, and with a transaction response as close to their
requirements as capacity allows. No other system complex has this ability to
react in such a positive way to major events without manual intervention.
For a scheduled shutdown of a major component, WLM works in the same way,
but in this case the changeover is probably much less dramatic. You will
probably use a management product like CP/SM to stop workloads from
processing in a much more controlled manner. This also makes the
In both cases, when the resource is ready to be returned to the sysplex, WLM
allows it to be brought into use transparently. Workload is sent to the new
resource and the workload automatically rebalances across the active members
of the sysplex.
An interesting side effect of WLM is the protection from “hung” systems resulting
from a CPU-intensive job being released at the wrong time. Because of the way
that WLM works, it prevents such a job from completely taking over the
processor on which it has been released. This means that although
performance will certainly be impacted on this processor, other work will
continue to process successfully, and the console will not be locked out. This
may not be a frequent occurrence in your installation, but preventing even one
such hit may be worthwhile.
From the high availability point of view, the key functions of internally detecting,
analyzing, and reacting to events that may cause a loss of service are essential.
This can be done without reference to an external product, although generic
alerts for NetView can still be flagged. If the event cannot be handled internally,
then NetView may be the normal route to signal that additional handling is
required. CP/SM can also use the ARM component of OS/390 to handle the
restart of a CICS region according the the ARM policy that is active.
Another important function of CP/SM is the single point of control which allows a
view of all CICS systems and resources no matter which processor or image
these resources are running on. This is essential for the human interaction that
cannot be avoided in day-to-day work. The task of trying to work out what is
running where is taken away, leaving it much easier for people to understand
and manage CICS MRO operations.
To solve the problem of matching business needs with subjective perception, the
concept of Service Level Agreements (SLA) was introduced.
First of all, you will have to define, for each on-line service, your own
requirements. This requirements, which can be called service level objectives ,
can include:
• The hours of service
• Maximum/minimum response time for different applications
• Maximum number of outages per time interval
• Mean Time To Repair (MTTR) and recovery time
• Process or print turnaround times
These service level objectives serve as input for determining your data
processing requirements. The service level objectives will have to be negotiated
with the end users and eventually be turned into Service Level Agreements. It is
recommended that objectives be documented in existing service level
agreements, or that new agreements be created if none exist.
Similar approaches can be used for both. Here are some examples:
• Rules of thumb
• Comparison with other systems
• Parametric methods
− A transaction profile (10 read calls, two update calls, eight physical I/Os)
− Cost of function (CICS: 15ms per physical I/O in a 3390 non-cached)
• Analytic (queueing theory) models (for example, the IBM capacity planning
tool CP90)
• Simulation using a computer program that has the essential logic of the
system (for example, the Snap/Shot modelling system from IBM)
• Benchmarks, Teleprocessing Network Simulator (TPNS)
In addition, performance and capacity work tend to require similar input data
from the system. In particular, both require resource consumption by workload
data
In addition, you can use the Spreadsheet Converter for further processing the
measurement data on a programmable workstation with the help of spreadsheet
applications.
Performance Reporter for MVS is a systems management tool that collects and
transforms detailed measurement data into meaningful reports that focus on
such key business areas as availability and performance management, and
service level reporting. This information is available at various levels (user,
application, subsystem, processor and network). Reports are available for
management and technical specialists.
Performance Reporter for MVS′s use of DB2 provides easy access to the stored
performance data, and makes access to distributed data possible. Performance
Reporter for MVS significantly improves user productivity and ease of
implementation through powerful dialogs and online help facilities. Based on the
user′s definition, Performance Reporter for MVS selects which data to collect
and how long to keep it. It also includes a set of reports that address the user′ s
requirements.
Performance Reporter for MVS is the follow-on product to Service Level Reporter
(SLR, 5665-397). Service Level Reporter Version 3 will be kept up-to-date for
supported products and is the preferred solution for managing and reporting of
systems measurements data in a non-DB2 environment.
By selecting the areas and level of detail that need attention, you can set
objectives for system management areas such as:
• Service levels
• Availability
• Performance management
• Capacity planning
• Operations
• Problem and change management
• Accounting
To produce these reports, SRL tracks the availability of hardware and software
starting at the processor and ending at the users′ terminals.
Availability data, to be most meaningful, must show whether a user service was
available when it was supposed to be. SLR availability reporting gives you this
information. If you enter your availability schedules in SLR, your reports will
show availability during these scheduled periods.
13.8 Summary
It would be convenient to be able to draw a map at this point to chart progress
and directions on the road to systems management suitable for supporting
continuous availability. Unfortunately, it is not so simple.
The OS/390 base and optional products are an obvious first and second step.
The IBM Open Blueprint shows the systems management facilities, along with
the operating system resources, as forming the backbone to support all other
Using NetView is a good third step, with its mainframe and network monitoring
and control abilities, but from this point on the priorities are probably personal.
Products like ADSM, OPC, and System Automation may seem obvious next
targets, but many would argue that the creation of a sound problem and
management database on Info/Man is the next best step (assuming you do not
already have an effective one in operation).
The important point is that all of the topics in Table 12 on page 229 need to be
covered by a process, ideally an analytical, reactive, system-based process, in
order to achieve the combination of high availability and continuous operation.
This chapter highlights some of the IBM S/390 software products that do this.
In the dedicated computing model of the 1970s and 1980s, storage requirements
were dominated by centralized data and centralized management within an
environment of homogeneous platforms.
In the late 1980s and early 1990s, the client/server computing model distributed
data across the enterprise on a wide variety of heterogeneous multi-platform
environments. This raised new requirements for accessing the same data from
this heterogeneous world. These requirements were usually addressed by
copying the host (server) data and transmitting it to the different distributed
platforms across the enterprise.
Universal data access requires not only the physical connection and
communication between many types of technologies, but also the ability to
actually make sense of the data once you get to it. Therefore, universal data
access means being able to get to any data anywhere at any time, while
protecting data integrity.
While there are many products that support data copy sharing, both storage
sharing and data sharing among heterogeneous platforms cannot be achieved
by any of the S/390 devices available today. However, we will also discuss these
ways of sharing data in order to have a better understanding about the possible
evolution of the products.
Figure 87 shows an example of such a file transfer. Id 15is example, both the
source system and the destination data were unavailable for about two hours
each. Moreover, at the end of the process, the destination system application is
working on a data image that is as old as the duration of the entire process (14
hours).
Moreover, the combination of CNT FileSpeed with IBM RAMAC Virtual Array
SnapShot high-speed database replication unlocks the OS/390 database for
extraction without impacting normal batch or online processing at the point in
time when the service is required.
In the previous example, the database data extract operation took two hours and
the S/390 DB was unavailable for use by batch or online applications during that
entire length of time.
In contrast, while this operation still takes two hours with SnapShot, it does not
require taking the S/390 DB offline, thus improving the S/390 data and
application availability.
Compared with the example in 14.2.1, “File Transfer” on page 257, this solution
has not improved the availability of both systems, but the image of the data in
this case is only four hours old.
Also in this case, a copy of the source database using the SnapShot function of
IBM RAMAC RVA allows the extract program to work on an immediate copy of
the source database, without requiring you to take both the database and the
application offline.
Data transmission is controlled by API functions on both the S/390 side and the
RS/6000 side. The AIX API allows the application to “think” it is writing data to a
local tape, but actually the data is written to tape or disk storage on the S/390 at
very high ESCON speeds.
Again, the ability to snap a copy of the source database using the IBM RVA
SnapShot function not only allows continuous availability for both data and
application at the source side, but at the end of the process, the data at the
destination side is only two hours old.
Using storage sharing is a good way to lower the price and total cost of
ownership. However, at the moment, there are no devices available to attach
S/390 and the other platforms directly.
Real data sharing means that all applications are able to read the other′s data
structure, or they use a common data structure, but they must also guarantee
data integrity on write operations.
In a shared environment, the multiple copies of the database manager uses two
methods to provide this guarantee: locking, and buffer consistency. These are
explained as follows:
There are basically five categories of building blocks: RISC processors and
memory, network and platform attachments, storage adapters, storage building
blocks (such as disks, tapes, and libraries), and software.
Furthermore, it was stated in the announcement letter that IBM also plans to
improve current S/390 DASD devices by extending the RAMAC array to support
UNIX and Windows NT data, as well as planning an open systems disk array that
can store UNIX, AS/400, and Windows NT data simultaneously.
So far we have limited our scope to the “system side” of continuous availability.
We stressed the need to select components with appropriate C/A attributes, the
need to exploit these attributes when necessary, the need to design the system
for availability, the need for system-wide automation, and the need to measure
and manage availability.
It is obvious, just as with the system, that applications must also be designed
and implemented to reduce or eliminate the duration and frequency of both
scheduled and unscheduled outages, and to limit their scope.
In this section we briefly review application and development, and provide tips
on exploiting the availability attributes of the S/390 line of products in order to
improve application availability.
In the following list we discuss four of the most important criteria: correctness,
robustness, extendibility, and reusability.
• Correctness
On the other hand, correctness and robustness can only be achieved with a
quality-controlled development process, coupled with an advanced testing
process.
CICS/ESA, since Version 3.1, has been substantially re-engineered and has an
object-based structure. It is structured in domains , where each domain (also
referred to as a module or an object ) contains a data structure , that is, control
blocks and the function required to change their contents. A domain can send
and receive service requests to/from other domains by using a central
dispatching domain. This architecture reduces the interfaces to the minimum n-1
if the number of domains is n .
For this reason, the application testing should be as near as possible to the
production conditions that the application must sustain.
As the last phase of testing, you also need to establish a test environment for
future changes; the following items can help you create such an environment:
Testing with CICS EDF: The CICS Execution Diagnostic Facility (EDF), introduced
in CICS/VS V1R3 and enhanced in CICS/MVS and CICS/ESA with a major subset
in CICS OS/2, is an application development tool that can be used to test
command-level programs interactively, without having to modify the code or
preparation procedure.
With the enhancements to the CICS attachment code in DB2 V2R3, the
programmer can now also see information both before and after each SQL
statement. The use of the debugging capability in facilities such as EDF will
potentially result in more thorough testing and more rapid debugging, thus
leading to faster problem resolution and a decrease in future software-related
outages.
These can only be achieved if your application design supports such features as
redundancy, failure detection, and recovery or bypass algorithms.
15.3.1 Redundancy/Modularity
Redundancy is the main feature that guarantees your having spare components
that can take over the failed component′s activities.
Where it is not possible to have redundant components, you can use modularity
to guaranteed that only a subset of the total application functions are affected by
the failure.
Modularity by itself does not provide fault tolerance for the single failed
component, but it is useful in limiting the impact of the problem, from the
perception of the user.
For more information about these subjects, refer to 12.5, “How MQSeries
Contributes to Continuous Availability” on page 222.
Code Failure: In case of code failure, the detection algorithm must provide all
information useful in isolating and identifying both the failure and the component
that is affected.
Data Failure: In case of a data failure, the detection algorithm, before reporting
the problem, may have to identify the reason for the failure. The failure may be
due to:
• Security problems (insufficient authority)
• Physical problems (a hardware error, data corruption, and so on)
• Logical unavailability (data locked by others, and so on)
15.3.3 Recovery/Bypass
This feature allows your application to do a fast recovery or bypass of the
problem; in this case also, we can separate the algorithms into two main parts
as follows:
When read operations are unsuccessful, your application design should evaluate
the following possibilities:
• Are there any alternate sources of data?
• Are there any default values available?
• Can a subset of the business process be satisfied without the data?
• Would a day-old copy of the data be satisfactory or useful?
15.4.1 Isolation
During your design, you must define very strictly the interface protocols and data
structures that these applications use to communicate both with others and
within themselves.
Moreover, your application design should consider being able to isolate the
applications from external dependencies, including operator interventions,
because the service that your application needs may:
• Be unavailable when invoked
• Be unable to process your request
• Fail while processing your request
• Never reply to your request
As discussed in 15.3.2, “Failure Detection” on page 272, in this case also your
design must interface with problem and availability management in order to
report any problems that are encountered.
For more information about Message Queueing, refer to 12.5, “How MQSeries
Contributes to Continuous Availability” on page 222.
You may also take advantage of the following strategies when designing for
Application Isolation.
SSSP can be used by CICS/ESA to prevent CICS system code, control blocks, or
data from being overwritten by application code. It enables subsystems to
isolate the failure at the application level.
Note: CICS applications can take instant advantage of this feature without any
modification.
Using SSSP reduces the the time required to diagnose errant application
programs that might potentially overwrite CICS system storage, to solve the
problem, and to recover the application. SSSP can also help application
programmers identify potentially faulty programs before they are put in
production. This is a fault-avoidance attribute at the application level.
During the design phase, you should try to provide your application with features
that allow you to dynamically modify its environment, parameters, and data.
End User: The considerations include both logical and physical views of the
end-user. At a minimum, your design should consider:
• Adding/deleting authorized users
• Changing a user′s authority
• Changing a user′s location
• Terminal types and software levels
• Terminal location and hours of operation
• Terminal connection protocols and speeds
Time, Date, and Parameters: These considerations include all paramenters that
normally affect applications, such as:
• Date and time
• Year 2000
• Timezone and daylight saving anomalies
• Clock corrections
• Application-related parameters without restart
Application Data: Finally, your design should also provide the ability to
dynamically change all or part of the application data. This can be achieved by
both IMS and DB2, using the Database Partitioning feature. For more
information, refer to 10.5.1, “Fast Path DEDB Online Change” on page 194, and
11.4.5, “Partition Independence” on page 208.
Use DB2 Package Binding and Versioning: The package concept, introduced
since DB2 2.3, is a result of binding a Database Request Module (DBRM).
A DBRM is one of two outputs from the DB2 pre-compiler, and it contains the
SQL statements for a given program module. The new application plan can now
consist solely of a list of the packages (bound DBRMs) that are used by the
application. This three-level hierarchy, DBRM-package-plan, makes the
re-binding process which results from SQL changes to the program much less
disruptive.
With the package bind, only the changed DBRMs need be rebound. In addition,
with the support of multiple versions of packages (DB2 2.3), the binding of
changed DBRMs can be done ahead of time, without disrupting the production
application.
Multi-versioning support also eliminates the need to re-bind in order to fall back
to a previous version if problems arise with modules newly introduced into
production. The package bind, together with the multi-versioning support,
provides very significant improvement in the continuous operation capabilities of
DB2 applications.
However, to be transparent to the user, you need also to apply VTAM Persistent
Session Support, which provides you with dynamic reconnection of the end user
to the application. With this feature, the end user perceives the application
switching as a “wait” on the terminal.
For more information on XRF, VTAM Generic Session, and their usage with CICS
and IMS, refer to 9.4, “CICS Parallel Sysplex Exploitation” on page 171 and 10.4,
“IMS Parallel Sysplex Exploitation” on page 190
VSAM Record Level Sharing: Using VSAM record level sharing, you can share
VSAM files between CICS online transactions and batch processing. For more
information on VSAM Record Level Sharing, refer to 9.4.1, “VSAM Record Level
Sharing” on page 171.
IMS and BMPs: Since IMS/ESA 3.1, DBCTL makes batch message processing
(BMP) available for concurrent updating of data by both online and batch
programs. IMS BMP is used to perform batch-type processing online.
IMS Data Sharing: IMS Data Sharing offers an alternative approach to the
concurrent sharing of IMS data between batch and online. In addition, it
provides the ability to share IMS data between two separate IMS TM and/or
CICS subsystems, which can reside in separate MVS images.
Since most batch jobs use sequential data sets and are I/O-bound, sequential
data striping should help reduce the batch window and enhance the continuous
operation and continuous availability characteristics of the service provided by
an installation.
Smartbatch for OS/390: IBM′s Smartbatch for OS/390 helps reduce the batch
window by using techniques to speed up the flow of batch work through a
processor or a group of processors. These techniques are parallelism, I/O
optimization, and workload distribution.
For more information, refer to 13.5.1, “Smartbatch for OS/390” on page 243.
Fault-tolerant systems are often designed so that they will not stop working if
they have an internal component failure or if one of their devices suffers a
failure. This not only requires a suitable degree of redundancy, but it requires
constantly checkpointing work as the state of the work being done changes so
that in the event of a failure, that work can be rescued by a back-up facility.
The publication Fire in the Computer Room, What Now? covers all the design
aspects of disaster recovery, while Disaster Recovery Library - S/390 Technology
Guide describes all the S/390 technologies available and how you can apply
them to implement the recovery solution.
In a disaster situation, users normally are aware that an outage has happened to
the central computer facility, and the duration of the outage is mainly dependent
on the recovery solution.
At SHARE 78, Anaheim, 1992, in session M028, the Automated Remote Site
Recovery Task Force presented seven tiers of recoverability, which were ranked
For example, some installations can tolerate resuming their services after longer
periods of time, but with maximum data currency. Other installations must
resume their services as soon as possible, regardless the currency of their data.
Still others need both a short recovery time and maximum data currency.
Figure 105 on page 286 illustrates the data loss and service loss aspects of each
recovery solution.
While you must of course be prepared to react to a disaster, the solution you
would apply in this case may be more of a recovery solution than a continuous
availability solution. A recovery solution can be defined by making a trade-off
among implementation costs, maintenance costs, and financial impacts resulting
from performing a business impact analysis of your business.
Furthermore, as you can see from the tiers definitions, only a Disaster Recovery
Tier 6 solution can be compared to a continuous availability solution, although
currently there is no technology available that fulfill that definition.
These two (or more) buildings should also be located in such way that they will
not be impacted by the same disaster; for example if they are located 100 km
apart, but are situated next to the same river, it is likely they could both be
damaged by the same flood.
These two (or more) buildings must also be connected with high bandwidth
connections to provide device sharing among all processors. The best way to do
this is to use the flexibility provided by ESCON channels and ESCON Directors
together. Moreover, you can limit the number of fiber connections between the
sites by using the IBM 9729, which lets you share the same fiber trunk among 10
ESCON channels.
For more information about these connections, refer to Chapter 3, “S/390 ESCON
Architecture” on page 35.
The last consideration concerns the distance between the sites. The further
apart the sites are, the longer it takes to access data since the signals travel at
a fixed speed.
For more information on PPRC and P/DAS, refer to 6.3.2.4, “Remote Copy” on
page 109.
Under these conditions you can afford a controlled workload and processing
transfer between the two sites. However, in a disaster situation, the time that
your operators need to perform this transfer might not be available.
The cost associated with such a solution is very high and few installations can
justify it by having very stringent recovery needs, even in case of disaster.
However, the S/390 platform provides you with this solution by using the same
technologies you may already be using every day in your business.
Next you should map each business process to the I/S applications that support
it, and identify how important the application is to that business process.
Information in this book was developed in conjunction with use of the equipment
specified, and is limited in application to those specific hardware and software
products and levels.
IBM may have patents or pending patent applications covering subject matter in
this document. The furnishing of this document does not give you any license to
these patents. You can send license inquiries, in writing, to the IBM Director of
Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785.
Licensees of this program who wish to have information about it for the purpose
of enabling: (i) the exchange of information between independently created
programs and other programs (including this one) and (ii) the mutual use of the
information which has been exchanged, should contact IBM Corporation, Dept.
600A, Mail Drop 1329, Somers, NY 10589 USA.
The information contained in this document has not been submitted to any
formal IBM test and is distributed AS IS. The information about non-IBM
(″vendor″) products in this manual has been supplied by the vendor and IBM
assumes no responsibility for its accuracy or completeness. The use of this
information or the implementation of any of these techniques is a customer
responsibility and depends on the customer′s ability to evaluate and integrate
them into the customer′s operational environment. While each item may have
been reviewed by IBM for accuracy in a specific situation, there is no guarantee
that the same or similar results will be obtained elsewhere. Customers
attempting to adapt these techniques to their own environments do so at their
own risk.
Any pointers in this publication to external Web sites are provided for
convenience only and do not in any manner serve as an endorsement of these
Web sites.
The following document contains examples of data and reports used in daily
business operations. To illustrate them as completely as possible, the examples
contain the names of individuals, companies, brands, and products. All of these
names are fictitious and any similarity to the names and addresses used by an
actual business enterprise is entirely coincidental.
AnyNet ACF/VTAM
ADSTAR AIX
APPN AS/400
BatchPipes CICS
CICS OS/2 CICS/ESA
CICS/MVS CICSPlex
DB2 DFSMS
DFSMS/MVS DFSMSdss
DFSMShsm ES/9000
ESA/390 ESCON
ESCON XDF First Failure Support Technology
FFST GDDM
Hiperbatch IBM
IMS IMS/ESA
Magstar Multiprise
MQ MQSeries
MVS/ESA MVS/SP
MVS/XA Net.Data
Network Station NetView
Open Blueprint OpenEdition
OS/2 OS/390
Parallel Sysplex Predictive Failure Analysis
PR/SM PROFS
RACF RAMAC
RMF RS/6000
S/370 S/390
S/390 Parallel Enterprise Server Seascape
SecureWay Sysplex Timer
System/390 SystemView
SP Ultrastar
VM/ESA VSE/ESA
VTAM 400
Microsoft, Windows, Windows NT, and the Windows 95 logo are trademarks
or registered trademarks of Microsoft Corporation.
The publications listed in this section are considered particularly suitable for a
more detailed discussion of the topics covered in this redbook.
IBMMAIL Internet
In United States: usib6fpl at ibmmail [email protected]
I n Canada: caibmbkz at ibmmail [email protected]
Outside North America: dkibmbsh at ibmmail [email protected]
• Telephone Orders
This information was current at the time of publication, but is continually subject to change. The latest information
for customers may be found at http://www.redbooks.ibm.com/ and for IBM employees at
http://w3.itso.ibm.com/.
Company
Address
We accept American Express, Diners, Eurocard, Master Card, and Visa. Payment by credit card not
available in all countries. Signature mandatory for credit card payment.
Index 299
Dynamic Enterprise Systems Connections
Application update with XRF 277 and Geographically Dispersed Parallel
Coupling Facility Dispatching 11, 17 Sysplex 287
DASD address switching (P/DAS) 111 Error Correction Code 11
ESCD port reassignment 46 ESCD
ESCD port upgrade 46 see ESCON Director
I/O Reconfiguration Management (DRM) 13 ESCM
Memory Sparing 12 see ESCON Manager
SAP Sparing / Reassignment 11 ESCON
Storage Reconfiguration (DSR 2) 11 see Enterprise Systems Connection
Transaction Balancing 169 ESCON Channel Extender
Transaction Routing 168 9036 and PPRC 51
Workload Balancing 147 models 51
Dynamic disk reconstruction 118 ESCON Converter
types 50
ESCON Director
E ESCON Manager 236
ECC 11 ESCON Manager 236
eliminating batch window 278 ESCON Multiple Image Facility
EMIF sharing ESCON channels 39
see ESCON Multiple Image Facility sharing ESCON CTCs 40
Enterprise Systems Connection ESCV
9032 46 see ESCON Converter
9033 46 Extended Distance Feature
9034 50 Extended Recovery Facility (XRF)
9035 50 recovering CICS 174
9036 and PPRC 51 recovering IMS 188
9036 ESCON extender 51 recovering MQSeries 219
9036 models 51 Extended Remote Copy
9729 carrying ESCON channels 51 see XRC
Byte Multiplexer Channel 38 External CICS Interface (EXCI) 164, 167
CBY 38
chained ESCDs 37
channel path configuration 44 F
CNC 38 failure detection 272
configuring channels for availability 41 code failure 272
Converter Channel 38 FFST 108
CTC 38 Fiber Optic
CTC path configuration 45 advantages 35
CVC 38 ESCON Channel 35
Distance 37 multi-mode 37
error checking 35 single mode 38
ESCD control 56 Forward Recovery
ESCD path availability configuration 51 IMS 187
ESCON Channel Types 38 Function Shipping (FS) 164, 166
ESCON Converter 50
ESCON CTCs with EMIF 40
introduction 35 G
LED card 37 Global Enterprise Manager 241
Non-disruptive change 38 Global Resource Serialization (GRS)
Parallel Channel Conversion 38 ring configuration 60
PPRC and ESCON Directors 50 star configuration 60
Serial Channel 35
sharing ESCON channels with EMIF 39
Speed 37
H
Hardware Configuration Dialog
Topology 36
changing ESCON Channel types 38
use of ESCON Manager 57
ESCD control 56
use of HCD 38
XDF card 38
Index 301
MQSeries (continued)
J Recovery Features (continued)
Java Page data sets 218
CICS/ESA internet gateway 87 recovery with XRF 219
Syncpoints 220
two-phase commit 221
L Unit of works 220
LOGR couple dataset 146
Remote queue 214
Reply-to queue 214
M Simple and Staged Transfer 215
Message and Queueing Syncpoints 220
see MQSeries Temporary Dynamic queue 214
Microprocessor Complex PCI Card 21 Transmission queue 214
Migration: DASD 132 Trigger events 215
MQSeries Triggers 215
, and CICS recovery 219 two-phase commit 221
, and IMS recovery 219 Undelivered message queue 214
, Extended Recovery Facility (XRF) 219 Unit of works 220
Alias queue 214 , and message persistence 221
Bootstrap data set 219 getting messages 221
Checkpoint records 219 putting messages 220
Command queue 214 triggering messages 221
Data conversion 216 Multi-Region Operation (MRO) 164
Dead-letter queue 214 Multiprise 2000 20, 99
Distributed Queue Management (DQM) 215 Internal disk 21
Simple and Staged Transfer 215 Muxmaster
Event queue 214 see Wavelength Division Multiplexer
Initiation queue 214
Internet gateway 88
Local queue 214
N
Net.Data
Log data set 219
DB2 WWW connection 87
Message channel agent 215
NetView for OS/390 235
Message channels 215
NetView Performance Monitor 238
Message queue 214
Network Computing
Messages
access from CICS applications 182
Application data 212
AIF gateway 88
Control Information 213
CICS/ESA gateway 87
Expiry time 213
CICS/ESA gateway for Java 87
Message identifier 213
connection from IMS 197
Message persistence 213
DB2 WWW connection 87
Message type 213
e-Business 94
Priority 213
IMS WWW Connection 88
Model queue 214
Internet Connection Secure Server (ICSS) 64, 90
Page data sets 218
MQSeries gateway 88
Permanent Dynamic queue 214
S/390 Gateways 86
Programs communication 216
Secure Electronic Transaction (SET) 90
Bidirectional 216
Network File System (NFS) 61
Client-server 217
Remote 217
Unidirectional 216 O
Queue Managers 213 object-oriented programming
Queue Types 214 application design 269
Queues 213 data abstraction 270
Recovery Features 218 OpenEdition
Bootstrap data set 219 Setup Verification Program (SVP) 89
Checkpoint records 219 OpenEdition Hierarchical File System (HFS) 61, 80
CICS recovery 219 Backuping up Open HFS Data sets 82
IMS recovery 219 Distributed File Service (DFS) 61
Log data set 219
Index 303
OSA-2 (continued) Peer-to-Peer Remote Copy (continued)
Load Balancing 25 P/DAS 111
OSA/SF 26 P/DAS Control Bloc 112
Redundant/backup connections 25 PPRC 100
SNA/APPN Networking Protocol 24 PPRC and ESCON Directors 50
TCP/IP Networking Protocol 23 shsring paths over chained ESCDs 55
virtual IP addressing 24 Peer-to-Peer Remote Copy (PPRC) 74
OSA/SF 26 Performance Management 250
Performance Reporter for MVS 237, 251
planning, Continuous Availability 1
P PPRC 104, 110, 133
P/DAS 111 see Peer to Peer Remote Copy
and geographically dispersed Parallel Sysplex 287 PPRC Dynamic Address switching
Parallel Channel 35 see P/DAS
Parallel Sysplex PR/SM
9037 sysplex timer 143 Dynamic Storage Reconfiguration 11
asynchronous CF requests 159 EMIF 12
capacity 138 Time Management Reporting 12
CFRM policy overview 246 UNIX Processes 86
components 139 Up to 15 LPARs 12
configuration discussions 147 Processor 5
coupling facility 140 Processor Availability Facility (PAF) 11
Coupling Links 142
CP/SM 247
datasharing 137 R
daylight saving considerations 145 RAID 98
EPSO service offering 153 Berkeley 98
Geographically dispersed 287 History 98
ICMF 156, 159 Level 0 99
in a Parallel Sysplex 147 Level 1 99
Internal Coupling Facility use 141 Level 5 100
Internal Coupling Migration Facility use 141 Level 6 101
introduction to 137 RAMAC 3 Array 106
monoplex 160 RAMAC 3 Array: microcode upgrade 118
non-disruptive change 151 RAMAC Array Family 105
SFM 151 RAMAC Array subsystem 99
sharing DASD with GRS 161 RAMAC Drawers 101, 117
Single System Parallel Sysplex 154 RAMAC Scalable Array 99, 101, 128
synchronous CF requests 159 RAMAC subsystem 120
Sysplex Couple Datasets 146 RAMAC Virtual Array 99, 121
Sysplex Failure Manager 245 Deleted Data Space Release 125
systems management 244 Disk Array 123
testing 152 Internal Structure 121
WLM 157 Log-Structured Array 125
WLM overview 245 SnapShot Copy 127
WLM policy 151 Recoveable Resource Manager
workload balancing 137 and IMS recovery 197
PARMLIB Recoverable Resource Management Services 233
Concatanation 60 Recoverable Resource Manager
Symbolic Pre-processor 61 and DB2 recovery 209
Partial I/O Restart 13 recovery
Partial Memory Restart 13 code failure 272, 273
PC Server System/390 21 data failure 273
Peer to Peer Remote Copy Recovery Tiers
ICKDSF 61 tier 0 - No Offsite Data 282
Peer-to-Peer Remote Copy tier 1 - PTAM 282
9036 and PPRC 51 tier 2 - PTAM + Hot Site 283
and geographically dispersed Parallel Sysplex 287 tier 3 - Electronic Vaulting 283
Data Flow 110 tier 4 - Active Secondary Site 283
Index 305
Tivoli Management Environment 10 (continued)
Performance Reporter 237
System Automation for OS/390 236
TME framework 234
Transaction Manager
see CICS/TS
Transaction Routing (TR) 164, 166
U
Unix
Services in OS/390 63
Verifying Availability 89
V
VTAM Generic Resource
and CICS support 173
and IMS support 194
and TSO support 62
VTAM Persistent Session 64
and CICS support 174
W
Wavelength Division Multiplexer
and Geographically Dispersed Parallel
Sysplex 287
Wavelength Division Multiplexor
backup fibre 51
carrying ESCON channels 51
WLM couple dataset 146
Workload Manager
DB2 Attachment Facility 182
Internet Connection Secure Server (ICSS) 92
policy 151
Workload Manager (WLM)
OS/390 Availability Enhancements 62
X
XCF couple dataset 146
XRC 74, 100, 105, 114, 133
Data Flow 116
XRF
see Extended Recovery Facility
Your feedback is very important to help us maintain the quality of ITSO redbooks. Please complete this
questionnaire and Fax it to: USA International Access Code + 1 914 432 8264 or:
• Use the online evaluation form found at http://www.redbooks.ibm.com
• Send your comments in an Internet note to [email protected]
Please rate your overall satisfaction with this book using the scale:
(1 = very good, 2 = good, 3 = average, 4 = poor, 5 = very poor)
Overall Satisfaction ____________
Was this redbook published in time for your needs? Yes____ No____
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
_____________________________________________________________________________________________________
IBML
SG24-2086-00