Red Hat Enterprise Linux-5-Cluster Administration-en-US PDF
Red Hat Enterprise Linux-5-Cluster Administration-en-US PDF
Red Hat Enterprise Linux-5-Cluster Administration-en-US PDF
Cluster Administration
Configuring and Managing a Red Hat Cluster
Cluster Administration
The text of and illustrations in this document are licensed by Red Hat under a Creative Commons
Attribution–Share Alike 3.0 Unported license ("CC-BY-SA"). An explanation of CC-BY-SA is available
at http://creativecommons.org/licenses/by-sa/3.0/. In accordance with CC-BY-SA, if you distribute this
document or an adaptation of it, you must provide the URL for the original version.
Red Hat, as the licensor of this document, waives the right to enforce, and agrees not to assert,
Section 4d of CC-BY-SA to the fullest extent permitted by applicable law.
Red Hat, Red Hat Enterprise Linux, the Shadowman logo, JBoss, MetaMatrix, Fedora, the Infinity
Logo, and RHCE are trademarks of Red Hat, Inc., registered in the United States and other countries.
Linux® is the registered trademark of Linus Torvalds in the United States and other countries.
XFS® is a trademark of Silicon Graphics International Corp. or its subsidiaries in the United States
and/or other countries.
MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other
countries.
Configuring and Managing a Red Hat Cluster describes the configuration and management of
Red Hat cluster systems for Red Hat Enterprise Linux 5. It does not include information about Red
Hat Linux Virtual Servers (LVS). Information about installing and configuring LVS is in a separate
document.
Introduction v
1. Document Conventions ................................................................................................... vi
1.1. Typographic Conventions ..................................................................................... vi
1.2. Pull-quote Conventions ........................................................................................ vii
1.3. Notes and Warnings ........................................................................................... viii
2. Feedback ..................................................................................................................... viii
1. Red Hat Cluster Configuration and Management Overview 1
1.1. Configuration Basics ..................................................................................................... 1
1.1.1. Setting Up Hardware ......................................................................................... 1
1.1.2. Installing Red Hat Cluster software ..................................................................... 2
1.1.3. Configuring Red Hat Cluster Software ................................................................ 2
1.2. Conga ......................................................................................................................... 4
1.3. system-config-cluster Cluster Administration GUI ................................................. 6
1.3.1. Cluster Configuration Tool .............................................................................. 6
1.3.2. Cluster Status Tool .......................................................................................... 8
1.4. Command Line Administration Tools .............................................................................. 9
2. Before Configuring a Red Hat Cluster 11
2.1. General Configuration Considerations .......................................................................... 11
2.2. Compatible Hardware ................................................................................................. 12
2.3. Enabling IP Ports ....................................................................................................... 12
2.3.1. Enabling IP Ports on Cluster Nodes .................................................................. 12
2.3.2. Enabling IP Ports on Computers That Run luci ................................................. 13
2.4. Configuring ACPI For Use with Integrated Fence Devices ............................................. 14
2.4.1. Disabling ACPI Soft-Off with chkconfig Management ...................................... 15
2.4.2. Disabling ACPI Soft-Off with the BIOS .............................................................. 15
2.4.3. Disabling ACPI Completely in the grub.conf File ............................................ 17
2.5. Considerations for Configuring HA Services ................................................................. 18
2.6. Configuring max_luns ................................................................................................. 20
2.7. Considerations for Using Quorum Disk ........................................................................ 21
2.8. Red Hat Cluster Suite and SELinux ............................................................................. 22
2.9. Multicast Addresses ................................................................................................... 22
2.10. Enabling Multicast Traffic .......................................................................................... 23
2.11. Considerations for Using Conga ................................................................................ 23
3. Configuring Red Hat Cluster With Conga 25
3.1. Configuration Tasks .................................................................................................... 25
3.2. Starting luci and ricci ................................................................................................ 25
3.3. Creating A Cluster ...................................................................................................... 27
3.4. Global Cluster Properties ............................................................................................ 27
3.5. Configuring Fence Devices ......................................................................................... 30
3.5.1. Creating a Shared Fence Device ...................................................................... 30
3.5.2. Modifying or Deleting a Fence Device ............................................................... 32
3.6. Configuring Cluster Members ...................................................................................... 32
3.6.1. Initially Configuring Members ............................................................................ 33
3.6.2. Adding a Member to a Running Cluster ............................................................ 33
3.6.3. Deleting a Member from a Cluster .................................................................... 34
3.7. Configuring a Failover Domain .................................................................................... 35
3.7.1. Adding a Failover Domain ................................................................................ 36
3.7.2. Modifying a Failover Domain ............................................................................ 37
3.8. Adding Cluster Resources .......................................................................................... 38
3.9. Adding a Cluster Service to the Cluster ....................................................................... 38
3.10. Configuring Cluster Storage ...................................................................................... 40
4. Managing Red Hat Cluster With Conga 43
iii
Cluster Administration
iv
Introduction
This document provides information about installing, configuring and managing Red Hat Cluster
components. Red Hat Cluster components are part of Red Hat Cluster Suite and allow you to connect
a group of computers (called nodes or members) to work together as a cluster. This document
does not include information about installing, configuring, and managing Linux Virtual Server (LVS)
software. Information about that is in a separate document.
The audience of this document should have advanced working knowledge of Red Hat Enterprise Linux
and understand the concepts of clusters, storage, and server computing.
For more information about Red Hat Enterprise Linux 5, refer to the following resources:
• Red Hat Enterprise Linux Installation Guide — Provides information regarding installation of Red
Hat Enterprise Linux 5.
• Red Hat Enterprise Linux Deployment Guide — Provides information regarding the deployment,
configuration and administration of Red Hat Enterprise Linux 5.
For more information about Red Hat Cluster Suite for Red Hat Enterprise Linux 5, refer to the following
resources:
• Red Hat Cluster Suite Overview — Provides a high level overview of the Red Hat Cluster Suite.
• Logical Volume Manager Administration — Provides a description of the Logical Volume Manager
(LVM), including information on running LVM in a clustered environment.
• Global File System: Configuration and Administration — Provides information about installing,
configuring, and maintaining Red Hat GFS (Red Hat Global File System).
• Global File System 2: Configuration and Administration — Provides information about installing,
configuring, and maintaining Red Hat GFS2 (Red Hat Global File System 2).
v
Introduction
• Using Device-Mapper Multipath — Provides information about using the Device-Mapper Multipath
feature of Red Hat Enterprise Linux 5.
• Using GNBD with Global File System — Provides an overview on using Global Network Block
Device (GNBD) with Red Hat GFS.
• Red Hat Cluster Suite Release Notes — Provides information about the current release of Red Hat
Cluster Suite.
Red Hat Cluster Suite documentation and other Red Hat documents are available in HTML,
PDF, and RPM versions on the Red Hat Enterprise Linux Documentation CD and online at http://
www.redhat.com/docs/.
1. Document Conventions
This manual uses several conventions to highlight certain words and phrases and draw attention to
specific pieces of information.
1
In PDF and paper editions, this manual uses typefaces drawn from the Liberation Fonts set. The
Liberation Fonts set is also used in HTML editions if the set is installed on your system. If not,
alternative but equivalent typefaces are displayed. Note: Red Hat Enterprise Linux 5 and later includes
the Liberation Fonts set by default.
Mono-spaced Bold
Used to highlight system input, including shell commands, file names and paths. Also used to highlight
keycaps and key combinations. For example:
The above includes a file name, a shell command and a keycap, all presented in mono-spaced bold
and all distinguishable thanks to context.
Key combinations can be distinguished from keycaps by the hyphen connecting each part of a key
combination. For example:
The first paragraph highlights the particular keycap to press. The second highlights two key
combinations (each a set of three keycaps with each set pressed simultaneously).
1
https://fedorahosted.org/liberation-fonts/
vi
Pull-quote Conventions
If source code is discussed, class names, methods, functions, variable names and returned values
mentioned within a paragraph will be presented as above, in mono-spaced bold. For example:
File-related classes include filesystem for file systems, file for files, and dir for
directories. Each class has its own associated set of permissions.
Proportional Bold
This denotes words or phrases encountered on a system, including application names; dialog box text;
labeled buttons; check-box and radio button labels; menu titles and sub-menu titles. For example:
Choose System → Preferences → Mouse from the main menu bar to launch Mouse
Preferences. In the Buttons tab, click the Left-handed mouse check box and click
Close to switch the primary mouse button from the left to the right (making the mouse
suitable for use in the left hand).
The above text includes application names; system-wide menu names and items; application-specific
menu names; and buttons and text found within a GUI interface, all presented in proportional bold and
all distinguishable by context.
Whether mono-spaced bold or proportional bold, the addition of italics indicates replaceable or
variable text. Italics denotes text you do not input literally or displayed text that changes depending on
circumstance. For example:
To see the version of a currently installed package, use the rpm -q package
command. It will return a result as follows: package-version-release.
Note the words in bold italics above — username, domain.name, file-system, package, version and
release. Each word is a placeholder, either for text you enter when issuing a command or for text
displayed by the system.
Aside from standard usage for presenting the title of a work, italics denotes the first use of a new and
important term. For example:
vii
Introduction
Source-code listings are also set in mono-spaced roman but add syntax highlighting as follows:
package org.jboss.book.jca.ex1;
import javax.naming.InitialContext;
System.out.println("Created Echo");
Note
Notes are tips, shortcuts or alternative approaches to the task at hand. Ignoring a note should
have no negative consequences, but you might miss out on a trick that makes your life easier.
Important
Important boxes detail things that are easily missed: configuration changes that only apply to
the current session, or services that need restarting before an update will apply. Ignoring a box
labeled 'Important' will not cause data loss but may cause irritation and frustration.
Warning
Warnings should not be ignored. Ignoring warnings will most likely cause data loss.
2. Feedback
If you spot a typo, or if you have thought of a way to make this manual better, we would love to
hear from you. Please submit a report in Bugzilla (http://bugzilla.redhat.com/bugzilla/) against the
component
viii Documentation-cluster.
Feedback
Cluster_Administration(EN)-5 (2011-07-21T15:52)
By mentioning this manual's identifier, we know exactly which version of the guide you have.
If you have a suggestion for improving the documentation, try to be as specific as possible. If you have
found an error, please include the section number and some of the surrounding text so we can find it
easily.
ix
x
Chapter 1.
Configuring and managing a Red Hat Cluster consists of the following basic steps:
2. Installing Red Hat Cluster software. Refer to Section 1.1.2, “Installing Red Hat Cluster software”.
3. Configuring Red Hat Cluster Software. Refer to Section 1.1.3, “Configuring Red Hat Cluster
Software”.
• Cluster nodes — Computers that are capable of running Red Hat Enterprise Linux 5 software, with
at least 1GB of RAM. The maximum number of nodes supported in a Red Hat Cluster is 16.
• Ethernet switch or hub for public network — This is required for client access to the cluster.
• Ethernet switch or hub for private network — This is required for communication among the cluster
nodes and other cluster hardware such as network power switches and Fibre Channel switches.
• Fibre Channel switch — A Fibre Channel switch provides access to Fibre Channel storage. Other
options are available for storage according to the type of storage interface; for example, iSCSI or
GNBD. A Fibre Channel switch can be configured to perform fencing.
• Storage — Some type of storage is required for a cluster. The type required depends on the
purpose of the cluster.
1
Chapter 1. Red Hat Cluster Configuration and Management Overview
2
Configuring Red Hat Cluster Software
The following cluster configuration tools are available with Red Hat Cluster:
• Conga — This is a comprehensive user interface for installing, configuring, and managing Red Hat
clusters, computers, and storage attached to clusters and computers.
• system-config-cluster — This is a user interface for configuring and managing a Red Hat
cluster.
• Command line tools — This is a set of command line tools for configuring and managing a Red Hat
cluster.
3
Chapter 1. Red Hat Cluster Configuration and Management Overview
1.2. Conga
Conga is an integrated set of software components that provides centralized configuration and
management of Red Hat clusters and storage. Conga provides the following major features:
• No Need to Re-Authenticate
The primary components in Conga are luci and ricci, which are separately installable. luci is a server
that runs on one computer and communicates with multiple clusters and computers via ricci. ricci is
an agent that runs on each computer (either a cluster member or a standalone computer) managed by
Conga.
luci is accessible through a Web browser and provides three major functions that are accessible
through the following tabs:
• homebase — Provides tools for adding and deleting computers, adding and deleting users, and
configuring user privileges. Only a system administrator is allowed to access this tab.
• cluster — Provides tools for creating and configuring clusters. Each instance of luci lists clusters
that have been set up with that luci. A system administrator can administer all clusters listed on this
tab. Other users can administer only clusters that the user has permission to manage (granted by an
administrator).
• storage — Provides tools for remote administration of storage. With the tools on this tab, you can
manage storage on computers whether they belong to a cluster or not.
You can populate the database of one luci instance from another luciinstance. That capability
provides a means of replicating a luci server instance and provides an efficient upgrade and testing
path. When you install an instance of luci, its database is empty. However, you can import part or all of
a luci database from an existing luci server when deploying a new luci server.
Each luci instance has one user at initial installation — admin. Only the admin user may add systems
to a luci server. Also, the admin user can create additional user accounts and determine which users
are allowed to access clusters and computers registered in the luci database. It is possible to import
users as a batch operation in a new luci server, just as it is possible to import clusters and computers.
The following figures show sample displays of the three major luci tabs: homebase, cluster, and
storage.
4
Conga
For more information about Conga, refer to Chapter 3, Configuring Red Hat Cluster With Conga,
Chapter 4, Managing Red Hat Cluster With Conga, and the online help available with the luci server.
5
Chapter 1. Red Hat Cluster Configuration and Management Overview
Note
6
Cluster Configuration Tool
The Cluster Configuration Tool represents cluster configuration components in the configuration file
(/etc/cluster/cluster.conf) with a hierarchical graphical display in the left panel. A triangle
icon to the left of a component name indicates that the component has one or more subordinate
components assigned to it. Clicking the triangle icon expands and collapses the portion of the tree
below a component. The components displayed in the GUI are summarized as follows:
• Cluster Nodes — Displays cluster nodes. Nodes are represented by name as subordinate
elements under Cluster Nodes. Using configuration buttons at the bottom of the right frame (below
Properties), you can add nodes, delete nodes, edit node properties, and configure fencing methods
for each node.
• Fence Devices — Displays fence devices. Fence devices are represented as subordinate
elements under Fence Devices. Using configuration buttons at the bottom of the right frame (below
Properties), you can add fence devices, delete fence devices, and edit fence-device properties.
Fence devices must be defined before you can configure fencing (with the Manage Fencing For
This Node button) for each node.
7
Chapter 1. Red Hat Cluster Configuration and Management Overview
• Failover Domains — For configuring one or more subsets of cluster nodes used to run a high-
availability service in the event of a node failure. Failover domains are represented as subordinate
elements under Failover Domains. Using configuration buttons at the bottom of the right frame
(below Properties), you can create failover domains (when Failover Domains is selected) or edit
failover domain properties (when a failover domain is selected).
Note
The Cluster Configuration Tool provides the capability to configure private resources, also.
A private resource is a resource that is configured for use with only one service. You can
configure a private resource within a Service component in the GUI.
8
Command Line Administration Tools
The nodes and services displayed in the Cluster Status Tool are determined by the cluster
configuration file (/etc/cluster/cluster.conf). You can use the Cluster Status Tool to enable,
disable, restart, or relocate a high-availability service.
9
Chapter 1. Red Hat Cluster Configuration and Management Overview
10
Chapter 2.
Important
Make sure that your deployment of Red Hat Cluster Suite meets your needs and can be
supported. Consult with an authorized Red Hat representative to verify Cluster Suite and GFS
configuration prior to deployment. In addition, allow time for a configuration burn-in period to test
failure modes.
• Section 2.4, “Configuring ACPI For Use with Integrated Fence Devices”
GFS/GFS2
Although a GFS/GFS2 file system can be implemented in a standalone system or as part of a
cluster configuration, for the RHEL 5.5 release and later, Red Hat does not support the use of
GFS/GFS2 as a single-node file system. Red Hat does support a number of high-performance
single-node file systems that are optimized for single node, and thus have generally lower
overhead than a cluster file system. Red Hat recommends using those file systems in preference
to GFS/GFS2 in cases where only a single node needs to mount the file system. Red Hat will
continue to support single-node GFS/GFS2 file systems for existing customers.
When you configure a GFS/GFS2 file system as a cluster file system, you must ensure that all
nodes in the cluster have access to the shared file system. Asymmetric cluster configurations in
which some nodes have access to the file system and others do not are not supported.This does
not require that all nodes actually mount the GFS/GFS2 file system itself.
11
Chapter 2. Before Configuring a Red Hat Cluster
Alternatively, a low-cost cluster can be set up to provide less availability than a no-single-point-of-
failure cluster. For example, you can set up a cluster with a single-controller RAID array and only a
single Ethernet channel.
Certain low-cost alternatives, such as host RAID controllers, software RAID without cluster
support, and multi-initiator parallel SCSI configurations are not compatible or appropriate for use
as shared cluster storage.
12
Enabling IP Ports on Computers That Run luci
Note
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.
Note
Table 2.1, “Enabled IP Ports on Red Hat Cluster Nodes” shows no IP ports to enable for
rgmanager. For Red Hat Enterprise Linux 5.1 and later, rgmanager does not use TCP or UDP
sockets.
Note
If a cluster node is running luci, port 11111 should already have been enabled.
If your server infrastructure incorporates more than one network and you want to access luci from the
internal network only, you can configure the stunnel component to listen on one IP address only by
editing the LUCI_HTTPS_PORT parameter in the /etc/sysconfig/luci file as follows:
13
Chapter 2. Before Configuring a Red Hat Cluster
LUCI_HTTPS_PORT=10.10.10.10:8084
Note
For the most current information about integrated fence devices supported by Red Hat Cluster
1
Suite, refer to http://www.redhat.com/cluster_suite/hardware/ .
If a cluster node is configured to be fenced by an integrated fence device, disable ACPI Soft-Off for
that node. Disabling ACPI Soft-Off allows an integrated fence device to turn off a node immediately
and completely rather than attempting a clean shutdown (for example, shutdown -h now).
Otherwise, if ACPI Soft-Off is enabled, an integrated fence device can take four or more seconds to
turn off a node (refer to note that follows). In addition, if ACPI Soft-Off is enabled and a node panics
or freezes during shutdown, an integrated fence device may not be able to turn off the node. Under
those circumstances, fencing is delayed or unsuccessful. Consequently, when a node is fenced
with an integrated fence device and ACPI Soft-Off is enabled, a cluster recovers slowly or requires
administrative intervention to recover.
Note
The amount of time required to fence a node depends on the integrated fence device used.
Some integrated fence devices perform the equivalent of pressing and holding the power button;
therefore, the fence device turns off the node in four to five seconds. Other integrated fence
devices perform the equivalent of pressing the power button momentarily, relying on the operating
system to turn off the node; therefore, the fence device turns off the node in a time span much
longer than four to five seconds.
To disable ACPI Soft-Off, use chkconfig management and verify that the node turns off immediately
when fenced. The preferred way to disable ACPI Soft-Off is with chkconfig management: however,
if that method is not satisfactory for your cluster, you can disable ACPI Soft-Off with one of the
following alternate methods:
• Changing the BIOS setting to "instant-off" or an equivalent setting that turns off the node without
delay
Note
Disabling ACPI Soft-Off with the BIOS may not be possible with some computers.
1
http://www.redhat.com/cluster_suite/hardware/
14
Disabling ACPI Soft-Off with chkconfig Management
• Appending acpi=off to the kernel boot command line of the /boot/grub/grub.conf file
Important
This method completely disables ACPI; some computers do not boot correctly if ACPI is
completely disabled. Use this method only if the other methods are not effective for your
cluster.
The following sections provide procedures for the preferred method and alternate methods of disabling
ACPI Soft-Off:
• Section 2.4.1, “Disabling ACPI Soft-Off with chkconfig Management” — Preferred method
• Section 2.4.2, “Disabling ACPI Soft-Off with the BIOS” — First alternate method
• Section 2.4.3, “Disabling ACPI Completely in the grub.conf File” — Second alternate method
Note
Disable ACPI Soft-Off with chkconfig management at each cluster node as follows:
• chkconfig --del acpid — This command removes acpid from chkconfig management.
— OR —
• chkconfig --level 2345 acpid off — This command turns off acpid.
3. When the cluster is configured and running, verify that the node turns off immediately when
fenced.
Note
You can fence the node with the fence_node command or Conga.
Note
Disabling ACPI Soft-Off with the BIOS may not be possible with some computers.
You can disable ACPI Soft-Off by configuring the BIOS of each cluster node as follows:
1. Reboot the node and start the BIOS CMOS Setup Utility program.
3. At the Power menu, set the Soft-Off by PWR-BTTN function (or equivalent) to Instant-Off (or the
equivalent setting that turns off the node via the power button without delay). Example 2.1, “BIOS
CMOS Setup Utility: Soft-Off by PWR-BTTN set to Instant-Off” shows a Power menu with
ACPI Function set to Enabled and Soft-Off by PWR-BTTN set to Instant-Off.
Note
The equivalents to ACPI Function, Soft-Off by PWR-BTTN, and Instant-Off may vary
among computers. However, the objective of this procedure is to configure the BIOS so that
the computer is turned off via the power button without delay.
4. Exit the BIOS CMOS Setup Utility program, saving the BIOS configuration.
5. When the cluster is configured and running, verify that the node turns off immediately when
fenced.
Note
You can fence the node with the fence_node command or Conga.
Example 2.1. BIOS CMOS Setup Utility: Soft-Off by PWR-BTTN set to Instant-Off
+---------------------------------------------|-------------------+
| ACPI Function [Enabled] | Item Help |
| ACPI Suspend Type [S1(POS)] |-------------------|
| x Run VGABIOS if S3 Resume Auto | Menu Level * |
| Suspend Mode [Disabled] | |
| HDD Power Down [Disabled] | |
| Soft-Off by PWR-BTTN [Instant-Off | |
| CPU THRM-Throttling [50.0%] | |
| Wake-Up by PCI card [Enabled] | |
| Power On by Ring [Enabled] | |
| Wake Up On LAN [Enabled] | |
| x USB KB Wake-Up From S3 Disabled | |
| Resume by Alarm [Disabled] | |
| x Date(of Month) Alarm 0 | |
| x Time(hh:mm:ss) Alarm 0 : 0 : | |
| POWER ON Function [BUTTON ONLY | |
| x KB Power ON Password Enter | |
| x Hot Key Power ON Ctrl-F1 | |
| | |
16
Disabling ACPI Completely in the grub.conf File
| | |
+---------------------------------------------|-------------------+
This example shows ACPI Function set to Enabled, and Soft-Off by PWR-BTTN set to Instant-
Off.
Important
This method completely disables ACPI; some computers do not boot correctly if ACPI is
completely disabled. Use this method only if the other methods are not effective for your cluster.
You can disable ACPI completely by editing the grub.conf file of each cluster node as follows:
4. When the cluster is configured and running, verify that the node turns off immediately when
fenced.
Note
You can fence the node with the fence_node command or Conga.
root (hd0,0)
kernel /vmlinuz-2.6.18-36.el5 ro root=/dev/VolGroup00/LogVol00
console=ttyS0,115200n8 acpi=off
initrd /initrd-2.6.18-36.el5.img
In this example, acpi=off has been appended to the kernel boot command line — the line starting
with "kernel /vmlinuz-2.6.18-36.el5".
To create an HA service, you must configure it in the cluster configuration file. An HA service
comprises cluster resources. Cluster resources are building blocks that you create and manage in the
cluster configuration file — for example, an IP address, an application initialization script, or a Red Hat
GFS shared partition.
An HA service can run on only one cluster node at a time to maintain data integrity. You can specify
failover priority in a failover domain. Specifying failover priority consists of assigning a priority level to
each node in a failover domain. The priority level determines the failover order — determining which
node that an HA service should fail over to. If you do not specify failover priority, an HA service can fail
over to any node in its failover domain. Also, you can specify if an HA service is restricted to run only
on nodes of its associated failover domain. (When associated with an unrestricted failover domain, an
HA service can start on any cluster node in the event no member of the failover domain is available.)
Figure 2.1, “Web Server Cluster Service Example” shows an example of an HA service that is a web
server named "content-webserver". It is running in cluster node B and is in a failover domain that
consists of nodes A, B, and D. In addition, the failover domain is configured with a failover priority to
fail over to node D before node A and to restrict failover to nodes only in that failover domain. The HA
service comprises these cluster resources:
• An application resource named "httpd-content" — a web server application init script /etc/
init.d/httpd (specifying httpd).
18
Considerations for Configuring HA Services
Clients access the HA service through the IP address 10.10.10.201, enabling interaction with the web
server application, httpd-content. The httpd-content application uses the gfs-content-webserver file
system. If node B were to fail, the content-webserver HA service would fail over to node D. If node
D were not available or also failed, the service would fail over to node A. Failover would occur with
minimal service interruption to the cluster clients. For example, in an HTTP service, certain state
information may be lost (like session data). The HA service would be accessible from another cluster
node via the same IP address as it was before failover.
Note
For more information about HA services and failover domains, refer to Red Hat Cluster Suite
Overview. For information about configuring failover domains, refer to Section 3.7, “Configuring a
Failover Domain” (using Conga) or Section 5.6, “Configuring a Failover Domain” (using system-
config-cluster).
An HA service is a group of cluster resources configured into a coherent entity that provides
specialized services to clients. An HA service is represented as a resource tree in the cluster
configuration file, /etc/cluster/cluster.conf (in each cluster node). In the cluster configuration
19
Chapter 2. Before Configuring a Red Hat Cluster
file, each resource tree is an XML representation that specifies each resource, its attributes, and its
relationship among other resources in the resource tree (parent, child, and sibling relationships).
Note
At the root of each resource tree is a special type of resource — a service resource. Other types of
resources comprise the rest of a service, determining its characteristics. Configuring an HA service
consists of creating a service resource, creating subordinate cluster resources, and organizing them
into a coherent entity that conforms to hierarchical restrictions of the service.
• Apache
• Application (Script)
• MySQL
• NFS
• Open LDAP
• Oracle
• PostgreSQL 8
• Samba
• SAP
• Tomcat 5
There are two major considerations to take into account when configuring an HA service:
The types of resources and the hierarchy of resources depend on the type of service you are
configuring.
The types of cluster resources are listed in Appendix C, HA Resource Parameters. Information about
parent, child, and sibling relationships among resources is described in Appendix D, HA Resource
Behavior.
In Red Hat Enterprise Linux releases prior to Red Hat Enterprise Linux 5, if RAID storage in a cluster
presents multiple LUNs, it is necessary to enable access to those LUNs by configuring max_luns
20
Considerations for Using Quorum Disk
(or max_scsi_luns for 2.4 kernels) in the /etc/modprobe.conf file of each node. In Red
Hat Enterprise Linux 5, cluster nodes detect multiple LUNs without intervention required; it is not
necessary to configure max_luns to detect multiple LUNs.
Note
Configuring qdiskd is not required unless you have special requirements for node health. An
example of a special requirement is an "all-but-one" configuration. In an all-but-one configuration,
qdiskd is configured to provide enough quorum votes to maintain quorum even though only one
node is working.
Important
Overall, heuristics and other qdiskd parameters for your Red Hat Cluster depend on the site
environment and special requirements needed. To understand the use of heuristics and other
qdiskd parameters, refer to the qdisk(5) man page. If you require assistance understanding and
using qdiskd for your site, contact an authorized Red Hat support representative.
If you need to use qdiskd, you should take into account the following considerations:
Fencing
To ensure reliable fencing when using qdiskd, use power fencing. While other types of fencing
(such as watchdog timers and software-based solutions to reboot a node internally) can be reliable
for clusters not configured with qdiskd, they are not reliable for a cluster configured with qdiskd.
21
Chapter 2. Before Configuring a Red Hat Cluster
Maximum nodes
A cluster configured with qdiskd supports a maximum of 16 nodes. The reason for the limit
is because of scalability; increasing the node count increases the amount of synchronous I/O
contention on the shared quorum disk device.
Note
Using JBOD as a quorum disk is not recommended. A JBOD cannot provide dependable
performance and therefore may not allow a node to write to it quickly enough. If a node is
unable to write to a quorum disk device quickly enough, the node is falsely evicted from a
cluster.
• Red Hat Enterprise Linux 5.5 and later — enforcing or permissive state with the SELinux
policy type set to targeted (or with the state set to disabled).
For more information about SELinux, refer to Deployment Guide for Red Hat Enterprise Linux 5.
Note
Procedures for configuring network switches and associated networking equipment vary
according each product. Refer to the appropriate vendor documentation or other information
about configuring network switches and associated networking equipment to enable multicast
addresses and IGMP.
22
Enabling Multicast Traffic
Note
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.
For openais:
iptables -I INPUT -p udp -m state --state NEW -m multiport --dports 5404,5405 -j ACCEPT
For ricci:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 11111 -j ACCEPT
For modcluster:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 16851 -j ACCEPT
For gnbd:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 14567 -j ACCEPT
For luci:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 8084 -j ACCEPT
For DLM:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 21064 -j ACCEPT
There is no special consideration for rgmanager on Red Hat Enterprise Linux 5; it uses ports
5404/5405.
23
Chapter 2. Before Configuring a Red Hat Cluster
network. If the computer running luci is on another network (for example, a public network rather
than a private network that the cluster is communicating on), contact an authorized Red Hat support
representative to make sure that the appropriate host name is configured for each cluster node.
24
Chapter 3.
1. Configuring and running the Conga configuration user interface — the luci server. Refer to
Section 3.2, “Starting luci and ricci”.
3. Configuring global cluster properties. Refer to Section 3.4, “Global Cluster Properties”.
8. Creating cluster services. Refer to Section 3.9, “Adding a Cluster Service to the Cluster”.
1. At each node to be administered by Conga, install the ricci agent. For example:
25
Chapter 3. Configuring Red Hat Cluster With Conga
3. Select a computer to host luci and install the luci software on that computer. For example:
Note
Typically, a computer in a server cage or a data center hosts luci; however, a cluster
computer can host luci.
4. At the computer running luci, initialize the luci server using the luci_admin init command.
For example:
# luci_admin init
Initializing the Luci server
Please wait...
The admin password has been successfully set.
Generating SSL certificates...
Luci server has been successfully initialized
6. At a Web browser, place the URL of the luci server into the URL address box and click Go (or the
equivalent). The URL syntax for the luci server is https://luci_server_hostname:8084.
26
Creating A Cluster
The first time you access luci, two SSL certificate dialog boxes are displayed. Upon
acknowledging the dialog boxes, your Web browser displays the luci login page.
3. At the Cluster Name text box, enter a cluster name. The cluster name cannot exceed 15
characters. Add the node name and password for each cluster node. Enter the node name for
each node in the Node Hostname column; enter the root password for each node in the in the
Root Password column. Check the Enable Shared Storage Support checkbox if clustered
storage is required.
c. Cluster configuration file to be created and propagated to each node in the cluster.
A progress page shows the progress of those actions for each node in the cluster.
When the process of creating a new cluster is complete, a page is displayed providing a
configuration interface for the newly created cluster.
1. General tab — This tab displays cluster name and provides an interface for configuring the
configuration version and advanced cluster properties. The parameters are summarized as
follows:
• The Cluster Name text box displays the cluster name; it does not accept a cluster name
change. You cannot change the cluster name. The only way to change the name of a Red Hat
cluster is to create a new cluster configuration with the new name.
• The Configuration Version value is set to 1 by default and is automatically incremented each
time you modify your cluster configuration. However, if you need to set it to another value, you
can specify it at the Configuration Version text box.
27
Chapter 3. Configuring Red Hat Cluster With Conga
• You can enter advanced cluster properties by clicking Show advanced cluster properties.
Clicking Show advanced cluster properties reveals a list of advanced properties. You can
click any advanced property for online help about the property.
Enter the values required and click Apply for changes to take effect.
2. Fence tab — This tab provides an interface for configuring these Fence Daemon Properties
parameters: Post-Fail Delay and Post-Join Delay. The parameters are summarized as follows:
• The Post-Fail Delay parameter is the number of seconds the fence daemon (fenced) waits
before fencing a node (a member of the fence domain) after the node has failed. The Post-Fail
Delay default value is 0. Its value may be varied to suit cluster and network performance.
• The Post-Join Delay parameter is the number of seconds the fence daemon (fenced) waits
before fencing a node after the node joins the fence domain. The Post-Join Delay default
value is 3. A typical setting for Post-Join Delay is between 20 and 30 seconds, but can vary
according to cluster and network performance.
Enter values required and Click Apply for changes to take effect.
Note
For more information about Post-Join Delay and Post-Fail Delay, refer to the fenced(8) man
page.
3. Multicast tab — This tab provides an interface for configuring these Multicast Configuration
parameters: Let cluster choose the multicast address and Specify the multicast address
manually. The default setting is Let cluster choose the multicast address. If you need to use
a specific multicast address, click Specify the multicast address manually, enter a multicast
address into the text box, and click Apply for changes to take effect.
Note
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.
If you do not specify a multicast address, the Red Hat Cluster software (specifically, cman, the
Cluster Manager) creates one. It forms the upper 16 bits of the multicast address with 239.192 and
forms the lower 16 bits based on the cluster ID.
Note
The cluster ID is a unique identifier that cman generates for each cluster. To view the cluster
ID, run the cman_tool status command on a cluster node.
If you do specify a multicast address, you should use the 239.192.x.x series that cman uses.
Otherwise, using a multicast address outside that range may cause unpredictable results. For
example, using 224.0.0.x (which is "All hosts on the network") may not be routed correctly, or even
routed at all by some hardware.
28
Global Cluster Properties
Note
If you specify a multicast address, make sure that you check the configuration of routers
that cluster packets pass through. Some routers may take a long time to learn addresses,
seriously impacting cluster performance.
4. Quorum Partition tab — This tab provides an interface for configuring these Quorum Partition
Configuration parameters: Do not use a Quorum Partition, Use a Quorum Partition, Interval,
Votes, TKO, Minimum Score, Device, Label, and Heuristics. The Do not use a Quorum
Partition parameter is enabled by default. Table 3.1, “Quorum-Disk Parameters” describes the
parameters. If you need to use a quorum disk, click Use a Quorum Partition, enter quorum disk
parameters, click Apply, and restart the cluster for the changes to take effect.
Important
Quorum-disk parameters and heuristics depend on the site environment and the special
requirements needed. To understand the use of quorum-disk parameters and heuristics, refer
to the qdisk(5) man page. If you require assistance understanding and using quorum disk,
contact an authorized Red Hat support representative.
Note
Clicking Apply on the Quorum Partition tab propagates changes to the cluster configuration
file (/etc/cluster/cluster.conf) in each cluster node. However, for the quorum disk to
operate, you must restart the cluster (refer to Section 4.1, “Starting, Stopping, and Deleting
Clusters”).
29
Chapter 3. Configuring Red Hat Cluster With Conga
Parameter Description
Device The storage device the quorum daemon uses. The device must be the
same on all nodes.
Label Specifies the quorum disk label created by the mkqdisk utility. If this field
contains an entry, the label overrides the Device field. If this field is used,
the quorum daemon reads /proc/partitions and checks for qdisk
signatures on every block device found, comparing the label against the
specified label. This is useful in configurations where the quorum device
name differs among nodes.
Heuristics Path to Program — The program used to determine if this heuristic is
alive. This can be anything that can be executed by /bin/sh -c. A return
value of 0 indicates success; anything else indicates failure. This field is
required.
Interval — The frequency (in seconds) at which the heuristic is polled. The
default interval for every heuristic is 2 seconds.
Score — The weight of this heuristic. Be careful when determining scores
for heuristics. The default score for each heuristic is 1.
Apply Propagates the changes to the cluster configuration file (/etc/cluster/
cluster.conf) in each cluster node.
Note
If you are creating a new cluster, you can create fence devices when you configure cluster nodes.
Refer to Section 3.6, “Configuring Cluster Members”.
With Conga you can create shared and non-shared fence devices. For information on supported fence
devices and there parameters, refer to Appendix B, Fence Device Parameters.
• Creating shared fence devices — Refer to Section 3.5.1, “Creating a Shared Fence Device”. The
procedures apply only to creating shared fence devices. You can create non-shared (and shared)
fence devices while configuring nodes (refer to Section 3.6, “Configuring Cluster Members”).
• Modifying or deleting fence devices — Refer to Section 3.5.2, “Modifying or Deleting a Fence
Device”. The procedures apply to both shared and non-shared fence devices.
The starting point of each procedure is at the cluster-specific page that you navigate to from Choose a
cluster to administer displayed on the cluster tab.
30
Creating a Shared Fence Device
1. At the detailed menu for the cluster (below the clusters menu), click Shared Fence Devices.
Clicking Shared Fence Devices causes the display of the fence devices for a cluster and causes
the display of menu items for fence device configuration: Add a Fence Device and Configure a
Fence Device.
Note
If this is an initial cluster configuration, no fence devices have been created, and therefore
none are displayed.
2. Click Add a Fence Device. Clicking Add a Fence Device causes the Add a Sharable Fence
Device page to be displayed (refer to Figure 3.1, “Fence Device Configuration”).
3. At the Add a Sharable Fence Device page, click the drop-down box under Fencing Type and
select the type of fence device to configure.
31
Chapter 3. Configuring Red Hat Cluster With Conga
4. Specify the information in the Fencing Type dialog box according to the type of fence device.
Refer to Appendix B, Fence Device Parameters for more information about fence device
parameters.
6. Clicking Add this shared fence device causes a progress page to be displayed temporarily. After
the fence device has been added, the detailed cluster properties menu is updated with the fence
device under Configure a Fence Device.
1. At the detailed menu for the cluster (below the clusters menu), click Shared Fence Devices.
Clicking Shared Fence Devices causes the display of the fence devices for a cluster and causes
the display of menu items for fence device configuration: Add a Fence Device and Configure a
Fence Device.
2. Click Configure a Fence Device. Clicking Configure a Fence Device causes the display of a list
of fence devices under Configure a Fence Device.
3. Click a fence device in the list. Clicking a fence device in the list causes the display of a Fence
Device Form page for the fence device selected from the list.
• To modify the fence device, enter changes to the parameters displayed. Refer to Appendix B,
Fence Device Parameters for more information about fence device parameters. Click Update
this fence device and wait for the configuration to be updated.
• To delete the fence device, click Delete this fence device and wait for the configuration to be
updated.
Note
You can create shared fence devices on the node configuration page, also. However, you
can only modify or delete a shared fence device via Shared Fence Devices at the detailed
menu for the cluster (below the clusters menu).
32
Initially Configuring Members
1. At the detailed menu for the cluster (below the clusters menu), click Nodes. Clicking Nodes
causes the display of an Add a Node element and a Configure element with a list of the nodes
already configured in the cluster.
2. Click a link for a node at either the list in the center of the page or in the list in the detailed menu
under the clusters menu. Clicking a link for a node causes a page to be displayed for that link
showing how that node is configured.
3. At the bottom of the page, under Main Fencing Method, click Add a fence device to this level.
4. Select a fence device and provide parameters for the fence device (for example port number).
Note
You can choose from an existing fence device or create a new fence device.
5. Click Update main fence properties and wait for the change to take effect.
1. At the detailed menu for the cluster (below the clusters menu), click Nodes. Clicking Nodes
causes the display of an Add a Node element and a Configure element with a list of the nodes
already configured in the cluster. (In addition, a list of the cluster nodes is displayed in the center
of the page.)
2. Click Add a Node. Clicking Add a Node causes the display of the Add a node to cluster
name page.
3. At that page, enter the node name in the Node Hostname text box; enter the root password in
the Root Password text box. Check the Enable Shared Storage Support checkbox if clustered
storage is required. If you want to add more nodes, click Add another entry and enter node name
and password for the each additional node.
b. Cluster software to be installed (or verification that the appropriate software packages are
installed) onto the added node.
c. Cluster configuration file to be updated and propagated to each node in the cluster —
including the added node.
33
Chapter 3. Configuring Red Hat Cluster With Conga
A progress page shows the progress of those actions for each added node.
5. When the process of adding a node is complete, a page is displayed providing a configuration
interface for the cluster.
6. At the detailed menu for the cluster (below the clusters menu), click Nodes. Clicking Nodes
causes the following displays:
• The Add a Node element and the Configure element with a list of the nodes configured in the
cluster at the detailed menu for the cluster (below the clusters menu)
7. Click the link for an added node at either the list in the center of the page or in the list in the
detailed menu under the clusters menu. Clicking the link for the added node causes a page to be
displayed for that link showing how that node is configured.
8. At the bottom of the page, under Main Fencing Method, click Add a fence device to this level.
9. Select a fence device and provide parameters for the fence device (for example port number).
Note
You can choose from an existing fence device or create a new fence device.
10. Click Update main fence properties and wait for the change to take effect.
1. Click the link of the node to be deleted. Clicking the link of the node to be deleted causes a page
to be displayed for that link showing how that node is configured.
Note
To allow services running on a node to fail over when the node is deleted, skip the next step.
Note
Repeat this step for each service that needs to be disabled or started on another node.
34
Configuring a Failover Domain
a. Under Services on this Node, click the link for a service. Clicking that link cause a
configuration page for that service to be displayed.
b. On that page, at the Choose a taskdrop-down box, choose to either disable the service are
start it on another node and click Go.
c. Upon confirmation that the service has been disabled or started on another node, click the
cluster tab. Clicking the cluster tab causes the Choose a cluster to administer page to be
displayed.
d. At the Choose a cluster to administer page, click the link of the node to be deleted. Clicking
the link of the node to be deleted causes a page to be displayed for that link showing how that
node is configured.
3. On that page, at the Choose a taskdrop-down box, choose Delete this node and click Go. When
the node is deleted, a page is displayed that lists the nodes in the cluster. Check the list to make
sure that the node has been deleted.
• Unrestricted — Allows you to specify that a subset of members are preferred, but that a cluster
service assigned to this domain can run on any available member.
• Restricted — Allows you to restrict the members that can run a particular cluster service. If none
of the members in a restricted failover domain are available, the cluster service cannot be started
(either manually or by the cluster software).
• Unordered — When a cluster service is assigned to an unordered failover domain, the member on
which the cluster service runs is chosen from the available failover domain members with no priority
ordering.
• Ordered — Allows you to specify a preference order among the members of a failover domain. The
member at the top of the list is the most preferred, followed by the second member in the list, and so
on.
• Failback — Allows you to specify whether a service in the failover domain should fail back to the
node that it was originally running on before that node failed. Configuring this characteristic is useful
in circumstances where a node repeatedly fails and is part of an ordered failover domain. In that
circumstance, if a node is the preferred node in a failover domain, it is possible for a service to fail
over and fail back repeatedly between the preferred node and another node, causing severe impact
on performance.
Note
35
Chapter 3. Configuring Red Hat Cluster With Conga
Note
Note
In a cluster with several members, using a restricted failover domain can minimize the work to set up
the cluster to run a cluster service (such as httpd), which requires you to set up the configuration
identically on all members that run the cluster service). Instead of setting up the entire cluster to
run the cluster service, you must set up only the members in the restricted failover domain that you
associate with the cluster service.
Note
To configure a preferred member, you can create an unrestricted failover domain comprising only
one cluster member. Doing that causes a cluster service to run on that cluster member primarily
(the preferred member), but allows the cluster service to fail over to any of the other members.
The following sections describe adding a failover domain and modifying a failover domain:
1. At the detailed menu for the cluster (below the clusters menu), click Failover Domains. Clicking
Failover Domains causes the display of failover domains with related services and the display of
menu items for failover domains: Add a Failover Domain and Configure a Failover Domain .
2. Click Add a Failover Domain. Clicking Add a Failover Domain causes the display of the Add a
Failover Domain page.
3. At the Add a Failover Domain page, specify a failover domain name at the Failover Domain
Name text box.
Note
The name should be descriptive enough to distinguish its purpose relative to other names
used in your cluster.
36
Modifying a Failover Domain
4. To enable setting failover priority of the members in the failover domain, click the Prioritized
checkbox. With Prioritized checked, you can set the priority value, Priority, for each node
selected as members of the failover domain.
5. To restrict failover to members in this failover domain, click the checkbox next to Restrict failover
to this domain's members. With Restrict failover to this domain's members checked,
services assigned to this failover domain fail over only to nodes in this failover domain.
6. To specify that a node does not fail back in this failover domain, click the checkbox next to Do not
fail back services in this domain. With Do not fail back services in this domain checked, if a
service fails over from a preferred node, the service does not fail back to the original node once it
has recovered.
7. Configure members for this failover domain. Under Failover domain membership, click the
Member checkbox for each node that is to be a member of the failover domain. If Prioritized is
checked, set the priority in the Priority text box for each member of the failover domain.
8. Click Submit. Clicking Submit causes a progress page to be displayed followed by the display
of the Failover Domain Form page. That page displays the added resource and includes the
failover domain in the cluster menu to the left under Domain.
9. To make additional changes to the failover domain, continue modifications at the Failover Domain
Form page and click Submit when you are done.
1. At the detailed menu for the cluster (below the clusters menu), click Failover Domains. Clicking
Failover Domains causes the display of failover domains with related services and the display of
menu items for failover domains: Add a Failover Domain and Configure a Failover Domain .
2. Click Configure a Failover Domain. Clicking Configure a Failover Domain causes the display
of failover domains under Configure a Failover Domain at the detailed menu for the cluster
(below the clusters menu).
3. At the detailed menu for the cluster (below the clusters menu), click the failover domain to modify.
Clicking the failover domain causes the display of the Failover Domain Form page. At the
Failover Domain Form page, you can modify the failover domain name, prioritize failover, restrict
failover to this domain, and modify failover domain membership.
4. Modifying failover name — To change the failover domain name, modify the text at the Failover
Domain Name text box.
Note
The name should be descriptive enough to distinguish its purpose relative to other names
used in your cluster.
5. Failover priority — To enable or disable prioritized failover in this failover domain, click the
Prioritized checkbox. With Prioritized checked, you can set the priority value, Priority, for each
37
Chapter 3. Configuring Red Hat Cluster With Conga
node selected as members of the failover domain. With Prioritized not checked, setting priority
levels is disabled for this failover domain.
6. Restricted failover — To enable or disable restricted failover for members in this failover domain,
click the checkbox next to Restrict failover to this domain's members. With Restrict failover
to this domain's members checked, services assigned to this failover domain fail over only to
nodes in this failover domain. With Restrict failover to this domain's members not checked,
services assigned to this failover domain can fail over to nodes outside this failover domain.
7. Failback — To enable or disable failback in a failover domain, click the checkbox next to Do not
fail back services in this domain. With Do not fail back services in this domain checked, if a
service fails over from a preferred node, the service does not fail back to the original node once it
has recovered.
8. Modifying failover domain membership — Under Failover domain membership, click the
Membercheckbox for each node that is to be a member of the failover domain. A checked box for
a node means that the node is a member of the failover domain. If Prioritized is checked, you can
adjust the priority in the Priority text box for each member of the failover domain.
9. Click Submit. Clicking Submit causes a progress page to be displayed followed by the display
of the Failover Domain Form page. That page displays the added resource and includes the
failover domain in the cluster menu to the left under Domain.
10. To make additional changes to the failover domain, continue modifications at the Failover Domain
Form page and click Submit when you are done.
1. At the detailed menu for the cluster (below the clusters menu), click Resources. Clicking
Resources causes the display of resources in the center of the page and causes the display of
menu items for resource configuration: Add a Resource and Configure a Resource.
2. Click Add a Resource. Clicking Add a Resource causes the Add a Resource page to be
displayed.
3. At the Add a Resource page, click the drop-down box under Select a Resource Type and select
the type of resource to configure. Appendix C, HA Resource Parameters describes resource
parameters.
4. Click Submit. Clicking Submit causes a progress page to be displayed followed by the display
of Resources forcluster name page. That page displays the added resource (and other
resources).
1. At the detailed menu for the cluster (below the clusters menu), click Services. Clicking Services
causes the display of services in the center of the page and causes the display of menu items for
services configuration: Add a Service and Configure a Service.
38
Adding a Cluster Service to the Cluster
2. Click Add a Service. Clicking Add a Service causes the Add a Service page to be displayed.
3. On the Add a Service page, at the Service name text box, type the name of the service.
Below the Service name text box is an checkbox labeled Automatically start this service.
The checkbox is checked by default. When the checkbox is checked, the service is started
automatically when a cluster is started and running. If the checkbox is not checked, the service
must be started manually any time the cluster comes up from the stopped state.
Note
Use a descriptive name that clearly distinguishes the service from other services in the
cluster.
4. Add a resource to the service; click Add a resource to this service. Clicking Add a resource
to this service causes the display of two drop-down boxes: Add a new local resource and Use
an existing global resource. Adding a new local resource adds a resource that is available only
to this service. The process of adding a local resource is the same as adding a global resource
described in Section 3.8, “Adding Cluster Resources”. Adding a global resource adds a resource
that has been previously added as a global resource (refer to Section 3.8, “Adding Cluster
Resources”).
5. At the drop-down box of either Add a new local resource or Use an existing global resource,
select the resource to add and configure it according to the options presented. (The options are
the same as described in Section 3.8, “Adding Cluster Resources”.)
Note
6. If you want to add resources to that resource, click Add a child. Clicking Add a child causes
the display of additional options to local and global resources. You can continue adding children
resources to the resource to suit your requirements. To view children resources, click the triangle
icon to the left of Show Children.
7. When you have completed adding resources to the service, and have completed adding children
resources to resources, click Submit. Clicking Submit causes a progress page to be displayed
followed by a page displaying the added service (and other services).
39
Chapter 3. Configuring Red Hat Cluster With Conga
Note
To verify the existence of the IP service resource used in a cluster service, you must use the /
sbin/ip addr list command on a cluster node. The following output shows the /sbin/ip
addr list command executed on a node running a cluster service:
The storage tab allows you to monitor and configure storage on remote systems. It provides a
means for configuring disk partitions, logical volumes (clustered and single system use), file system
parameters, and mount points. The storage tab provides an interface for setting up shared storage for
clusters and offers GFS and other file systems as file system options. When a you select the storage
tab, the Welcome to Storage Configuration Interface page shows a list of systems available to
the you in a navigation table to the left. A small form allows you to choose a storage unit size to suit
your preference. That choice is persisted and can be changed at any time by returning to this page.
In addition, you can change the unit type on specific configuration forms throughout the storage user
interface. This general choice allows you to avoid difficult decimal representations of storage size (for
example, if you know that most of your storage is measured in gigabytes, terabytes, or other more
familiar representations).
Additionally, the Welcome to Storage Configuration Interface page lists systems that you are
authorized to access, but currently are unable to administer because of a problem. Examples of
problems:
• A computer has been re-imaged and the luci server admin must re-authenticate with the ricci agent
on the computer.
A reason for the trouble is displayed if the storage user interface can determine it.
Only those computers that the user is privileged to administer is shown in the main navigation table. If
you have no permissions on any computers, a message is displayed.
After you select a computer to administer, a general properties page is displayed for the computer.
This page is divided into three sections:
• Hard Drives
• Partitions
40
Configuring Cluster Storage
• Volume Groups
Each section is set up as an expandable tree, with links to property sheets for specific devices,
partitions, and storage entities.
Configure the storage for your cluster to suit your cluster requirements. If you are configuring Red Hat
GFS, configure clustered logical volumes first, using CLVM. For more information about CLVM and
GFS refer to Red Hat documentation for those products.
Note
Shared storage for use in Red Hat Cluster Suite requires that you be running the cluster logical
volume manager daemon (clvmd) or the High Availability Logical Volume Management agents
(HA-LVM). If you are not able to use either the clvmd daemon or HA-LVM for operational
reasons or because you do not have the correct entitlements, you must not use single-instance
LVM on the shared disk as this may result in data corruption. If you have any concerns please
contact your Red Hat service representative.
41
42
Chapter 4.
• Restart a cluster.
• Start a cluster.
• Stop a cluster.
• Delete a cluster.
To perform one of the functions in the preceding list, follow the steps in this section. The starting point
of the procedure is at the cluster tab (at the Choose a cluster to administer page).
1. At the right of the Cluster Name for each cluster listed on the Choose a cluster to administer
page is a drop-down box. By default, the drop-down box is set to Restart this cluster. Clicking
the drop-down box box reveals all the selections available: Restart this cluster, Stop this
cluster/Start this cluster, and Delete this cluster. The actions of each function are summarized
as follows:
• Restart this cluster — Selecting this action causes the cluster to be restarted. You can select
this action for any state the cluster is in.
• Stop this cluster/Start this cluster — Stop this cluster is available when a cluster is running.
Start this cluster is available when a cluster is stopped.
Selecting Stop this cluster shuts down cluster software in all cluster nodes.
• Delete this cluster — Selecting this action halts a running cluster, disables cluster software
from starting automatically, and removes the cluster configuration file from each node. You can
select this action for any state the cluster is in. Deleting a cluster frees each node in the cluster
for use in another cluster.
3. Clicking Go causes a progress page to be displayed. When the action is complete, a page is
displayed showing either of the following pages according to the action selected:
• For Restart this cluster and Stop this cluster/Start this cluster — Displays a page with the
list of nodes for the cluster.
43
Chapter 4. Managing Red Hat Cluster With Conga
• For Delete this cluster — Displays the Choose a cluster to administer page in the cluster
tab, showing a list of clusters.
• Fence a node.
• Reboot a node.
• Delete a node.
To perform one the functions in the preceding list, follow the steps in this section. The starting point
of the procedure is at the cluster-specific page that you navigate to from Choose a cluster to
administer displayed on the cluster tab.
1. At the detailed menu for the cluster (below the clusters menu), click Nodes. Clicking Nodes
causes the display of nodes in the center of the page and causes the display of an Add a Node
element and a Configure element with a list of the nodes already configured in the cluster.
2. At the right of each node listed on the page displayed from the preceding step, click the Choose
a task drop-down box. Clicking Choose a task drop-down box reveals the following selections:
Have node leave cluster/Have node join cluster, Fence this node, Reboot this node, and
Delete. The actions of each function are summarized as follows:
• Have node leave cluster/Have node join cluster — Have node leave cluster is available
when a node has joined of a cluster. Have node join cluster is available when a node has left a
cluster.
Selecting Have node leave cluster shuts down cluster software and makes the node leave the
cluster. Making a node leave a cluster prevents the node from automatically joining the cluster
when it is rebooted.
Selecting Have node join cluster starts cluster software and makes the node join the cluster.
Making a node join a cluster allows the node to automatically join the cluster when it is rebooted.
• Fence this node — Selecting this action causes the node to be fenced according to how the
node is configured to be fenced.
• Reboot this node — Selecting this action causes the node to be rebooted.
• Delete — Selecting this action causes the node to be deleted from the cluster configuration.
It also stops all cluster services on the node, and deletes the cluster.conf file from /etc/
cluster/.
4. Clicking Go causes a progress page to be displayed. When the action is complete, a page is
displayed showing the list of nodes for the cluster.
44
Managing High-Availability Services
• Configure a service.
• Restart a service.
• Delete a service
To perform one the functions in the preceding list, follow the steps in this section. The starting point
of the procedure is at the cluster-specific page that you navigate to from Choose a cluster to
administer displayed on the cluster tab.
1. At the detailed menu for the cluster (below the clusters menu), click Services. Clicking Services
causes the display of services for the cluster in the center of the page.
2. At the right of each service listed on the page, click the Choose a task drop-down box. Clicking
Choose a task drop-down box reveals the following selections depending on if the service is
running:
• If service is running — Configure this service, Restart this service, and Stop this service.
• If service is not running — Configure this service, Start this service, and Delete this service.
• Configure this service — Configure this service is available when the service is running
or not running. Selecting Configure this service causes the services configuration page for
the service to be displayed. On that page, you can change the configuration of the service. For
example, you can add a resource to the service. (For more information about adding resources
and services, refer to Section 3.8, “Adding Cluster Resources” and Section 3.9, “Adding a
Cluster Service to the Cluster”.) In addition, a drop-down box on the page provides other
functions depending on if the service is running.
When a service is running, the drop-down box provides the following functions: restarting,
disabling, and relocating the service.
When a service is not running, the drop-down box on the configuration page provides the
following functions: enabling and deleting the service.
If you are making configuration changes, save the changes by clicking Save. Clicking Save
causes a progress page to be displayed. When the change is complete, another page is
displayed showing a list of services for the cluster.
If you have selected one of the functions in the drop-down box on the configuration page,
click Go. Clicking Go causes a progress page to be displayed. When the change is complete,
another page is displayed showing a list of services for the cluster.
• Restart this service and Stop this service — These selections are available when the
service is running. Select either function and click Go to make the change take effect. Clicking
Go causes a progress page to be displayed. When the change is complete, another page is
displayed showing a list of services for the cluster.
45
Chapter 4. Managing Red Hat Cluster With Conga
• Start this service and Delete this service — These selections are available when the service
is not running. Select either function and click Go to make the change take effect. Clicking
Go causes a progress page to be displayed. When the change is complete, another page is
displayed showing a list of services for the cluster.
46
Chapter 5.
Note
4. Creating cluster members. Refer to Section 5.5, “Adding and Deleting Members”.
47
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
9. Starting the cluster software. Refer to Section 5.10, “Starting the Cluster Software”.
$ ssh -Y root@nano-01
.
.
.
# system-config-cluster
2. If this is the first time you have started the Cluster Configuration Tool, the program prompts you
to either open an existing configuration or create a new one. Click Create New Configuration to
start a new configuration file (refer to Figure 5.1, “Starting a New Configuration File”).
Note
The Cluster Management tab for the Red Hat Cluster Suite management GUI is available
after you save the configuration file with the Cluster Configuration Tool, exit, and restart
the Red Hat Cluster Suite management GUI (system-config-cluster). (The Cluster
Management tab displays the status of the cluster service manager, cluster nodes, and
resources, and shows statistics concerning cluster service operation. To manage the cluster
system further, choose the Cluster Configuration tab.)
3. Clicking Create New Configuration causes the New Configuration dialog box to be displayed
(refer to Figure 5.2, “Creating A New Configuration”). The New Configuration dialog box provides
48
Custom Configure Multicast
a text box for cluster name and the following checkboxes: Custom Configure Multicast and Use
a Quorum Disk. In most circumstances you only need to configure the cluster name.
Note
Choose the cluster name carefully. The only way to change the name of a Red Hat cluster is
to create a new cluster configuration with the new name.
Note
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.
If you do not specify a multicast address, the Red Hat Cluster software (specifically, cman, the
Cluster Manager) creates one. It forms the upper 16 bits of the multicast address with 239.192 and
forms the lower 16 bits based on the cluster ID.
Note
The cluster ID is a unique identifier that cman generates for each cluster. To view the cluster
ID, run the cman_tool status command on a cluster node.
If you do specify a multicast address, you should use the 239.192.x.x series that cman uses.
Otherwise, using a multicast address outside that range may cause unpredictable results. For
example, using 224.0.0.x (which is "All hosts on the network") may not be routed correctly, or even
routed at all by some hardware.
Note
If you specify a multicast address, make sure that you check the configuration of routers
that cluster packets pass through. Some routers may take a long time to learn addresses,
seriously impacting cluster performance.
49
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
Important
Quorum-disk parameters and heuristics depend on the site environment and special
requirements needed. To understand the use of quorum-disk parameters and heuristics, refer
to the qdisk(5) man page. If you require assistance understanding and using quorum disk,
contact an authorized Red Hat support representative.
Note
Overall:
50
Use a Quorum Disk
4. When you have completed entering the cluster name and other parameters in the New
Configuration dialog box, click OK. Clicking OK starts the Cluster Configuration Tool,
displaying a graphical representation of the configuration (Figure 5.3, “The Cluster Configuration
Tool”).
51
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
52
Configuring Cluster Properties
Parameter Description
the quorum daemon reads /proc/partitions and checks for qdisk
signatures on every block device found, comparing the label against the
specified label. This is useful in configurations where the quorum device
name differs among nodes.
Quorum Disk Program — The program used to determine if this heuristic is alive. This
Heuristics can be anything that can be executed by /bin/sh -c. A return value of 0
indicates success; anything else indicates failure. This field is required.
Score — The weight of this heuristic. Be careful when determining scores
for heuristics. The default score for each heuristic is 1.
Interval — The frequency (in seconds) at which the heuristic is polled. The
default interval for every heuristic is 2 seconds.
2. At the bottom of the right frame (labeled Properties), click the Edit Cluster Properties button.
Clicking that button causes a Cluster Properties dialog box to be displayed. The Cluster
Properties dialog box presents text boxes for Cluster Alias, Config Version, and two Fence
Daemon Properties parameters: Post-Join Delay and Post-Fail Delay.
3. (Optional) At the Cluster Alias text box, specify a cluster alias for the cluster. The default cluster
alias is set to the true cluster name provided when the cluster is set up (refer to Section 5.2,
“Starting the Cluster Configuration Tool”). The cluster alias should be descriptive enough to
distinguish it from other clusters and systems on your network (for example, nfs_cluster or
httpd_cluster). The cluster alias cannot exceed 15 characters.
4. (Optional) The Config Version value is set to 1 by default and is automatically incremented each
time you save your cluster configuration. However, if you need to set it to another value, you can
specify it at the Config Version text box.
5. Specify the Fence Daemon Properties parameters: Post-Join Delay and Post-Fail Delay.
a. The Post-Join Delay parameter is the number of seconds the fence daemon (fenced) waits
before fencing a node after the node joins the fence domain. The Post-Join Delay default
value is 3. A typical setting for Post-Join Delay is between 20 and 30 seconds, but can vary
according to cluster and network performance.
b. The Post-Fail Delay parameter is the number of seconds the fence daemon (fenced) waits
before fencing a node (a member of the fence domain) after the node has failed.The Post-Fail
Delay default value is 0. Its value may be varied to suit cluster and network performance.
Note
For more information about Post-Join Delay and Post-Fail Delay, refer to the fenced(8) man
page.
53
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
1. Click Fence Devices. At the bottom of the right frame (labeled Properties), click the Add a Fence
Device button. Clicking Add a Fence Device causes the Fence Device Configuration dialog
box to be displayed (refer to Figure 5.4, “Fence Device Configuration”).
2. At the Fence Device Configuration dialog box, click the drop-down box under Add a New Fence
Device and select the type of fence device to configure.
3. Specify the information in the Fence Device Configuration dialog box according to the type of
fence device. Refer to Appendix B, Fence Device Parameters for more information about fence
device parameters.
4. Click OK.
5. Choose File => Save to save the changes to the cluster configuration.
54
Adding a Member to a Cluster
2. At the bottom of the right frame (labeled Properties), click the Add a Cluster Node button.
Clicking that button causes a Node Properties dialog box to be displayed. The Node Properties
dialog box presents text boxes for Cluster Node Name and Quorum Votes (refer to Figure 5.5,
“Adding a Member to a New Cluster”).
3. At the Cluster Node Name text box, specify a node name. The entry can be a name or an IP
address of the node on the cluster subnet.
Note
Each node must be on the same subnet as the node from which you are running the Cluster
Configuration Tool and must be defined either in DNS or in the /etc/hosts file of each
cluster node.
Note
The node on which you are running the Cluster Configuration Tool must be explicitly added
as a cluster member; the node is not automatically added to the cluster configuration as a
result of running the Cluster Configuration Tool.
4. Optionally, at the Quorum Votes text box, you can specify a value; however in most
configurations you can leave it blank. Leaving the Quorum Votes text box blank causes the
quorum votes value for that node to be set to the default value of 1.
5. Click OK.
b. At the bottom of the right frame (below Properties), click Manage Fencing For This Node.
Clicking Manage Fencing For This Node causes the Fence Configuration dialog box to be
displayed.
c. At the Fence Configuration dialog box, bottom of the right frame (below Properties), click
Add a New Fence Level. Clicking Add a New Fence Level causes a fence-level element (for
55
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
example, Fence-Level-1, Fence-Level-2, and so on) to be displayed below the node in the
left frame of the Fence Configuration dialog box.
e. At the bottom of the right frame (below Properties), click Add a New Fence to this Level.
Clicking Add a New Fence to this Level causes the Fence Properties dialog box to be
displayed.
f. At the Fence Properties dialog box, click the Fence Device Type drop-down box and select
the fence device for this node. Also, provide additional information required (for example, Port
and Switch for an APC Power Device).
g. At the Fence Properties dialog box, click OK. Clicking OK causes a fence device element to
be displayed below the fence-level element.
h. To create additional fence devices at this fence level, return to step 6d. Otherwise, proceed to
the next step.
i. To create additional fence levels, return to step 6c. Otherwise, proceed to the next step.
j. If you have configured all the fence levels and fence devices for this node, click Close.
7. Choose File => Save to save the changes to the cluster configuration.
Section 5.5.2.1, “Adding a Member to a Running Cluster That Contains Only Two Nodes”
Section 5.5.2.2, “Adding a Member to a Running Cluster That Contains More Than Two Nodes”
2. Click Send to Cluster to propagate the updated configuration to other running nodes in the
cluster.
3. Use the scp command to send the updated /etc/cluster/cluster.conf file from one of the
existing cluster nodes to the new node.
4. At the Red Hat Cluster Suite management GUI Cluster Status Tool tab, disable each service
listed under Services.
56
Adding a Member to a Running Cluster
5. Stop the cluster software on the two running nodes by running the following commands at each
node in this order:
c. service clvmd stop, if CLVM has been used to create clustered volumes
6. Start cluster software on all cluster nodes (including the added one) by running the following
commands in this order:
b. service clvmd start, if CLVM has been used to create clustered volumes
7. Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify
that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services
are running as expected.
2. Click Send to Cluster to propagate the updated configuration to other running nodes in the
cluster.
3. Use the scp command to send the updated /etc/cluster/cluster.conf file from one of the
existing cluster nodes to the new node.
4. Start cluster services on the new node by running the following commands in this order:
b. service clvmd start, if CLVM has been used to create clustered volumes
5. Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify
that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services
are running as expected.
57
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
1. At one of the running nodes (not to be removed), run the Red Hat Cluster Suite management GUI.
At the Cluster Status Tool tab, under Services, disable or relocate each service that is running
on the node to be deleted.
2. Stop the cluster software on the node to be deleted by running the following commands at that
node in this order:
c. service clvmd stop, if CLVM has been used to create clustered volumes
3. At the Cluster Configuration Tool (on one of the running members), delete the member as
follows:
a. If necessary, click the triangle icon to expand the Cluster Nodes property.
b. Select the cluster node to be deleted. At the bottom of the right frame (labeled Properties),
click the Delete Node button.
c. Clicking the Delete Node button causes a warning dialog box to be displayed requesting
confirmation of the deletion (Figure 5.6, “Confirm Deleting a Member”).
e. Propagate the updated configuration by clicking the Send to Cluster button. (Propagating the
updated configuration automatically saves the configuration.)
4. Stop the cluster software on the remaining running nodes by running the following commands at
each node in this order:
c. service clvmd stop, if CLVM has been used to create clustered volumes
58
Configuring a Failover Domain
5. Start cluster software on all remaining cluster nodes by running the following commands in this
order:
b. service clvmd start, if CLVM has been used to create clustered volumes
6. Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify
that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services
are running as expected.
• Unrestricted — Allows you to specify that a subset of members are preferred, but that a cluster
service assigned to this domain can run on any available member.
• Restricted — Allows you to restrict the members that can run a particular cluster service. If none
of the members in a restricted failover domain are available, the cluster service cannot be started
(either manually or by the cluster software).
• Unordered — When a cluster service is assigned to an unordered failover domain, the member on
which the cluster service runs is chosen from the available failover domain members with no priority
ordering.
• Ordered — Allows you to specify a preference order among the members of a failover domain. The
member at the top of the list is the most preferred, followed by the second member in the list, and so
on.
Note
Note
In a cluster with several members, using a restricted failover domain can minimize the work to set up
the cluster to run a cluster service (such as httpd), which requires you to set up the configuration
identically on all members that run the cluster service). Instead of setting up the entire cluster to
run the cluster service, you must set up only the members in the restricted failover domain that you
associate with the cluster service.
59
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
Note
To configure a preferred member, you can create an unrestricted failover domain comprising only
one cluster member. Doing that causes a cluster service to run on that cluster member primarily
(the preferred member), but allows the cluster service to fail over to any of the other members.
The following sections describe adding a failover domain, removing a failover domain, and removing
members from a failover domain:
1. At the left frame of the Cluster Configuration Tool, click Failover Domains.
2. At the bottom of the right frame (labeled Properties), click the Create a Failover Domain button.
Clicking the Create a Failover Domain button causes the Add Failover Domain dialog box to be
displayed.
3. At the Add Failover Domain dialog box, specify a failover domain name at the Name for new
Failover Domain text box and click OK. Clicking OK causes the Failover Domain Configuration
dialog box to be displayed (Figure 5.7, “Failover Domain Configuration: Configuring a Failover
Domain”).
Note
The name should be descriptive enough to distinguish its purpose relative to other names
used in your cluster.
60
Adding a Failover Domain
4. Click the Available Cluster Nodes drop-down box and select the members for this failover
domain.
5. To restrict failover to members in this failover domain, click (check) the Restrict Failover To This
Domains Members checkbox. (With Restrict Failover To This Domains Members checked,
services assigned to this failover domain fail over only to nodes in this failover domain.)
6. To prioritize the order in which the members in the failover domain assume control of a failed
cluster service, follow these steps:
a. Click (check) the Prioritized List checkbox (Figure 5.8, “Failover Domain Configuration:
Adjusting Priority”). Clicking Prioritized List causes the Priority column to be displayed next
to the Member Node column.
61
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
b. For each node that requires a priority adjustment, click the node listed in the Member Node/
Priority columns and adjust priority by clicking one of the Adjust Priority arrows. Priority is
indicated by the position in the Member Node column and the value in the Priority column.
The node priorities are listed highest to lowest, with the highest priority node at the top of the
Member Node column (having the lowest Priority number).
8. At the Cluster Configuration Tool, perform one of the following actions depending on whether
the configuration is for a new cluster or for one that is operational and running:
• New cluster — If this is a new cluster, choose File => Save to save the changes to the cluster
configuration.
• Running cluster — If this cluster is operational and running, and you want to propagate the
change immediately, click the Send to Cluster button. Clicking Send to Cluster automatically
saves the configuration change. If you do not want to propagate the change immediately,
choose File => Save to save the changes to the cluster configuration.
1. At the left frame of the Cluster Configuration Tool, click the failover domain that you want to
delete (listed under Failover Domains).
2. At the bottom of the right frame (labeled Properties), click the Delete Failover Domain button.
Clicking the Delete Failover Domain button causes a warning dialog box do be displayed asking
62
Removing a Member from a Failover Domain
if you want to remove the failover domain. Confirm that the failover domain identified in the
warning dialog box is the one you want to delete and click Yes. Clicking Yes causes the failover
domain to be removed from the list of failover domains under Failover Domains in the left frame
of the Cluster Configuration Tool.
3. At the Cluster Configuration Tool, perform one of the following actions depending on whether
the configuration is for a new cluster or for one that is operational and running:
• New cluster — If this is a new cluster, choose File => Save to save the changes to the cluster
configuration.
• Running cluster — If this cluster is operational and running, and you want to propagate the
change immediately, click the Send to Cluster button. Clicking Send to Cluster automatically
saves the configuration change. If you do not want to propagate the change immediately,
choose File => Save to save the changes to the cluster configuration.
1. At the left frame of the Cluster Configuration Tool, click the failover domain that you want to
change (listed under Failover Domains).
2. At the bottom of the right frame (labeled Properties), click the Edit Failover Domain Properties
button. Clicking the Edit Failover Domain Properties button causes the Failover Domain
Configuration dialog box to be displayed (Figure 5.7, “Failover Domain Configuration:
Configuring a Failover Domain”).
3. At the Failover Domain Configuration dialog box, in the Member Node column, click the node
name that you want to delete from the failover domain and click the Remove Member from
Domain button. Clicking Remove Member from Domain removes the node from the Member
Node column. Repeat this step for each node that is to be deleted from the failover domain.
(Nodes must be deleted one at a time.)
5. At the Cluster Configuration Tool, perform one of the following actions depending on whether
the configuration is for a new cluster or for one that is operational and running:
• New cluster — If this is a new cluster, choose File => Save to save the changes to the cluster
configuration.
• Running cluster — If this cluster is operational and running, and you want to propagate the
change immediately, click the Send to Cluster button. Clicking Send to Cluster automatically
saves the configuration change. If you do not want to propagate the change immediately,
choose File => Save to save the changes to the cluster configuration.
1. On the Resources property of the Cluster Configuration Tool, click the Create a Resource
button. Clicking the Create a Resource button causes the Resource Configuration dialog box to
be displayed.
63
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
2. At the Resource Configuration dialog box, under Select a Resource Type, click the drop-down
box. At the drop-down box, select a resource to configure. Appendix C, HA Resource Parameters
describes resource parameters.
4. Choose File => Save to save the change to the /etc/cluster/cluster.conf configuration
file.
2. At the bottom of the right frame (labeled Properties), click the Create a Service button. Clicking
Create a Service causes the Add a Service dialog box to be displayed.
3. At the Add a Service dialog box, type the name of the service in the Name text box and click OK.
Clicking OK causes the Service Management dialog box to be displayed (refer to Figure 5.9,
“Adding a Cluster Service”).
Note
Use a descriptive name that clearly distinguishes the service from other services in the
cluster.
64
Adding a Cluster Service to the Cluster
4. If you want to restrict the members on which this cluster service is able to run, choose a failover
domain from the Failover Domain drop-down box. (Refer to Section 5.6, “Configuring a Failover
Domain” for instructions on how to configure a failover domain.)
5. Autostart This Service checkbox — This is checked by default. If Autostart This Service is
checked, the service is started automatically when a cluster is started and running. If Autostart
This Service is not checked, the service must be started manually any time the cluster comes up
from stopped state.
6. Run Exclusive checkbox — This sets a policy wherein the service only runs on nodes that have
no other services running on them. For example, for a very busy web server that is clustered for
high availability, it would would be advisable to keep that service on a node alone with no other
services competing for his resources — that is, Run Exclusive checked. On the other hand,
services that consume few resources (like NFS and Samba), can run together on the same node
without little concern over contention for resources. For those types of services you can leave the
Run Exclusive unchecked.
65
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
Note
Circumstances that require enabling Run Exclusive are rare. Enabling Run Exclusive can
render a service offline if the node it is running on fails and no other nodes are empty.
7. Select a recovery policy to specify how the resource manager should recover from a service
failure. At the upper right of the Service Management dialog box, there are three Recovery
Policy options available:
• Restart — Restart the service in the node the service is currently located. The default setting is
Restart. If the service cannot be restarted in the current node, the service is relocated.
• Relocate — Relocate the service before restarting. Do not restart the node where the service is
currently located.
8. Click the Add a Shared Resource to this service button and choose the a resource listed that
you have configured in Section 5.7, “Adding Cluster Resources”.
Note
9. If needed, you may also create a private resource that you can create that becomes a subordinate
resource by clicking on the Attach a new Private Resource to the Selection button. The process
is the same as creating a shared resource described in Section 5.7, “Adding Cluster Resources”.
The private resource will appear as a child to the shared resource to which you associated with
the shared resource. Click the triangle icon next to the shared resource to display any private
resources associated.
11. Choose File => Save to save the changes to the cluster configuration.
66
Propagating The Configuration File: New Cluster
Note
To verify the existence of the IP service resource used in a cluster service, you must use the /
sbin/ip addr list command on a cluster node. The following output shows the /sbin/ip
addr list command executed on a node running a cluster service:
2. Using the scp command, copy the /etc/cluster/cluster.conf file to all nodes in the
cluster.
Note
Propagating the cluster configuration file this way is necessary for the first time a cluster is
created. Once a cluster is installed and running, the cluster configuration file is propagated
using the Red Hat cluster management GUI Send to Cluster button. For more information
about propagating the cluster configuration using the GUI Send to Cluster button, refer to
Section 6.3, “Modifying the Cluster Configuration”.
2. service clvmd start, if CLVM has been used to create clustered volumes
67
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
Note
Shared storage for use in Red Hat Cluster Suite requires that you be running the cluster
logical volume manager daemon (clvmd) or the High Availability Logical Volume
Management agents (HA-LVM). If you are not able to use either the clvmd daemon or HA-
LVM for operational reasons or because you do not have the correct entitlements, you must
not use single-instance LVM on the shared disk as this may result in data corruption. If you
have any concerns please contact your Red Hat service representative.
5. Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify
that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services
are running as expected.
68
Chapter 6.
Note
2. service clvmd start, if CLVM has been used to create clustered volumes
To stop the cluster software on a member, type the following commands in this order:
3. service clvmd stop, if CLVM has been used to create clustered volumes
Stopping the cluster services on a member causes its services to fail over to an active member.
69
Chapter 6. Managing Red Hat Cluster With system-config-cluster
You can use the Cluster Status Tool to enable, disable, restart, or relocate a high-availability service.
The Cluster Status Tool displays the current cluster status in the Services area and automatically
updates the status every 10 seconds.
To enable a service, you can select the service in the Services area and click Enable. To disable a
service, you can select the service in the Services area and click Disable. To restart a service, you
can select the service in the Services area and click Restart. To relocate a service from one node to
another, you can drag the service to another node and drop the service onto that node. Relocating a
service restarts the service on that node. (Relocating a service to its current node — that is, dragging
a service to its current node and dropping the service onto that node — restarts the service.)
The following tables describe the members and services status information displayed by the Cluster
Status Tool.
70
Modifying the Cluster Configuration
Warning
Do not manually edit the contents of the /etc/cluster/cluster.conf file without guidance
from an authorized Red Hat representative or unless you fully understand the consequences of
editing the /etc/cluster/cluster.conf file manually.
Important
Although the Cluster Configuration Tool provides a Quorum Votes parameter in the
Properties dialog box of each cluster member, that parameter is intended only for use during
initial cluster configuration. Furthermore, it is recommended that you retain the default Quorum
Votes value of 1. For more information about using the Cluster Configuration Tool, refer to
Chapter 5, Configuring Red Hat Cluster With system-config-cluster.
To edit the cluster configuration file, click the Cluster Configuration tab in the cluster configuration
GUI. Clicking the Cluster Configuration tab displays a graphical representation of the cluster
configuration. Change the configuration file according the following steps:
71
Chapter 6. Managing Red Hat Cluster With system-config-cluster
2. Propagate the updated configuration file throughout the cluster by clicking Send to Cluster.
Note
The Cluster Configuration Tool does not display the Send to Cluster button if the cluster
is new and has not been started yet, or if the node from which you are running the Cluster
Configuration Tool is not a member of the cluster. If the Send to Cluster button is not
displayed, you can still use the Cluster Configuration Tool; however, you cannot propagate
the configuration. You can still save the configuration file. For information about using the
Cluster Configuration Tool for a new cluster configuration, refer to Chapter 5, Configuring
Red Hat Cluster With system-config-cluster.
3. Clicking Send to Cluster causes a Warning dialog box to be displayed. Click Yes to save and
propagate the configuration.
4. Clicking Yes causes an Information dialog box to be displayed, confirming that the current
configuration has been propagated to the cluster. Click OK.
5. Click the Cluster Management tab and verify that the changes have been propagated to the
cluster members.
Each time you save a configuration file, the Cluster Configuration Tool saves backup copies of
the three most recently used configuration files as /etc/cluster/cluster.conf.bak.1, /
etc/cluster/cluster.conf.bak.2, and /etc/cluster/cluster.conf.bak.3. The
backup file /etc/cluster/cluster.conf.bak.1 is the newest backup, /etc/cluster/
cluster.conf.bak.2 is the second newest backup, and /etc/cluster/cluster.conf.bak.3
is the third newest backup.
If a cluster member becomes inoperable because of misconfiguration, restore the configuration file
according to the following steps:
1. At the Cluster Configuration Tool tab of the Red Hat Cluster Suite management GUI, click File
=> Open.
2. Clicking File => Open causes the system-config-cluster dialog box to be displayed.
3. At the system-config-cluster dialog box, select a backup file (for example, /etc/cluster/
cluster.conf.bak.1). Verify the file selection in the Selection box and click OK.
4. Increment the configuration version beyond the current working version number as follows:
b. At the Cluster Properties dialog box, change the Config Version value and click OK.
72
Disabling the Cluster Software
6. Clicking File => Save As causes the system-config-cluster dialog box to be displayed.
8. Clicking OK causes an Information dialog box to be displayed. At that dialog box, click OK.
9. Propagate the updated configuration file throughout the cluster by clicking Send to Cluster.
Note
The Cluster Configuration Tool does not display the Send to Cluster button if the cluster
is new and has not been started yet, or if the node from which you are running the Cluster
Configuration Tool is not a member of the cluster. If the Send to Cluster button is not
displayed, you can still use the Cluster Configuration Tool; however, you cannot propagate
the configuration. You can still save the configuration file. For information about using the
Cluster Configuration Tool for a new cluster configuration, refer to Chapter 5, Configuring
Red Hat Cluster With system-config-cluster.
10. Clicking Send to Cluster causes a Warning dialog box to be displayed. Click Yes to propagate
the configuration.
11. Click the Cluster Management tab and verify that the changes have been propagated to the
cluster members.
Use the /sbin/chkconfig command to stop the member from joining the cluster at boot-up as
follows:
Once the problems with the disabled cluster member have been resolved, use the following
commands to allow the member to rejoin the cluster:
You can then reboot the member for the changes to take effect or run the following commands in the
order shown to restart cluster software:
73
Chapter 6. Managing Red Hat Cluster With system-config-cluster
2. service clvmd start, if CLVM has been used to create clustered volumes
74
Appendix A. Example of Setting Up
Apache HTTP Server
This appendix provides an example of setting up a highly available Apache HTTP Server on a Red Hat
Cluster. The example describes how to set up a service to fail over an Apache HTTP Server. Variables
in the example apply to this example only; they are provided to assist setting up a service that suits
your requirements.
Note
This example uses the Cluster Configuration Tool (system-config-cluster). You can use
comparable Conga functions to make an Apache HTTP Server highly available on a Red Hat
Cluster.
When installing the Apache HTTP Server on the cluster systems, run the following command to
ensure that the cluster nodes do not automatically start the service when the system boots:
Rather than having the system init scripts spawn the httpd daemon, the cluster infrastructure
initializes the service on the active cluster node. This ensures that the corresponding IP address and
file system mounts are active on only one cluster node at a time.
When adding an httpd service, a floating IP address must be assigned to the service so that the IP
address will transfer from one cluster node to another in the event of failover or service relocation.
The cluster infrastructure binds this IP address to the network interface on the cluster system that is
currently running the Apache HTTP Server. This IP address ensures that the cluster node running
httpd is transparent to the clients accessing the service.
The file systems that contain the Web content cannot be automatically mounted on the shared storage
resource when the cluster nodes boot. Instead, the cluster software must mount and unmount the
file system as the httpd service is started and stopped. This prevents the cluster systems from
accessing the same data simultaneously, which may result in data corruption. Therefore, do not
include the file systems in the /etc/fstab file.
1. On one cluster node, use the interactive parted utility to create a partition to use for the
document root directory. Note that it is possible to create multiple document root directories on
different disk partitions.
75
Appendix A. Example of Setting Up Apache HTTP Server
2. Use the mkfs command to create an ext3 file system on the partition you created in the previous
step. Specify the drive letter and the partition number. For example:
3. Mount the file system that contains the document root directory. For example:
Do not add this mount information to the /etc/fstab file because only the cluster software can
mount and unmount file systems used in a service.
5. If you have CGI files or other files that must be in different directories or in separate partitions,
repeat these steps, as needed.
On all node in the cluster (or nodes in the failover domain, if used), install the httpd RPM package.
For example:
To configure the Apache HTTP Server as a cluster service, perform the following tasks:
1. Edit the /etc/httpd/conf/httpd.conf configuration file and customize the file according to
your configuration. For example:
• Specify the directory that contains the HTML files. Also specify this mount point when adding the
service to the cluster configuration. It is only required to change this field if the mount point for
the web site's content differs from the default setting of /var/www/html/. For example:
DocumentRoot "/mnt/httpdservice/html"
• Specify a unique IP address to which the service will listen for requests. For example:
Listen 192.168.1.100:80
This IP address then must be configured as a cluster resource for the service using the Cluster
Configuration Tool.
• If the script directory resides in a non-standard location, specify the directory that contains the
CGI programs. For example:
76
Installing and Configuring the Apache HTTP Server
• Specify the path that was used in the previous step, and set the access permissions to default to
that directory. For example:
<Directory /mnt/httpdservice/cgi-bin">
AllowOverride None
Options None
Order allow,deny
Allow from all
</Directory>
Additional changes may need to be made to tune the Apache HTTP Server or add module
functionality. For information on setting up other options, refer to the Red Hat Enterprise Linux
System Administration Guide and the Red Hat Enterprise Linux Reference Guide.
2. The standard Apache HTTP Server start script, /etc/rc.d/init.d/httpd is also used within
the cluster framework to start and stop the Apache HTTP Server on the active cluster node.
Accordingly, when configuring the service, specify this script by adding it as a Script resource in
the Cluster Configuration Tool.
3. Copy the configuration file over to the other nodes of the cluster (or nodes of the failover domain, if
configured).
Before the service is added to the cluster configuration, ensure that the Apache HTTP Server
directories are not mounted. Then, on one node, invoke the Cluster Configuration Tool to add the
service, as follows. This example assumes a failover domain named httpd-domain was created for
this service.
1. Add the init script for the Apache HTTP Server service.
• Select the Resources tab and click Create a Resource. The Resources Configuration
properties dialog box is displayed.
• Specify the path to the Apache HTTP Server init script (for example, /etc/rc.d/init.d/
httpd) in the File (with path) field.
• Click OK.
2. Add a device for the Apache HTTP Server content files and/or custom scripts.
• In the Resource Configuration dialog, select File System from the drop-down menu.
• Enter the mount point in the Mount Point field (for example, /var/www/html/).
• Enter the device special file name in the Device field (for example, /dev/sda3).
77
Appendix A. Example of Setting Up Apache HTTP Server
• Enter the IP Address to be associated with the Apache HTTP Server service.
• Click OK.
• Click Create a Service. Type a Name for the service in the Add a Service dialog.
• In the Service Management dialog, select a Failover Domain from the drop-down menu or
leave it as None.
• Click the Add a Shared Resource to this service button. From the available list, choose each
resource that you created in the previous steps. Repeat this step until all resources have been
added.
• Click OK.
78
Appendix B. Fence Device Parameters
This appendix provides tables with parameter descriptions of fence devices.
Note
The Name parameter for a fence device specifies an arbitrary name for the device that will be
used by Red Hat Cluster Suite. This is not the same as the DNS name for the device.
Note
Certain fence devices have an optional Password Script parameter. The Password Script
parameter allows specifying that a fence-device password is supplied from a script rather
than from the Password parameter. Using the Password Script parameter supersedes the
Password parameter, allowing passwords to not be visible in the cluster configuration file (/etc/
cluster/cluster.conf).
Table B.2. APC Power Switch over SNMP (Red Hat Enterprise Linux 5.2 and later)
Field Description
Name A name for the APC device connected to the cluster into which the fence
daemon logs via the SNMP protocol.
IP Address The IP address or hostname assigned to the device.
UDP/TCP port The UDP/TCP port to use for connection with the device; the default value is
161.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password Script The script that supplies a password for access to the fence device. Using
(optional) this supersedes the Password parameter.
79
Appendix B. Fence Device Parameters
Field Description
Port The port.
Switch (optional) The switch number for the APC switch that connects to the node when you
have multiple daisy-chained switches.
SNMP version The SNMP version to use (1, 2c, 3); the default value is 1.
SNMP community The SNMP community string; the default value is private.
SNMP security level The SNMP security level (noAuthNoPriv, authNoPriv, authPriv).
SNMP authentication The SNMP authentication protocol (MD5, SHA).
protocol
SNMP privacy The SNMP privacy protocol (DES, AES).
protocol
SNMP privacy The SNMP privacy protocol password.
protocol password
SNMP privacy The script that supplies a password for SNMP privacy protocol. Using this
protocol script supersedes the SNMP privacy protocol password parameter.
Power wait Number of seconds to wait after issuing a power off or power on command.
fence_apc_snmp The fence agent for APC that logs into the SNP device via the SNMP
protocol.
80
Table B.5. Cisco MDS (Red Hat Enterprise Linux 5.4 and later)
Field Description
Name A name for the Cisco MDS 9000 series device with SNMP enabled.
IP Address The IP address or hostname assigned to the device.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password Script The script that supplies a password for access to the fence device. Using
(optional) this supersedes the Password parameter.
Port The port.
SNMP version The SNMP version to use (1, 2c, 3).
SNMP community The SNMP community string.
SNMP authentication The SNMP authentication protocol (MD5, SHA).
protocol
SNMP security level The SNMP security level (noAuthNoPriv, authNoPriv, authPriv).
SNMP privacy The SNMP privacy protocol (DES, AES).
protocol
SNMP privacy The SNMP privacy protocol password.
protocol password
SNMP privacy The script that supplies a password for SNMP privacy protocol. Using this
protocol script supersedes the SNMP privacy protocol password parameter.
Power wait Number of seconds to wait after issuing a power off or power on command.
fence_cisco_mds The fence agent for Cisco MDS.
Table B.6. Cisco UCS (Red Hat Enterprise Linux 5.6 and later)
Field Description
Name A name for the Cisco UCS device.
IP Address The IP address or hostname assigned to the device.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password Script The script that supplies a password for access to the fence device. Using
(optional) this supersedes the Password parameter.
SSL The SSL connection.
IP port (optional) The TCP port to use to connect to the device.
Port Name of virtual machine.
Power wait Number of seconds to wait after issuing a power off or power on command.
Power timeout Number of seconds to test for a status change after issuing a power off or
power on command.
Shell timeout Number of seconds to wait for a command prompt after issuing a command.
Retry on Number of attempts to retry power on.
fence_cisco_ucs The fence agent for Cisco UCS.
81
Appendix B. Fence Device Parameters
82
Field Description
IP address The cluster name of the node to be fenced. Refer to the fence_gnbd(8) man page
for more information.
fence_gnbd The fence agent for GNBD-based GFS clusters.
Table B.12. HP iLO (Integrated Lights Out) MP (Red Hat Enterprise Linux 5.5 and later)
Field Description
Name A name for the server with HP iLO support.
Hostname The hostname assigned to the device.
IP port (optional) TCP port to use for connection with the device.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password Script The script that supplies a password for access to the fence device. Using
(optional) this supersedes the Password parameter.
SSH (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH
to access the device.
Path to the SSH The identity file for SSH.
identity file
Force command The command prompt to use. The default value is ’MP>’, ’hpiLO->’.
prompt
Power wait Number of seconds to wait after issuing a power off or power on command.
fence_ilo_mp The fence agent for HP iLO MP devices.
83
Appendix B. Fence Device Parameters
Field Description
Blade The blade of the device.
Use SSH (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to
access the device.
The fence agent for IBm BladeCenter.
fence_bladecenter
Table B.15. IF MIB (Red Hat Enterprise Linux 5.6 and later)
Field Description
Name A name for the IF MIB device connected to the cluster.
IP Address The IP address or hostname assigned to the device.
UDP/TCP The UDP/TCP port to use for connection with the device; the default value is
port(optiona) 161.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password Script The script that supplies a password for access to the fence device. Using
(optional) this supersedes the Password parameter.
SNMP version The SNMP version to use (1, 2c, 3); the default value is 1.
SNMP community The SNMP community string.
SNMP security level The SNMP security level (noAuthNoPriv, authNoPriv, authPriv).
SNMP authentication The SNMP authentication protocol (MD5, SHA).
protocol
SNMP privacy The SNMP privacy protocol (DES, AES).
protocol
SNMP privacy The SNMP privacy protocol password.
protocol password
SNMP privacy The script that supplies a password for SNMP privacy protocol. Using this
protocol script supersedes the SNMP privacy protocol password parameter.
Power timeout Number of seconds to test for a status change after issuing a power off or
power on command.
Shell timeout Number of seconds to wait for a command prompt after issuing a command.
Login timeout Number of seconds to wait for a command prompt after login.
Power wait Number of seconds to wait after issuing a power off or power on command.
84
Field Description
Retry on Number of attempts to retry power on.
Port Physical plug number or name of virtual machine.
fence_ifmib The fence agent for IF-MIB devices.
Warning
85
Appendix B. Fence Device Parameters
Field Description
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password The script that supplies a password for access to the fence device. Using this
Script supersedes the Password parameter.
(optional)
Port The switch outlet number.
fence_sanbox2The fence agent for QLogic SANBox2 FC switches.
Note
Use of SCSI persistent reservations as a fence method is supported with the following limitations:
• As of Red Hat Enterprise Linux 5.5 and fully-updated releases of Red Hat Enterprise Linux 5.4,
SCSI fencing can be used in a 2-node cluster; previous releases did not support this feature.
• When using SCSI fencing, all nodes in the cluster must register with the same devices so that
each node can remove another node's registration key from all the devices it is registered with.
• Devices used for the cluster volumes should be a complete LUN, not partitions. SCSI
persistent reservations work on an entire LUN, meaning that access is controlled to each LUN,
not individual partitions.
• As of Red Hat Enterprise Linux 5.5 and fully-updated releases of Red Hat Enterprise Linux
5.4, SCSI fencing can be used in conjunction with qdisk; previous releases did not support this
feature. You cannot use fence_scsi on the LUN where qdiskd resides; it must be a raw
LUN or raw partition of a LUN.
86
Table B.23. VMware (SOAP Interface) (Red Hat Enterprise Linux 5.7 and later)
Field Description
Name Name of the virtual machine fencing device.
Hostname The IP address or hostname assigned to the device.
IP Port The TCP port to use for connection with the device.
Login The login name used to access the device.
Password The password used to authenticate the connection to the device.
Password The script that supplies a password for access to the fence device. Using this
Script supersedes the Password parameter.
(optional)
Use SSL Use SSL connections to communicate with the device.
connections
Power wait Number of seconds to wait after issuing a power off or power on command.
Virtual Name of virtual machine in inventory path format (e.g., /datacenter/vm/
machine name Discovered_virtual_machine/myMachine).
Virtual The UUID of the virtual machine to fence.
machine UUID
The fence agent for VMWare over SOAP API.
fence_vmware_soap
87
88
Appendix C. HA Resource Parameters
This appendix provides descriptions of HA resource parameters. You can configure the parameters
with Luci, system-config-cluster, or by editing etc/cluster/cluster.conf. Table C.1, “HA
Resource Summary” lists the resources, their corresponding resource agents, and references to other
tables containing parameter descriptions. To understand resource agents in more detail you can view
them in /usr/share/cluster of any cluster node.
For a comprehensive list and description of cluster.conf elements and attributes, refer
to the cluster schema at /usr/share/system-config-cluster/misc/cluster.ng,
and the annotated schema at /usr/share/doc/system-config-cluster-X.Y.ZZ/
cluster_conf.html (for example /usr/share/doc/system-config-cluster-1.0.57/
cluster_conf.html).
89
Appendix C. HA Resource Parameters
Field Description
Server Root The default value is /etc/httpd.
Config File Specifies the Apache configuration file. The default valuer is /etc/httpd/conf.
httpd Options Other command line options for httpd.
Shutdown Wait Specifies the number of seconds to wait for correct end of service shutdown.
(seconds)
When creating a new file system resource, you can leave this field blank. Leaving
the field blank causes a file system ID to be assigned automatically after you
commit the parameter during configuration. If you need to assign a file system ID
explicitly, specify it in this field.
Force unmount If enabled, forces the file system to unmount. The default setting is disabled.
Force Unmount kills all processes using the mount point to free up the mount
when it tries to unmount.
Reboot If enabled, reboots the node if unmounting this file system fails. The default setting
host node if is disabled.
unmount fails
Check file If enabled, causes fsck to be run on the file system before mounting it. The
system before default setting is disabled.
mounting
90
Field Description
File System ID Note
When creating a new GFS resource, you can leave this field blank. Leaving the
field blank causes a file system ID to be assigned automatically after you commit
the parameter during configuration. If you need to assign a file system ID explicitly,
specify it in this field.
Force If enabled, forces the file system to unmount. The default setting is disabled.
Unmount Force Unmount kills all processes using the mount point to free up the mount
when it tries to unmount. With GFS resources, the mount point is not unmounted at
service tear-down unless Force Unmount is enabled.
Reboot If enabled and unmounting the file system fails, the node will immediately reboot.
Host Node if Generally, this is used in conjunction with force-unmount support, but it is not
Unmount Fails required.
(self fence)
91
Appendix C. HA Resource Parameters
Tip
Name the NFS Export resource so it is clearly distinguished from other NFS
resources.
Note
92
Table C.11. Open LDAP
Field Description
Name Specifies a service name for logging and other purposes.
Config File Specifies an absolute path to a configuration file. The default value is /etc/
openldap/slapd.conf.
URL List The default value is ldap:///.
slapd Options Other command line options for slapd.
Shutdown Wait Specifies the number of seconds to wait for correct end of service shutdown.
(seconds)
93
Appendix C. HA Resource Parameters
Field Description
ABAP stack is If you do not have an ABAP stack installed in the SAP database, enable this
not installed, parameter.
only Java stack
is installed
J2EE instance The fully qualified path the J2EE instance bootstrap directory. For example, /usr/
bootstrap sap/P01/J00/j2ee/cluster/bootstrap.
directory
J2EE security The fully qualified path the J2EE security store directory. For example, /usr/sap/
store path P01/SYS/global/security/lib/tools.
Note
Regarding Table C.16, “Samba Service”, when creating or editing a cluster service, connect a
Samba-service resource directly to the service, not to a resource within a service.
94
Field Description
Automatically If enabled, this service (or resource group) is started automatically after the
start this cluster forms a quorum. If this parameter is disabled, this service is not started
service automatically after the cluster forms a quorum; the service is put into the
disabled state.
Run exclusive If enabled, this service (resource group) can only be relocated to run on another
node exclusively; that is, to run on a node that has no other services running on it.
If no nodes are available for a service to run exclusively, the service is not restarted
after a failure. Additionally, other services do not automatically relocate to a node
running this service as Run exclusive. You can override this option by manual
start or relocate operations.
Failover Defines lists of cluster members to try in the event that a service fails.
Domain
Recovery Recovery policy provides the following options:
policy
• Disable — Disables the resource group if any component fails.
• Relocate — Tries to restart service in another node; that is, it does not try to
restart in the current node.
• Restart — Tries to restart failed parts of this service locally (in the current
node) before trying to relocate (default) to service to another node.
• Restart-Disable — (Red Hat Enterprise Linux release 5.6 and later) The
service will be restarted in place if it fails. However, if restarting the service fails
the service will be disabled instead of being moved to another host in the cluster.
95
Appendix C. HA Resource Parameters
Field Description
Tomcat User User who runs the Tomcat server. The default value is tomcat.
Catalina Other command line options for Catalina.
Options
Catalina Base Catalina base directory (differs for each service) The default value is /usr/share/
tomcat5.
Shutdown Wait Specifies the number of seconds to wait for correct end of service shutdown. The
(seconds) default value is 30.
Important
The path should never directly point to a virtual machine configuration file.
Automatically If enabled, this virtual machine is started automatically after the cluster forms
start this virtual a quorum. If this parameter is disabled, this virtual machine is not started
machine automatically after the cluster forms a quorum; the virtual machine is put into the
disabled state.
Run exclusive If enabled, this virtual machine can only be relocated to run on another node
exclusively; that is, to run on a node that has no other virtual machines running
on it. If no nodes are available for a virtual machine to run exclusively, the virtual
machine is not restarted after a failure. Additionally, other virtual machines do not
automatically relocate to a node running this virtual machine as Run exclusive.
You can override this option by manual start or relocate operations.
Failover Defines lists of cluster members to try in the event that a virtual machine fails.
Domain
Recovery Recovery policy provides the following options:
policy
• Disable — Disables the virtual machine if it fails.
• Relocate — Tries to restart the virtual machine in another node; that is, it does
not try to restart in the current node.
• Restart — Tries to restart the virtual machine locally (in the current node)
before trying to relocate (default) to virtual machine to another node.
• Restart-Disable — (Red Hat Enterprise Linux Release 5.6 and later) The
service will be restarted in place if it fails. However, if restarting the service fails
the service will be disabled instead of being moved to another host in the cluster.
Migration type Specifies a migration type of live or pause. The default setting is live.
96
Appendix D. HA Resource Behavior
This appendix describes common behavior of HA resources. It is meant to provide ancillary
information that may be helpful in configuring HA services. You can configure the parameters with
Luci, system-config-cluster, or by editing etc/cluster/cluster.conf. For descriptions
of HA resource parameters, refer to Appendix C, HA Resource Parameters. To understand resource
agents in more detail you can view them in /usr/share/cluster of any cluster node.
Note
To fully comprehend the information in this appendix, you may require detailed understanding of
resource agents and the cluster configuration file, /etc/cluster/cluster.conf.
An HA service is a group of cluster resources configured into a coherent entity that provides
specialized services to clients. An HA service is represented as a resource tree in the cluster
configuration file, /etc/cluster/cluster.conf (in each cluster node). In the cluster configuration
file, each resource tree is an XML representation that specifies each resource, its attributes, and its
relationship among other resources in the resource tree (parent, child, and sibling relationships).
Note
At the root of each resource tree is a special type of resource — a service resource. Other types of
resources comprise the rest of a service, determining its characteristics. Configuring an HA service
consists of creating a service resource, creating subordinate cluster resources, and organizing them
into a coherent entity that conforms to hierarchical restrictions of the service.
Note
The sections that follow present examples from the cluster configuration file, /etc/cluster/
cluster.conf, for illustration purposes only.
97
Appendix D. HA Resource Behavior
Example D.1, “Resource Hierarchy of Service foo” shows a sample resource tree of the service foo. In
the example, the relationships among the resources are as follows:
• fs:myfs (<fs name="myfs" ...>) and ip:10.1.1.2 (<ip address="10.1.1.2 .../>) are siblings.
• For a resource to be considered in good health, all its children must be in good health.
• Designates child-type attribute (typed child resource) — If the Service resource designates a child-
type attribute for a child resource, the child resource is typed. The child-type attribute explicitly
determines the start and the stop order of the child resource.
• Does not designate child-type attribute (non-typed child resource) — If the Service resource does
not designate a child-type attribute for a child resource, the child resource is non-typed. The Service
resource does not explicitly control the starting order and stopping order of a non-typed child
resource. However, a non-typed child resource is started and stopped according to its order in /
etc/cluster.cluster.conf In addition, non-typed child resources are started after all typed
child resources have started and are stopped before any typed child resources have stopped.
98
Typed Child Resource Start and Stop Ordering
Note
The only resource to implement defined child resource type ordering is the Service resource.
For more information about typed child resource start and stop ordering, refer to Section D.2.1, “Typed
Child Resource Start and Stop Ordering”. For more information about non-typed child resource start
and stop ordering, refer to Section D.2.2, “Non-typed Child Resource Start and Stop Ordering”.
Example D.2. Resource Start and Stop Values: Excerpt from Service Resource Agent,
service.sh
<special tag="rgmanager">
<attributes root="1" maxinstances="1"/>
<child type="lvm" start="1" stop="9"/>
<child type="fs" start="2" stop="8"/>
<child type="clusterfs" start="3" stop="7"/>
<child type="netfs" start="4" stop="6"/>
<child type="nfsexport" start="5" stop="5"/>
<child type="nfsclient" start="6" stop="4"/>
<child type="ip" start="7" stop="2"/>
<child type="smb" start="8" stop="3"/>
<child type="script" start="9" stop="1"/>
</special>
99
Appendix D. HA Resource Behavior
Ordering within a resource type is preserved as it exists in the cluster configuration file, /etc/
cluster/cluster.conf. For example, consider the starting order and stopping order of the typed
child resources in Example D.3, “Ordering Within a Resource Type”.
<service name="foo">
<script name="1" .../>
<lvm name="1" .../>
<ip address="10.1.1.1" .../>
<fs name="1" .../>
<lvm name="2" .../>
</service>
1. lvm:1 — This is an LVM resource. All LVM resources are started first. lvm:1 (<lvm
name="1" .../>) is the first LVM resource started among LVM resources because it is the first
LVM resource listed in the Service foo portion of /etc/cluster/cluster.conf.
2. lvm:2 — This is an LVM resource. All LVM resources are started first. lvm:2 (<lvm
name="2" .../>) is started after lvm:1 because it is listed after lvm:1 in the Service foo
portion of /etc/cluster/cluster.conf.
3. fs:1 — This is a File System resource. If there were other File System resources in Service foo,
they would start in the order listed in the Service foo portion of /etc/cluster/cluster.conf.
5. script:1 — This is a Script resource. If there were other Script resources in Service foo, they
would start in the order listed in the Service foo portion of /etc/cluster/cluster.conf.
1. script:1 — This is a Script resource. If there were other Script resources in Service foo,
they would stop in the reverse order listed in the Service foo portion of /etc/cluster/
cluster.conf.
3. fs:1 — This is a File System resource. If there were other File System resources in Service
foo, they would stop in the reverse order listed in the Service foo portion of /etc/cluster/
cluster.conf.
4. lvm:2 — This is an LVM resource. All LVM resources are stopped last. lvm:2 (<lvm
name="2" .../>) is stopped before lvm:1; resources within a group of a resource type are
stopped in the reverse order listed in the Service foo portion of /etc/cluster/cluster.conf.
100
Non-typed Child Resource Start and Stop Ordering
5. lvm:1 — This is an LVM resource. All LVM resources are stopped last. lvm:1 (<lvm
name="1" .../>) is stopped after lvm:2; resources within a group of a resource type are
stopped in the reverse order listed in the Service foo portion of /etc/cluster/cluster.conf.
For example, consider the starting order and stopping order of the non-typed child resources in
Example D.4, “Non-typed and Typed Child Resource in a Service”.
<service name="foo">
<script name="1" .../>
<nontypedresource name="foo"/>
<lvm name="1" .../>
<nontypedresourcetwo name="bar"/>
<ip address="10.1.1.1" .../>
<fs name="1" .../>
<lvm name="2" .../>
</service>
1. lvm:1 — This is an LVM resource. All LVM resources are started first. lvm:1 (<lvm
name="1" .../>) is the first LVM resource started among LVM resources because it is the first
LVM resource listed in the Service foo portion of /etc/cluster/cluster.conf.
2. lvm:2 — This is an LVM resource. All LVM resources are started first. lvm:2 (<lvm
name="2" .../>) is started after lvm:1 because it is listed after lvm:1 in the Service foo
portion of /etc/cluster/cluster.conf.
3. fs:1 — This is a File System resource. If there were other File System resources in Service foo,
they would start in the order listed in the Service foo portion of /etc/cluster/cluster.conf.
5. script:1 — This is a Script resource. If there were other Script resources in Service foo, they
would start in the order listed in the Service foo portion of /etc/cluster/cluster.conf.
101
Appendix D. HA Resource Behavior
3. script:1 — This is a Script resource. If there were other Script resources in Service foo,
they would stop in the reverse order listed in the Service foo portion of /etc/cluster/
cluster.conf.
5. fs:1 — This is a File System resource. If there were other File System resources in Service
foo, they would stop in the reverse order listed in the Service foo portion of /etc/cluster/
cluster.conf.
6. lvm:2 — This is an LVM resource. All LVM resources are stopped last. lvm:2 (<lvm
name="2" .../>) is stopped before lvm:1; resources within a group of a resource type are
stopped in the reverse order listed in the Service foo portion of /etc/cluster/cluster.conf.
7. lvm:1 — This is an LVM resource. All LVM resources are stopped last. lvm:1 (<lvm
name="1" .../>) is stopped after lvm:2; resources within a group of a resource type are
stopped in the reverse order listed in the Service foo portion of /etc/cluster/cluster.conf.
102
Failure Recovery and Independent Subtrees
Example D.5. NFS Service Set Up for Resource Reuse and Inheritance
<resources>
<nfsclient name="bob" target="bob.example.com" options="rw,no_root_squash"/>
<nfsclient name="jim" target="jim.example.com" options="rw,no_root_squash"/>
<nfsexport name="exports"/>
</resources>
<service name="foo">
<fs name="1" mountpoint="/mnt/foo" device="/dev/sdb1" fsid="12344">
<nfsexport ref="exports"> <!-- nfsexport's path and fsid
attributes are inherited from the mountpoint and fsid
attribute of the parent fs resource -->
<nfsclient ref="bob"/> <!-- nfsclient's path is inherited
from the mountpoint and the fsid is added to the options
string during export -->
<nfsclient ref="jim"/ >
</nfsexport>
</fs>
<fs name="2" mountpoint="/mnt/bar" device="/dev/sdb2" fsid="12345">
<nfsexport ref="exports">
<nfsclient ref="bob"/> <!-- Because all of the critical
data for this resource is either defined in the resources block
or inherited, we can reference it again! -->
<nfsclient ref="jim"/>
</nfsexport>
</fs>
<ip address="10.2.13.20"/>
</service>
If the service were flat (that is, with no parent/child relationships), it would need to be configured as
follows:
• The service would need four nfsclient resources — one per file system (a total of two for file
systems), and one per target machine (a total of two for target machines).
• The service would need to specify export path and file system ID to each nfsclient, which introduces
chances for errors in the configuration.
In Example D.5, “NFS Service Set Up for Resource Reuse and Inheritance” however, the NFS
client resources nfsclient:bob and nfsclient:jim are defined once; likewise, the NFS export resource
nfsexport:exports is defined once. All the attributes needed by the resources are inherited from parent
resources. Because the inherited attributes are dynamic (and do not conflict with one another), it is
possible to reuse those resources — which is why they are defined in the resources block. It may not
be practical to configure some resources in multiple places. For example, configuring a file system
resource in multiple places can result in mounting one file system on two nodes, therefore causing
problems.
103
Appendix D. HA Resource Behavior
<service name="foo">
<script name="script_one" ...>
<script name="script_two" .../>
</script>
<script name="script_three" .../>
</service>
<service name="foo">
<script name="script_one" __independent_subtree="1" ...>
<script name="script_two" __independent_subtree="1" .../>
<script name="script_three" .../>
</script>
<script name="script_four" .../>
</service>
In some circumstances, if a component of a service fails you may want to disable only that component
without disabling the entire service, to avoid affecting other services the use other components of
that service. As of the Red Hat Enterprise Linux 5.6 release, you can accomplish that by using the
__independent_subtree="2" attribute, which designates the independent subtree as non-critical.
Note
You may only use the non-critical flag on singly-referenced resources. The non-critical flag works
with all resources at all levels of the resource tree, but should not be used at the top level when
defining services or virtual machines.
As of the Red Hat Enterprise Linux 5.6 release, you can set maximum restart and restart expirations
on a per-node basis in the resource tree for independent subtrees. To set these thresholds, you can
use the following attributes:
• __max_restarts configures the maximum number of tolerated restarts prior to giving up.
104
Debugging and Testing Services and Resource Ordering
Start a service:
Stop a service:
105
Appendix D. HA Resource Behavior
Action Syntax
cluster.conf
files.
106
Appendix E. Cluster Service Resource
Check and Failover Timeout
This appendix describes how rgmanager monitors the status of cluster resources, and how to modify
the status check interval. The appendix also describes the _enforce_timeouts service parameter,
which indicates that a timeout for an operation should cause a service to fail.
Note
To fully comprehend the information in this appendix, you may require detailed understanding
of resource agents and the cluster configuration file, /etc/cluster/cluster.conf. For
a comprehensive list and description of cluster.conf elements and attributes, refer to
the cluster schema at /usr/share/system-config-cluster/misc/cluster.ng,
and the annotated schema at /usr/share/doc/system-config-cluster-X.Y.ZZ/
cluster_conf.html (for example /usr/share/doc/system-config-cluster-1.0.57/
cluster_conf.html).
Each resource agent specifies the amount of time between periodic status checks. Each resource
utilizes these timeout values unless explicitly overridden in the cluster.conf file using the special
<action> tag:
This tag is a special child of the resource itself in the cluster.conf file. For example, if you had a
file system resource for which you wanted to override the status check interval you could specify the
file system resource in the cluster.conf file as follows:
Some agents provide multiple "depths" of checking. For example, a normal file system status check
(depth 0) checks whether the file system is mounted in the correct place. A more intensive check
is depth 10, which checks whether you can read a file from the file system A more intensive check
is depth 20, which checks whether you can write to the file system. In the example given here, the
depth is set to *, which indicates that these values should be used for all depths. The result is that
the test file system is checked at the highest-defined depth provided by the resource-agent (in this
case, 20) every 10 seconds.
107
Appendix E. Cluster Service Resource Check and Failover Timeout
The following example shows a cluster service that has been configured with the
_enforce_timeouts attribute set for the netfs resource. With this attribute set, then during
recovery if it takes more than 30 seconds to unmount the NFS file system the operation will time out,
causing the service to enter the failed state.
</screen>
<rm>
<failoverdomains/>
<resources>
<netfs export="/nfstest" force_unmount="1" fstype="nfs" host="10.65.48.65"
mountpoint="/data/nfstest" name="nfstest_data" options="rw,sync,soft"/>
</resources>
<service autostart="1" exclusive="0" name="nfs_client_test" recovery="relocate">
<netfs ref="nfstest_data" __enforce_timeouts="1"/>
</service>
</rm>
108
Appendix F. Upgrading A Red Hat
Cluster from RHEL 4 to RHEL 5
This appendix provides a procedure for upgrading a Red Hat cluster from RHEL 4 to RHEL 5. The
procedure includes changes required for Red Hat GFS and CLVM, also. For more information about
Red Hat GFS, refer to Global File System: Configuration and Administration. For more information
about LVM for clusters, refer to LVM Administrator's Guide: Configuration and Administration.
Upgrading a Red Hat Cluster from RHEL 4 to RHEL 5 consists of stopping the cluster, converting the
configuration from a GULM cluster to a CMAN cluster (only for clusters configured with the GULM
cluster manager/lock manager), adding node IDs, and updating RHEL and cluster software. To
upgrade a Red Hat Cluster from RHEL 4 to RHEL 5, follow these steps:
c. Run service gfs stop, if you are using Red Hat GFS.
d. Run service clvmd stop, if CLVM has been used to create clustered volumes.
Note
The error message is the expected result when running service clvmd stop after
clvmd has stopped.
e. Depending on the type of cluster manager (either CMAN or GULM), run the following
command or commands:
3. Disable cluster software from starting during reboot. At each node, run /sbin/chkconfig as
follows:
109
Appendix F. Upgrading A Red Hat Cluster from RHEL 4 to RHEL 5
b. If your cluster is configured with GULM as the cluster manager, remove the GULM
XML elements — <gulm> and </gulm> — and their content from /etc/cluster/
cluster.conf. GULM is not supported in Red Hat Cluster Suite for RHEL 5. Example F.1,
“GULM XML Elements and Content” shows an example of GULM XML elements and content.
c. At the <clusternode> element for each node in the configuration file, insert
nodeid="number" after name="name". Use a number value unique to that node. Inserting
it there follows the format convention of the <clusternode> element in a RHEL 5 cluster
configuration file.
Note
The nodeid parameter is required in Red Hat Cluster Suite for RHEL 5. The parameter
is optional in Red Hat Cluster Suite for RHEL 4. If your configuration file already contains
nodeid parameters, skip this step.
d. When you have completed editing /etc/cluster/cluster.conf, save the file and copy it
to the other nodes in the cluster (for example, using the scp command).
5. If your cluster is a GULM cluster and uses Red Hat GFS, change the superblock of each GFS file
system to use the DLM locking protocol. Use the gfs_tool command with the sb and proto
options, specifying lock_dlm for the DLM locking protocol:
For example:
6. Update the software in the cluster nodes to RHEL 5 and Red Hat Cluster Suite for RHEL 5. You
can acquire and update software through Red Hat Network channels for RHEL 5 and Red Hat
Cluster Suite for RHEL 5.
8. Enable cluster software to start upon reboot. At each node run /sbin/chkconfig as follows:
110
# chkconfig --level 2345 gfs on
# chkconfig --level 2345 clvmd on
# chkconfig --level 2345 cman on
9. Reboot the nodes. The RHEL 5 cluster software should start while the nodes reboot. Upon
verification that the Red Hat cluster is running, the upgrade is complete.
<gulm>
<lockserver name="gulmserver1"/>
<lockserver name="gulmserver2"/>
<lockserver name="gulmserver3"/>
</gulm>
111
112
Appendix G. Revision History
Revision 5.7-1 Thu Jul 21 2011 Steven Levine [email protected]
Resolves: #713256
Documents new fence_vmware_soap agent.
Resolves: #446137
Documents procedure to configure a system to listen to luci from the internal network only.
Resolves: #515858
Provides information about cluster service status check and failover timeout.
Resolves: #560558
Provides rules to allow multicast traffic for cluster comunication
Resolves: #705131
Updates tables of fence agent parameters to reflect Red Hat Enterprise Linux 5.7 support.
Resolves: #705134
Documents non-critical resources and restart-disable configuration parameter.
Resolves: #480292
Adds pointer to cluster.conf schema in documentation of resource parameters.
Resolves: #515860
Updates example domains.
Resolves: #595711
Fixes minor typographical errors.
Resolves: #654176
Fixes minor typographical errors.
Resolves: #675809
Fixes incorrect table title reference.
Resolves: #527473
Adds information about cluster node-count limit.
Resolves: #568179
Adds information about support of and GFS/GFS2 deployment.
Resolves: #568483
Adds general support statement.
Resolves: #526540
113
Appendix G. Revision History
Resolves: #482936
Corrects Section 5.7 title and intro text.
Resolves: #488751
Corrects iptables rules. Removed examples.
Resolves: #502053
Corrects iptables rules for rgmanager.
Resolves: #511150
Adds content stating that SELinux must be disabled for Red Hat Cluster Suite.
Resolves: #513072
Adds information about limitations on using SCSI reservations as a fencing method.
Resolves: #450777
Includes content about configuring failover domains to not fail back a service (an added feature).
114
Index cluster resource status check, 107
cluster resource types, 20
cluster service
A displaying status, 9, 70
ACPI cluster service managers
configuring, 14 configuration, 38, 64, 67
Apache HTTP Server cluster services, 38, 64
httpd.conf , 76 (see also adding to the cluster configuration)
setting up service, 75 Apache HTTP Server, setting up, 75
httpd.conf , 76
B cluster software
behavior, HA resources, 97 configuration, 25
disabling, 73
C installation and configuration, 47
cluster starting and stopping, 69
administration, 11, 43, 69 cluster software installation and configuration, 47
diagnosing and correcting problems, 46, 74 cluster storage
disabling the cluster software, 73 configuration, 40
displaying status, 9, 70 command line tools table, 9
managing node, 44 configuration
starting, 67 HA service, 18
starting, stopping, restarting, and deleting, 43 configuration file
cluster administration, 11, 43, 69 propagation of, 67
backing up the cluster database, 72 configuring cluster storage , 40
compatible hardware, 12 Conga
configuring ACPI, 14 accessing, 2
configuring iptables, 12 considerations for cluster administration, 23
configuring max_luns, 20 overview, 4
Conga considerations, 23 Conga overview, 4
considerations for using qdisk, 21
considerations for using quorum disk, 21 F
diagnosing and correcting problems in a failover timeout, 107
cluster, 46, 74 feedback, viii, viii
disabling the cluster software, 73
displaying cluster and service status, 9, 70 G
enabling IP ports, 12
general
general considerations, 11
considerations for cluster administration, 11
managing cluster node, 44
managing high-availability services, 45
modifying the cluster configuration, 71
H
network switches and multicast addresses, 22 HA service configuration
restoring the cluster database, 72 overview, 18
SELinux, 22 hardware
starting and stopping the cluster software, 69 compatible, 12
starting, stopping, restarting, and deleting a HTTP services
cluster, 43 Apache HTTP Server
cluster configuration, 25 httpd.conf, 76
modifying, 71 setting up, 75
Cluster Configuration Tool
accessing, 8 I
cluster database integrated fence devices
backing up, 72 configuring ACPI, 14
restoring, 72 introduction, v
cluster resource relationships, 98 other Red Hat Enterprise Linux documents, v
115
Index
IP ports
enabling, 12
iptables
configuring, 12
iptables firewall, 23
M
max_luns
configuring, 20
multicast addresses
considerations for using with network switches
and multicast addresses, 22
multicast traffic, enabling, 23
P
parameters, fence device, 79
parameters, HA resources, 89
power controller connection, configuring, 79
power switch, 79
(see also power controller)
Q
qdisk
considerations for using, 21
quorum disk
considerations for using, 21
R
relationships
cluster resource, 98
S
SELinux
configuring, 22
starting the cluster software, 67
status check, cluster resource, 107
System V init , 69
T
table
command line tools, 9
tables
HA resources, parameters, 89
power controller connection, configuring, 79
timeout failover, 107
troubleshooting
diagnosing and correcting problems in a
cluster, 46, 74
types
cluster resource, 20
U
upgrading, RHEL 4 to RHEL 5, 109
116