CP R80.20 ClusterXL AdminGuide PDF
CP R80.20 ClusterXL AdminGuide PDF
CLUSTERXL
R80.20
Administration Guide
Classification: [Protected]
© 2018 Check Point Software Technologies Ltd.
All rights reserved. This product and related documentation are protected by copyright and
distributed under licensing restricting their use, copying, distribution, and decompilation. No part
of this product or related documentation may be reproduced in any form or by any means without
prior written authorization of Check Point. While every precaution has been taken in the
preparation of this book, Check Point assumes no responsibility for errors or omissions. This
publication and features described herein are subject to change without notice.
RESTRICTED RIGHTS LEGEND:
Use, duplication, or disclosure by the government is subject to restrictions as set forth in
subparagraph (c)(1)(ii) of the Rights in Technical Data and Computer Software clause at DFARS
252.227-7013 and FAR 52.227-19.
TRADEMARKS:
Refer to the Copyright page http://www.checkpoint.com/copyright.html for a list of our
trademarks.
Refer to the Third Party copyright notices http://www.checkpoint.com/3rd_party_copyright.html
for a list of relevant copyrights and third-party licenses.
Important Information
Latest Software
We recommend that you install the most recent software release to stay up-to-date
with the latest functional improvements, stability fixes, security enhancements and
protection against new and evolving attacks.
Certifications
For third party independent certification of Check Point products, see the Check Point
Certifications page
https://www.checkpoint.com/products-solutions/certified-check-point-solutions/.
Feedback
Check Point is engaged in a continuous effort to improve its documentation.
Please help us by sending your comments
mailto:[email protected]?subject=Feedback on ClusterXL R80.20
Administration Guide.
Revision History
Date Description
26 September 2018 First release of this document
Contents
Important Information................................................................................................... 3
Terms ............................................................................................................................ 8
Introduction to ClusterXL ............................................................................................ 18
The Need for Clusters ............................................................................................. 18
ClusterXL Solution .................................................................................................. 18
How ClusterXL Works ............................................................................................. 19
The Cluster Control Protocol .........................................................................................19
ClusterXL Requirements and Compatibility ................................................................ 20
Check Point Appliances and Open Servers ............................................................. 20
Supported Number of Cluster Members ................................................................. 20
Hardware Requirements for Cluster Members ...................................................... 20
Software Requirements for Cluster Members ........................................................ 21
Hardware Requirements for Switches and Routers ............................................... 21
High Availability and Load Sharing Unicast Modes ........................................................21
Load Sharing Multicast Mode ........................................................................................22
VMAC Mode....................................................................................................................23
Supported Topologies for Synchronization Network .............................................. 24
Clock Synchronization in ClusterXL ........................................................................ 26
IPv6 Support for ClusterXL ..................................................................................... 26
IPv6 in ClusterXL High Availability .................................................................................26
Synchronized Cluster Restrictions .......................................................................... 27
Check Point Software Compatibility ........................................................................ 28
ClusterXL Compatibility with IPS ...................................................................................28
ClusterXL Compatibility (Excluding IPS) ........................................................................28
Example Configuration of a Cisco Catalyst Routing Switch ..................................... 29
High Availability and Load Sharing in ClusterXL ......................................................... 31
Introduction to High Availability and Load Sharing ................................................. 31
High Availability .............................................................................................................31
Load Sharing .................................................................................................................32
Example ClusterXL Topology .................................................................................. 33
Defining the Cluster Member IP Addresses ...................................................................34
Defining the Cluster Virtual IP Addresses .....................................................................34
Defining the Synchronization Network ..........................................................................34
Configuring Cluster Addresses on Different Subnets ....................................................34
Implementing Planning Considerations .................................................................. 36
High Availability or Load Sharing...................................................................................36
Choosing the Load Sharing Mode ..................................................................................36
IP Address Migration .....................................................................................................36
ClusterXL Modes ..................................................................................................... 37
High Availability Mode ...................................................................................................37
Load Sharing Multicast Mode ........................................................................................39
Load Sharing Unicast Mode ...........................................................................................40
Mode Comparison Table ................................................................................................41
Cluster Failover ...................................................................................................... 43
When Does a Failover Occur? ........................................................................................43
What Happens When a Cluster Member Recovers? .......................................................44
How a Recovered Cluster Member Obtains the Security Policy .....................................44
Synchronizing Connections in the Cluster .................................................................. 45
The Check Point State Synchronization Solution .................................................... 45
The Synchronization Network ........................................................................................45
How State Synchronization Works .................................................................................46
Configuring Services to Synchronize After a Delay ........................................................46
Configuring Services not to Synchronize .......................................................................47
Sticky Connections ........................................................................................................49
Non-Sticky Connections ................................................................................................52
Synchronizing Clusters on a Wide Area Network...........................................................53
Synchronized Cluster Restrictions ................................................................................54
Configuring ClusterXL................................................................................................. 55
Installing Cluster Members .................................................................................... 55
Configuring Routing for Client Computers .............................................................. 56
Selecting the CCP Transport Mode on the Cluster Members .................................. 57
Configuring the Cluster Object and Members ......................................................... 59
Overview ........................................................................................................................59
Using the Wizard Mode ..................................................................................................59
Using the Manual Configuration ....................................................................................62
Configuring a ClusterXL in Bridge Mode ................................................................. 67
Configuring ClusterXL in Bridge Mode - Active/Standby with Two Switches..................67
Configuring ClusterXL in Bridge Mode - Active/Active with Two Switches .....................80
Configuring a ClusterXL in Bridge Mode - Active/Active with Four Switches .................93
Advanced Features and Procedures ........................................................................... 96
Working with VPNs and Clusters ............................................................................ 96
Configuring VPN and Clusters .......................................................................................96
Defining VPN Peer Clusters with Separate Security Management Servers ...................97
Working with NAT and Clusters .............................................................................. 98
Cluster Fold and Cluster Hide .......................................................................................98
Configuring NAT on the Cluster .....................................................................................98
Configuring NAT on a Cluster Member ..........................................................................98
Working with VLANS and Clusters ........................................................................ 100
VLAN Support in ClusterXL..........................................................................................100
Connecting Several Clusters on the Same VLAN .........................................................101
Monitoring the Interface Link State ...................................................................... 106
Enabling Interface Link State Monitoring ....................................................................106
Bonding and Clusters ............................................................................................ 107
Bond Interfaces (Link Aggregation) .............................................................................108
Bonding - High Availability Mode .................................................................................110
Bonding - Load Sharing Mode......................................................................................116
Group of Bonds ............................................................................................................118
Performance Guidelines for Bonding ..........................................................................124
Troubleshooting Issues with Bonded Interfaces ..........................................................125
Advanced Cluster Configuration ........................................................................... 127
How to Configure Reboot Survival ...............................................................................127
Controlling the Clustering and Synchronization Timers .............................................. 127
Blocking New Connections Under Load .......................................................................128
Reducing the Number of Pending Packets...................................................................129
Defining Non-Monitored Interfaces ...................................................................... 130
Configuring Policy Update Timeout ....................................................................... 131
Enhanced 3-Way TCP Handshake Enforcement .................................................... 132
Cluster IP Addresses on Different Subnets .......................................................... 133
Introduction .................................................................................................................133
Configuring Cluster Addresses on Different Subnets ..................................................133
Example of Cluster IP Addresses on Different Subnets ...............................................135
Limitations of Cluster Addresses on Different Subnets ...............................................136
Converting a Security Gateway to a ClusterXL ...................................................... 138
Converting a Standalone Deployment to ClusterXL .....................................................138
Creating the New Member ...........................................................................................140
Creating the ClusterXL Object .....................................................................................140
In SmartConsole, for Computer 'B' ..............................................................................140
Preparing Computer 'A' ...............................................................................................141
In SmartConsole, for Computer 'A' ..............................................................................141
Adding Another Member to an Existing Cluster .................................................... 142
Removing a Member from an Existing Cluster ..................................................... 144
Configuring ISP Redundancy on a Cluster ............................................................ 145
Configuring the ISP Links ............................................................................................146
Configuring Security Gateway as DNS .........................................................................147
Configuring the Firewall ..............................................................................................148
Configuring with VPN ..................................................................................................149
Force ISP Link State ....................................................................................................149
Editing the ISP Redundancy Script...............................................................................150
Enabling Dynamic Routing Protocols in a Cluster Deployment ............................ 151
Components of the System ..........................................................................................151
Wait for Clustering ......................................................................................................152
ClusterXL Configuration Commands......................................................................... 153
Configuring the Cluster Member ID Mode in Local Logs....................................... 157
Registering a Critical Device ................................................................................. 158
Unregistering a Critical Device ............................................................................. 159
Reporting the State of a Critical Device ................................................................ 160
Registering Critical Devices Listed in a File .......................................................... 161
Unregistering All Critical Devices ......................................................................... 162
Configuring the Cluster Control Protocol (CCP) Mode .......................................... 163
Initiating Manual Cluster Failover......................................................................... 164
Monitoring and Troubleshooting Clusters ................................................................ 166
ClusterXL Monitoring Commands ......................................................................... 167
Monitoring Cluster State .............................................................................................171
Monitoring Critical Devices..........................................................................................176
Monitoring Cluster Interfaces......................................................................................182
Monitoring Bond Interfaces .........................................................................................187
Monitoring Cluster Failover Statistics .........................................................................191
Monitoring MAC Magic and MAC Forward Magic Values .............................................. 192
Monitoring Delta Synchronization................................................................................193
Viewing IGMP Status ....................................................................................................202
Viewing Cluster Delta Sync Statistics for Connections Table ....................................... 203
Viewing Cluster IP Addresses ......................................................................................204
Viewing the Cluster Member ID Mode in Local Logs ....................................................205
Viewing Interfaces Monitored by RouteD .....................................................................206
Viewing Roles of RouteD Daemon on Cluster Members...............................................207
Viewing Cluster Correction Statistics ..........................................................................208
Monitoring Cluster Status Using SmartConsole Clients ....................................... 210
SmartConsole Logs View .............................................................................................210
Working with SNMP Traps .................................................................................... 213
cphastart and cphastop ......................................................................................... 214
How to Initiate Cluster Failover ............................................................................ 215
Troubleshooting Dynamic Routing (routed) Pnote ................................................ 217
Standard RouteD Pnote Behavior ................................................................................217
Basic Troubleshooting Steps .......................................................................................217
ClusterXL Error Messages .................................................................................... 219
Appendix A - The clusterXL_admin Script ................................................................ 220
Appendix B - The clusterXL_monitor_ips Script ....................................................... 222
Appendix C - The clusterXL_monitor_process Script ............................................... 224
Member_A will assume the Standby
state).
Terms Active(!)
In ClusterXL, state of the Active cluster
3rd party Cluster member that suffers from a failure. A
Cluster of Check Point Security Gateways problem was detected, but the cluster
that work together in a redundant member still forwards packets, because it is
configuration. These Check Point Security the only member in the cluster, or because
Gateways are installed on X-Series XOS, or there are no other Active members in the
IPSO OS. VRRP Cluster on Gaia OS is also cluster. In any other situation, the state of
considered a 3rd party cluster. The 3rd party the member is Down.
cluster handles the traffic, and Check Point • ACTIVE(!) - See above.
Security Gateways perform only State
Synchronization. • ACTIVE(!F) - See above. Cluster
member is in the freeze state.
Active • ACTIVE(!P) - See above. This is the
State of a cluster member that is fully Pivot cluster member in Load Sharing
operational: Unicast mode.
• In ClusterXL, this applies to the state of • ACTIVE(!FP) - See above. This is the
the Security Gateway component Pivot cluster member in Load Sharing
• In 3rd party / OPSEC cluster, this applies Unicast mode and it is in the freeze state.
to the state of the cluster State Active/Active Mode
Synchronization mechanism
See Load Sharing Mode.
Active Member
Active/Standby Mode
A cluster member that handles network
connections that pass through the cluster. See High Availability Mode.
• In a High Availability deployment, only ARP Forwarding
one cluster member is Active and can
handle connections. See sk111956
http://supportcontent.checkpoint.com/soluti
• In a Load Sharing deployment, all cluster ons?id=sk111956.
members are Active and can handle
connections. Backup
Active Up • In VRRP on Gaia cluster:
ClusterXL in High Availability mode that was State of a cluster member that is ready to
configured as Maintain current active be promoted to Master state (if Master
Cluster Member in the cluster object in member fails).
SmartConsole: • In VSX cluster configured in Virtual
• If the current Active member fails for System Load Sharing mode with three or
some reason, or is rebooted (for example, more cluster members:
Member_A), then failover occurs between State of a Virtual System on a third (and
cluster members - another Standby so on) VSX cluster member.
member will be promoted to be Active
A cluster member or Virtual System in this
(for example, Member_B).
state does not process any traffic passing
• When former Active member (Member_A) through cluster.
recovers from a failure, or boots, the
former Standby member (Member_B) will
remain to be in Active state (and
Blocking Mode The CCL provides connections stickiness by
Cluster operation mode, in which cluster "correcting" the packets to the correct
member does not forward any traffic (for cluster member:
example, caused by a failure). • In most cases, the CCL makes the
correction from the CoreXL SND.
Bond
• In some cases (like Dynamic Routing, or
A virtual interface that contains (enslaves) VPN), the CCL makes the correction from
two or more physical interfaces for the Firewall or SecureXL.
redundancy and load sharing. The physical
interfaces share one IP address and one MAC Cluster Interface
address. See Link Aggregation. An interface on a cluster member, whose
Network Type was set as Cluster in
Bonding SmartConsole in cluster object. This
See Link Aggregation. interface is monitored by cluster, and failure
on this interface will cause cluster failover.
Bridge Mode
Cluster Member
A Security Gateway or Virtual System that
works as a Layer-2 bridge device for easy A Security Gateway that is part of a cluster.
deployment in an existing topology.
Cluster Mode
Cluster Configuration of cluster members to work in
1. Two or more Security Gateways that work these redundant modes:
together in a redundant configuration - High • One cluster member processes all the
Availability or Load Sharing. traffic - High Availability or VRRP mode
2. In a virtualized environment - a set of • All traffic is processed in parallel by all
VMware ESX/i hosts used for High Availability cluster members - Load Sharing
or Load Sharing.
Cluster Topology
Cluster Control Protocol (CCP) Set of interfaces on all members of a cluster
Proprietary Check Point protocol that runs and their settings (Network Objective, IP
between cluster members on UDP port 8116, address/Net Mask, Topology, Anti-Spoofing,
and has the following roles: and so on).
• State Synchronization (Delta Sync) ClusterXL
• Health checks (state of cluster members Cluster of Check Point Security Gateways
and of cluster interfaces): that work together in a redundant
• Health-status Reports configuration. The ClusterXL both handles
• Cluster-member Probing the traffic and performs State
• State-change Commands Synchronization.
• Querying for Cluster Membership These Check Point Security Gateways are
installed on Gaia OS:
Note: CCP is located between the Check
Point Firewall kernel and the network • Up to 5 cluster members are supported in
interface (therefore, only TCPdump should ClusterXL.
be used for capturing this traffic). • Up to 2 cluster members are supported in
VRRP cluster.
Cluster Correction Layer (CCL)
• Up to 13 cluster members are supported
Proprietary Check Point mechanism that
in VSX VSLS cluster.
deals with asymmetric connections in Check
Point cluster. Note - In ClusterXL Load Sharing mode,
configuring more than 4 members
significantly decreases the cluster
Delta Sync
performance due to amount of Delta Sync.
Synchronization of kernel tables between all
CPHA working cluster members - exchange of CCP
packets that carry pieces of information
General term that stands for Check Point
about different connections and operations
High Availability (historic fact: the first
that should be performed on these
release of ClusterXL supported only High
connections in relevant kernel tables. This
Availability) that is used only for internal
Delta Sync process is performed directly by
references (for example, inside kernel
Check Point kernel. While performing Full
debug) to designate ClusterXL infrastructure.
Sync, the Delta Sync updates are not
Creating an Interface Bond in Load processed and saved in kernel memory. After
Sharing Mode Full Sync is complete, the Delta Sync packets
stored during the Full Sync phase are applied
Follow the instructions in the R80.20 Gaia
by order of arrival.
Administration Guide
https://sc1.checkpoint.com/documents/R80. Delta Sync Retransmission
20_GA/WebAdminGuides/EN/CP_R80.20_Gai
a_AdminGuide/html_frameset.htm - Chapter It is possible that Delta Sync packets will be
Network Management - Section Network lost or corrupted during the Delta Sync
Interfaces - Section Bond Interfaces (Link operations. In such cases, it is required to
Aggregation). make sure the Delta Sync packet is re-sent.
The cluster member requests the sending
Critical Device cluster member to retransmit the
lost/corrupted Delta Sync packet.
Also known as a Problem Notification, or
Each Delta Sync packet has a sequence
pnote. A special software device on each number.
cluster member, through which the critical
The sending member has a queue of sent
aspects for cluster operation are monitored.
Delta Sync packets.
When the critical monitored component on a
Each cluster member has a queue of packets
cluster member fails to report its state on
sent from each of the peer cluster members.
time, or when its state is reported as
If, for any reason, a Delta Sync packet was
problematic, the state of that member is
not received by a cluster member, it can ask
immediately changed to Down. The complete for a retransmission of this packet from the
list of the configured critical devices (pnotes) sending member.
is printed by the cphaprob -ia list The Delta Sync retransmission mechanism is
command or show cluster pnotes all somewhat similar to a TCP Window and TCP
command. retransmission mechanism.
When a member requests retransmission of
Dead
Delta Sync packet, which no longer exists on
State reported by a cluster member when it the sending member, the member prints a
goes out of the cluster (due to cphastop console messages that the sync is not
command (which is a part of cpstop), or complete.
reboot).
Down
Decision Function State of a cluster member during a failure
A special cluster algorithm applied by each when one of the Critical Devices reports its
cluster member on the incoming traffic in state as "problem":
order to decide, which cluster member • In ClusterXL, applies to the state of the
should process the received packet. Each Security Gateway component
cluster members maintains a table of hash
values generated based on connections tuple • In 3rd party / OPSEC cluster, applies to
(source and destination IP addresses/Ports, the state of the State Synchronization
and Protocol number). mechanism
A cluster member in this state does not Flush and ACK
process any traffic passing through cluster. Also, FnA, F&A. Cluster member forces the
Delta Sync packet about the incoming packet
Dying
and waiting for acknowledgments from all
State of a cluster member as assumed by other Active members and only then allows
peer members, if it did not report its state for the incoming packet to pass through.
0.7 sec.
In some scenarios, it is required that some
Failback information, written into the kernel tables,
will be Sync-ed promptly, or else a race
Also, Fallback. Recovery of a cluster member condition can occur. The race condition may
that suffered from a failure. The state of a occur if a packet that caused a certain
recovered cluster member is changed from change in kernel tables left cluster
Down to either Active, or Standby Member_A toward its destination and then
(depending on Cluster Mode). the return packet tries to go through cluster
Member_B.
Failed Member
In general, this kind of situation is called
A cluster member that cannot send or accept asymmetric routing. What may happen in this
traffic because of a hardware or software scenario is that the return packet arrives at
problem. cluster Member_B before the changes
induced by this packet were Sync-ed to this
Failover
Member_B.
Also, Fail-over. Transferring of a control over
Example of such a case is when a SYN packet
traffic (packet filtering) from a cluster
goes through cluster Member_A, causing
member that suffered a failure to another
multiple changes in the kernel tables and
cluster member (based on internal cluster
then leaves to a server. The SYN-ACK packet
algorithms).
from a server arrives at cluster Member_B,
Failure but the connection itself was not Synced yet.
In this condition, the cluster Member_B will
A hardware or software problem that causes drop the packet as an Out-of-State packet
a Security Gateway to be unable to serve as a (First packet isn't SYN). In order to
cluster member (for example, one of cluster prevent such conditions, it is possible
interface has failed, or one of the monitored tousethe"FlushandAck"(F&A)mechanism.
daemon has crashed). Cluster member that
suffered from a failure is declared as failed, This mechanism can send the Delta Sync
and its state is changed to Down (a physical packets with all the changes accumulated so
interface is considered Down only if all far in the Sync buffer to the other cluster
configured VLANs on that physical interface members, hold the original packet that
are Down). induced these changes and wait for
acknowledgment from all other (Active)
Flapping cluster members that they received the
information in the Delta Sync packet. When
Consequent changes in the state of either all acknowledgments arrived, the
cluster interfaces (cluster interface flapping), mechanism will release the held original
or cluster members (cluster member packet.
flapping). Such consequent changes in the
state are seen in the Logs & Monitor > This ensures that by the time the return
Logs (if in SmartConsole >cluster object, the packet arrived from a server at the cluster,
cluster administrator set the Track all the cluster members are aware of the
changes in the status of cluster connection.
members to Log). F&A is being operated at the end of the
Inbound chain and at the end of the Outbound
chain (it is more common at the Outbound).
Forwarding Packets that are sent on the Forwarding
Process of transferring of an incoming traffic Layer use a special source MAC address to
from one cluster member to another cluster inform the receiving cluster member that
member for processing. There are two types they have already been inspected by another
of forwarding the incoming traffic between cluster member. Thus, the receiving cluster
cluster members - Packet forwarding and member can safely hand over these packets
Chain forwarding. Also see Forwarding Layer to the local Operating System, without
and ARP Forwarding. further inspection.
The Forwarding Layer is a ClusterXL Also, Full HA Mode. A special Cluster Mode
mechanism that allows a cluster member to (supported only on Check Point appliances
pass packets to peer cluster members, after running Gaia OS or SecurePlatform OS,
they have been locally inspected by the where each cluster member also runs as a
firewall. This feature allows connections to Security Management Server. This provides
be opened from a cluster member to an redundancy both between Security Gateways
external host. (only High Availability is supported) and
between Security Management Servers (only
Packets originated by cluster members are High Availability is supported). See sk101539
hidden behind the Cluster Virtual IP address. http://supportcontent.checkpoint.com/soluti
Thus, a reply from an external host is sent to ons?id=sk101539 and sk39345
the cluster, and not directly to the source http://supportcontent.checkpoint.com/soluti
cluster member. This can pose problems in ons?id=sk39345.
the following situations:
• The cluster is working in High Availability Full Sync
mode, and the connection is opened from Process of full synchronization of applicable
the Standby cluster member. All packets kernel tables by a cluster member from the
from the external host are handled by the working cluster member(s) when it tries to
Active cluster member, instead. join the existing cluster. This process is
• The cluster is working in a Load Sharing meant to fetcha"snapshot" of the applicable
mode, and the decision function has kernel tables of already Active cluster
selected another cluster member to member(s).
handle this connection. This can happen Full Sync is performed during the
since packets directed at a Cluster IP initialization of Check Point software (during
address are distributed between cluster boot process, the first time the cluster
members as with any other connection. member runs policy installation, during
If a cluster member decides, upon the cpstart, during cphastart). Until the Full
completion of the firewall inspection process, Sync process completes successfully, this
that a packet is intended for another cluster cluster member remains in the Down state,
member, it can use the Forwarding Layer to because until it is fully synchronized with
hand the packet over to that cluster member. other cluster members, it can not function as
a cluster member.
In High Availability mode, packets are
forwarded over a Synchronization network Meanwhile, the Delta Sync packets continue
directly to peer cluster members. It is to arrive, and the cluster member that tries
important to use secured networks only, as to join the existing cluster, stores them in the
encrypted packets are decrypted during the kernel memory until the Full Sync
inspection process, and are forwarded as completes.
clear-text (unencrypted) data. The whole Full Sync process is performed by
In Load Sharing mode, packets are fwd daemons on TCP port 256 over the Sync
forwarded over a regular traffic network. network (if it fails over the Sync network, it
tries the other cluster interfaces). The
information is sent by fwd daemons in other Cluster Members still run their
chunks, while making sure they confirm Software Blades in the kernel space
getting the information before sending the (represented by the fw_worker processes). In
next chunk. the Hybrid Mode, Cluster Members are able
to synchronize the required information.
Also see Delta Sync.
Init
HA not started
State of a cluster member in the phase after
Output of the cphaprob <flag> command
the boot and until the Full Sync completes. A
or show cluster <option> command on
cluster member in this state does not
the cluster member. This output means that
process any traffic passing through cluster.
Check Point clustering software is not
started on this Security Gateway (for IP Tracking
example, this machine is not a part of a
cluster, or cphastop command was run, or Collecting and saving of Source IP addresses
some failure occurred that prevented the and Source MAC addresses from incoming IP
ClusterXL product from starting correctly). packets during the probing. IP tracking is a
useful for members within a cluster to
High Availability Mode determine whether the network connectivity
of the member is acceptable.
A redundant cluster mode, where only one
cluster member (Active member) processes IP Tracking Policy
all the traffic, while other cluster members
(Standby members) are ready to be promoted Setting that controls, which IP addresses
to Active state if the current Active member should be tracked during IP tracking:
fails. • Only IP addresses from the subnet of
In the High Availability mode, the Cluster cluster VIP, or from subnet of physical
Virtual IP address (that represents the cluster interface (this is the default)
cluster on that network) is associated: • All IP addresses, also outside the cluster
• With physical MAC Address of Active subnet
member Link Aggregation
• With virtual MAC Address (see sk50840 A technology that joins multiple physical
http://supportcontent.checkpoint.com/sol interfaces together into one virtual interface,
utions?id=sk50840) known as a bond interface. Also known as
HTU Interface Bonding.
Stands for "HA Time Unit". All internal time Load Sharing Mode
in ClusterXL is measured in HTUs (the times
Also, Load Balancing mode. A redundant
in cluster debug also appear in HTUs).
cluster mode, where all cluster members
Formula in the Check Point software: 1 HTU
process all incoming traffic in parallel. See
= 10 x fwha_timer_base_res = 10 x 10
milliseconds = 100 ms Load Sharing Multicast Mode and Load
Sharing Unicast Mode.
Hybrid Mode
Load Sharing Multicast Mode
Starting in R80.20, on Security Gateways with
40 or more CPU cores, Software Blades run Load Sharing Cluster Mode, where all cluster
in the user space (as fwk processes). The members process all traffic in parallel. Each
Hybrid Mode refers to the state when you cluster member is assigned the equal load of
upgrade Cluster Members from R80.10 (or [ 100% / number_of_members ].
below) to R80.20 (or above). The Hybrid Mode The Cluster Virtual IP address (that
is the state, in which the upgraded Cluster represents the cluster on that network) is
Members already run their Software Blades associated with Multicast MAC Address
in the user space (as fwk processes), while 01:00:5E:X:Y:Z (which is generated based on
last 3 bytes of cluster Virtual IP address on interface state appears in the output of the
that network). cphaprob -a if command.
A ClusterXL decision algorithm (Decision
Function) on all cluster members decides, Non-Sticky Connection
which cluster member should process the A connection is called non-sticky, if the reply
given packet. packet returns via a different cluster
member, than the original packet (for
Load Sharing Unicast Mode
example, if network administrator has
Load Sharing Cluster Mode, where one configured asymmetric routing). In Load
cluster member (called Pivot) accepts all Sharing mode, all cluster members are
traffic. Then, the Pivot member decides to Active, and in Static NAT and encrypted
process this traffic, or to forward it to other connections, the Source and Destination IP
non-Pivot cluster members. addresses change. Therefore, Static NAT and
The traffic load is assigned to cluster encrypted connections through a Load
members based on the hard-coded formula Sharing cluster may be non-sticky.
per the value of Pivot_overhead attribute
(see sk34668 Packet Selection
http://supportcontent.checkpoint.com/soluti Distinguishing between different kinds of
ons?id=sk34668). packets coming from the network, and
The Cluster Virtual IP address (that selecting, which member should handle a
represents the cluster on that network) is specific packet (Decision Function
associated with: mechanism):
• Physical MAC Address of Pivot member • CCP packet from another member of this
• Virtual MAC Address (see sk50840 cluster
http://supportcontent.checkpoint.com/sol • CCP packet from another cluster or from
utions?id=sk50840) a cluster member with another version
Management Server (usually older version of CCP)
Traffic
The flow of data between network devices.
VLAN
Virtual Local Area Network. Open servers or
appliances connected to a virtual network,
which are not physically connected to the
same network.
VLAN Trunk
A connection between two switches that
contains multiple VLANs.
VMAC
Virtual MAC address. When this feature is
enabled on cluster members, all cluster
members in High Availability mode and Load
Sharing Unicast mode associate the same
Virtual MAC address with Virtual IP address.
This allows avoiding issues when Gratuitous
ARP packets sent by cluster during failover
are not integrated into ARP cache table on
switches surrounding the cluster. See
sk50840
http://supportcontent.checkpoint.com/soluti
ons?id=sk50840.
CHAPTE R 1
Introduction to ClusterXL
In This Section:
The Need for Clusters ..................................................................................................18
ClusterXL Solution ........................................................................................................18
How ClusterXL Works ..................................................................................................19
ClusterXL Solution
ClusterXL is a Check Point software-based cluster solution for Security Gateway redundancy and
Load Sharing. A ClusterXL Security Cluster contains identical Check Point Security Gateways.
• A High Availability Security Cluster ensures Security Gateway and VPN connection redundancy
by providing transparent failover to a backup Security Gateway in the event of failure.
• A Load Sharing Security Cluster provides reliability and also increases performance, as all
members are active
Item Description
1 Internal network
5 Internet
Note - There is no requirement for throughput of Sync interface to be identical to, or larger than
throughput of traffic interfaces (although, to prevent a possible bottle neck, a good practice for
throughput of Sync interface is to be at least identical to throughput of traffic interfaces).
When using CCP in multicast mode, configure the following settings on the router:
By default, when ClusterXL is configured in High Availability mode or Load Sharing Unicast Mode,
the unicast Cluster Virtual IP addresses are mapped to unicast MAC addresses of the physical
interfaces on the Active or Pivot cluster member. To let the traffic reach the cluster, configure the
settings below on the surrounding routers.
When working with Load Sharing in Multicast mode, configure the following settings on the
router:
By default, when ClusterXL is configured in High Availability mode or Load Sharing Multicast
Mode, the unicast Cluster Virtual IP addresses are mapped to multicast MAC addresses (these are
generated automatically by the Management Server based on the configured Cluster Virtual IP
addresses). To let the traffic reach the cluster, the router must support sending unicast IP packets
with multicast MAC addresses.
IGMP and Some routers require disabling of IGMP snooping, or configuration of static
Static CAMs CAMs to support sending of unicast IP packets with multicast MAC addresses.
Disabling Some routers send multicast traffic to the router itself. This may cause a
forwarding of multicast storm through the network. Disable such configuration on your
multicast router.
traffic to the
router itself
VMAC Mode
When ClusterXL is configured in HA mode or Load Sharing Unicast mode (not Multicast) a single
cluster member is associated with the Cluster Virtual IP address. In a High Availability
environment, the single member is the Active member. In a Load Sharing environment, the single
member is the Pivot.
After fail-over, the new Active member (or Pivot member) broadcasts a series of Gratuitous ARP
Requests (G-ARPs). The G-ARPS associate the Virtual IP address of the cluster with the physical
MAC address of the new Active member or the new Pivot. When this happens:
• A member with a large number of Static NAT entries can transmit too many GARPs
Switches may not integrate these GARP updates quickly enough into their ARP tables.
Switches continue to send traffic to the physical MAC address of the member that failed. This
results in traffic outage until the switches have fully updated ARP cache tables.
• Network components, such as VoIP phones, ignore GARPs
These components continue to send traffic to the MAC address of the failed member.
To minimize possible traffic outage during a fail-over, configure the cluster to use a virtual MAC
address (VMAC).
By enabling Virtual MAC in ClusterXL High Availability mode, or Load Sharing Unicast mode, all
cluster members associate the same Virtual MAC address with all Cluster Virtual Interfaces and
the Virtual IP address. In Virtual MAC mode, the VMAC that is advertised by the cluster members
(through G-ARP Requests) keeps the real MAC address of each member and adds a Virtual MAC
address on top of it.
For local connections and sync connections, the real physical MAC address of each cluster
member is still associated with its real IP address.
Note - Traffic originating from the cluster members will be sent with the VMAC Source address.
You can enable VMAC in SmartConsole, or on the command line. See sk50840
http://supportcontent.checkpoint.com/solutions?id=sk50840.
Failover time in a cluster with enabled VMAC mode is shorter than a failover in a cluster that uses
a physical MAC addresses.
3. Click OK.
4. Install the Access Control Policy on this cluster object.
Topology 2
• Sync interface is a Bond of several physical slave interfaces.
To work with this topology, you can configure the Bond interface in High Availability or Load
Sharing mode.
• On each cluster member, physical slave interfaces of the same Bond connect to different
switches.
• The switches must connect to each other through a cable.
Limitations
• IPv6 is not supported for Load Sharing clusters.
• You cannot define IPv6 address for synchronization interfaces.
FTP, HTTP and SMTP Security Servers Yes (2, 5) Yes (2)
Notes:
1. If there is a cluster failover when fragments are being received, the packet will be lost.
2. Does not survive cluster failover.
3. Requires unidirectional stickiness. This means that the same cluster member must receive all
external packets, and the same cluster member must receive all internal packets, but the
same cluster member does not have to receive both internal and external packets.
4. Requires bidirectional connection stickiness.
5. Uses the cluster Forwarding Layer.
Notes:
1. If there is a cluster failover when fragments are being received, the packet will be lost.
2. Does not survive cluster failover.
3. Requires unidirectional stickiness. This means that the same cluster member must receive all
external packets, and the same cluster member must receive all internal packets, but the
same cluster member does not have to receive both internal and external packets.
4. Requires bidirectional connection stickiness.
5. Uses the cluster Forwarding Layer.
2. For a network x.y.z.0 that does not have a Cluster Virtual IP address, such as the Sync, you
use the same procedure and substitute fa instead of 0 for the last octet of the MAC address.
• For example: 01:00:5e:00:00:fa for the 10.0.0.X network
To add a permanent CAM entry for the multicast MAC address for module 1 - port 1, and module 2
- ports 1, 3, and 8 through 12:
Cisco> (enable) set cam permanent 01-40-5e-28-0a-64 1/1,2/1,2/3,2/8-12
Permanent multicast entry added to CAM table.
Cisco> (enable)
High Availability
In a High Availability cluster, only one member is active (Active/Standby operation). In the event
that the active cluster member becomes unavailable, all connections are re-directed to a
designated standby without interruption. In a synchronized cluster, the standby cluster members
are updated with the state of the connections of the active cluster member.
In a High Availability cluster, each member is assigned a priority. The highest priority member
serves as the Security Gateway in normal circumstances. If this member fails, control is passed to
the next highest priority member. If that member fails, control is passed to the next member, and
so on.
Upon Security Gateway recovery, you can maintain the current active Security Gateway (Active Up),
or to change to the highest priority Security Gateway (Primary Up).
ClusterXL High Availability mode supports both IPv4 and IPv6.
Load Sharing
ClusterXL Load Sharing distributes traffic within a cluster so that the total throughput of multiple
members is increased. In Load Sharing configurations, all functioning members in the cluster are
active, and handle network traffic (Active/Active operation).
If any member in a cluster becomes unreachable, transparent failover occurs to the remaining
operational members in the cluster, thus providing High Availability. All connections are shared
between the remaining Security Gateways without interruption.
ClusterXL Load Sharing modes do not support IPv6.
Item Description
1 Internal network
6 Internet
Each cluster member has three interfaces: one external interface, one internal interface, and one
for synchronization. Cluster member interfaces facing in each direction are connected via a hub or
switch.
All cluster member interfaces facing the same direction must be in the same network. For
example, there must not be a router between cluster members.
The Security Management Server can be located anywhere, and connection should be established
to either the internal or external cluster IP addresses.
These sections present ClusterXL configuration concepts shown in the example.
Note - In these examples, RFC 1918 private addresses in the range 192.168.0.0 to
192.168.255.255 are treated as public IP addresses.
• Allow organizations to use only one public IP address for the ClusterXL Cluster. This saves
public IP addresses.
IP Address Migration
If you wish to provide High Availability or Load Sharing to an existing Security Gateways
configuration, we recommend taking the existing IP addresses from the Active Security Gateway,
and make these the Cluster Virtual IP addresses, when feasible. Doing so will avoid altering
current IPsec endpoint identities, as well keep Hide NAT configurations the same in many cases.
ClusterXL Modes
ClusterXL has four working modes. This section briefly describes each mode and its relative
advantages and disadvantages.
• High Availability Mode
• Load Sharing Multicast Mode
• Load Sharing Unicast Mode
Note - Many examples in the section refer to the sample deployment shown in the ClusterXL
example ("Example ClusterXL Topology" on page 33).
Example
This scenario describes a connection from a Client computer on the external network to a Web
server behind the Cluster (on the internal network). The cluser of two members is configured in
High Availability mode.
Example topology:
[Client on]
[external network]
{IP 192.168.10.78/24}
{DG 192.168.10.100}
|
|
{VIP 192.168.10.100/24}
/ \
| |
{IP 192.168.10.1/24} {IP 192.168.10.2/24}
| |
{Active} {Standby}
[Member1]-----sync-----[Member2]
| |
{IP 10.10.0.1/24} {IP 10.10.0.2/24}
| |
\ /
{VIP 10.10.0.100/24}
|
|
{DG 10.10.0.100}
{IP 10.10.0.34/24}
[Web server on]
[internal network]
Chain of events:
1. The user requests tries to connect from his Client computer 192.168.10.78 to the Web server
10.10.0.34.
2. The Default Gateway on Client computer is 192.168.10.100 (the cluster Virtual IP address).
3. The Client computer issues an ARP Request for IP 192.168.10.100.
4. The Active cluster member (Member1) handles the ARP Request for IP 192.168.10.100.
5. The Active cluster member sends the ARP Reply with the MAC address of the external
interface, on which the IP address 192.168.10.1 is configured.
6. The Client computer sends the HTTP request packet to the Active cluster member - to the VIP
address 192.168.10.100 and MAC address of the corresponding external interface.
7. The Active cluster member handles the HTTP request packet.
8. The Active cluster member sends the HTTP request packet to the Web server 10.10.0.34.
9. The Web server handles the HTTP request packet.
10. The Web server generates the HTTP response packet.
11. The Default Gateway on Web server computer is 10.10.0.100 (the cluster Virtual IP address).
12. The Web server issues an ARP Request for IP 10.10.0.100.
13. The Active cluster member handles the ARP Request for IP 10.10.0.100.
14. The Active cluster member sends the ARP Reply with the MAC address of the internal
interface, on which the IP address 10.10.0.1 is configured.
15. The Web server sends the HTTP response packet to the Active cluster member - to VIP
address 10.10.0.100 and MAC address of the corresponding internal interface.
ClusterXL Administration Guide R80.20 | 38
High Availability and Load Sharing in ClusterXL
16. The Active cluster member handles the HTTP response packet.
17. The Active cluster member sends the HTTP response packet to the Client computer
192.168.10.78.
18. From now, all traffic between the Client computer and the Web server is routed through the
Active cluster member (Member1).
19. If a failure occurs on the current Active cluster member (Member1), the cluster fails over.
20. The Standby cluster member (Member2) assumes the role of the Active cluster member.
21. The Standby cluster member sends Gratuitous ARP Requests to both the 192.168.10.x and the
10.10.0.x networks.
These GARP Requests associate the Cluster Virtual IP addresses with the MAC addresses of
the physical interfaces on the new Active cluster member (former Standby cluster member):
• Cluster VIP address 192.168.10.100 and MAC address of the corresponding external
interface, on which the IP address 192.168.10.2 is configured.
• Cluster VIP address 10.10.0.100 and MAC address of the corresponding internal interface,
on which the IP address 10.10.0.2 is configured.
22. From now, all traffic between the Client computer and the Web server is routed through the
new Active cluster member (Member2).
23. The former Active member (Member1) is now considered to be "down". Upon the recovery of a
former Active cluster member, the role of the Active cluster member may or may not be
switched back to that cluster member, depending on the cluster object configuration.
traffic is not blocked), and that no two cluster members will handle the same packets (so that
traffic is not duplicated).
An additional requirement of the decision function, is to route each connection through a cluster
member, to ensure that packets that belong to a single connection will be processed by the same
cluster member. Unfortunately, this requirement cannot always be enforced, and in some cases,
packets of the same connection will be handled by different cluster members. ClusterXL handles
these situations using its State Synchronization mechanism, which mirrors connections on all
cluster members.
Example
This scenario describes a user logging from the Internet to a Web server behind the cluster that is
configured in Load Sharing Multicast mode.
1. The user requests a connection from 192.168.10.78 (his computer) to 10.10.0.34 (the Web
server).
2. A router on the 192.168.10.x network recognizes 192.168.10.100 (the cluster Virtual IP
address) as the default gateway to the 10.10.0.x network.
3. The router issues an ARP Request for IP 192.168.10.100.
4. One of the Active cluster members processes the ARP Request, and responds with the
Multicast MAC assigned to the cluster Virtual IP address of 192.168.10.100.
5. When the Web server responds to the user requests, it recognizes 10.10.0.100 as its default
gateway to the Internet.
6. The Web server issues an ARP Request for IP 10.10.0.100.
7. One of the Active members processes the ARP Request, and responds with the Multicast MAC
address assigned to the cluster Virtual IP address of 10.10.0.100.
8. All packets sent between the user and the Web server reach every cluster member, which
decide whether to process each packet.
9. When a cluster failover occurs, one of the cluster members goes down. However, traffic still
reaches all of the Active cluster members. Therefore, there is no need to make changes in the
network ARP routing. The only thing that changes, is the cluster decision function, which takes
into account the new state of the cluster members.
When a failure occurs in a non-Pivot member, its handled connections are redistributed between
active non-Pivot cluster members, providing the same High Availability capabilities of High
Availability and Load Sharing Multicast. When the Pivot cluster member encounters a problem, a
regular failover event occurs, and, in addition, another active non-Pivot member assumes the role
of the new Pivot. The Pivot member is always the active member with the highest priority. This
means that when a former Pivot recuperates, it will retain its previous role.
Example
In this scenario, we use a Load Sharing Unicast cluster as the Security Gateway between the end
user computer and the Web server.
1. The user requests a connection from 192.168.10.78 (his computer) to 10.10.0.34 (the Web
server).
2. A router on the 192.168.10.x network recognizes 192.168.10.100 (the cluster Virtual IP
address) as the default gateway to the 10.10.0.x network.
3. The router issues an ARP Request for IP 192.168.10.100.
4. The Pivot cluster member handles the ARP Request, and responds with the MAC address that
corresponds to its own unique IP address of 192.168.10.1.
5. When the Web server responds to the user requests, it recognizes 10.10.0.100 as its default
gateway to the Internet.
6. The Web server issues an ARP Request for IP 10.10.0.100.
7. The Pivot cluster member handles the ARP Request, and responds with the MAC address that
corresponds to its own unique IP address of 10.10.0.1.
8. The user request packet reaches the Pivot cluster member on interface 192.168.10.1.
9. The Pivot cluster member decides that the second non-Pivot cluster member should handle
this packet, and forwards it to 192.168.10.2.
10. The second cluster member recognizes the packet as a forwarded packet, and handles it.
11. Further packets are processed by either the Pivot cluster member, or forwarded and
processed by the non-Pivot cluster member.
12. When a failover occurs from the current Pivot cluster member, the second cluster member
assumes the role of Pivot.
13. The new Pivot member sends Gratuitous ARP Requests to both the 192.168.10.x and the
10.10.0.x networks. These G-ARP requests associate the cluster Virtual IP address of
192.168.10.100 with the MAC address that corresponds to the unique IP address of
192.168.10.2, and the cluster Virtual IP address of 10.10.0.100 with the MAC address that
correspond to the unique IP address of 10.10.0.2.
14. Traffic sent to the cluster is now received by the new Pivot cluster member, and processed by
it (as it is currently the only active cluster member).
15. When the former Pivot cluster member recovers, it re-assumes the role of Pivot, by
associating the cluster Virtual IP addresses with its own unique MAC addresses.
Hardware Support All routers Not all routers are All routers
supported
Cluster Failover
Failover is a cluster redundancy operation that automatically occurs if a cluster member is not
functional. When this occurs, other cluster members take over for the failed cluster member.
In a High Availability mode:
• If the Active cluster member detects that it can not function as a cluster member, it notifies the
peer Standby cluster members that it must go down. One of the Standby cluster members
(with the next highest priority) will promote itself to the Active state.
• If one of the Standby cluster members stops receiving Cluster Control Protocol (CCP) packets
from the current Active cluster member, that Standby cluster member can assume that the
current Active cluster member failed. As a result, one of the Standby cluster members (with
the next highest priority) will promote itself to the Active state.
• If you do not use State Synchronization in the cluster, existing connections are interrupted
when cluster failover occurs.
• If a cluster member detects that it can not function as a cluster member, it notifies the peer
cluster members that it must go down. Traffic load will be redistributed between the working
cluster members.
• If the cluster members stop receiving Cluster Control Protocol (CCP) packets from one of their
peer cluster member, those working cluster members can assume that their peer cluster
member failed. As a result, traffic load will be redistributed between the working cluster
members.
• Because by design, all cluster members are always synchronized, current connections are not
interrupted when cluster failover occurs.
To tell each cluster member that the other cluster members are alive and functioning, the
ClusterXL Cluster Control Protocol (CCP) maintains a heartbeat between cluster members. If
after a predefined time, no CCP packets are received from a cluster member, it is assumed that
the cluster member is down. As a result, cluster failover can occur.
Note that more than one cluster member may encounter a problem that will result in a cluster
failover event. In cases where all cluster members encounter such problems, ClusterXL will try to
choose a single cluster member to continue operating. The state of the chosen member will be
reported as Active Attention. This situation lasts until another cluster member fully recovers. For
example, if a cross cable connecting the sync interfaces on cluster members malfunctions, both
cluster members will detect an interface problem. One of them will change to the Down state, and
the other to Active Attention state.
Procedure:
1. In SmartConsole, click Objects > Object Explorer.
2. In the left tree, click the small arrow on the left of the Services to expand this category
3. In the left tree, select TCP.
4. Search for the applicable TCP service.
5. Double-click the applicable TCP service.
6. In the TCP service properties window, click Advanced page.
7. At the top, select Override default settings (on Domain Management Server: Override global
domain settings).
8. At the bottom, in the Cluster and synchronization section, select Start synchronizing and
enter the desired value.
Important - This change applies to all policies that use this service.
9. Click OK.
10. Close the Object Explorer.
11. Publish the session.
12. Install the Access Control Policy on the cluster object.
Note - The Delayed Notifications setting in the service object is ignored, if Connection Templates
are not offloaded by the Firewall to SecureXL. For additional information about the Connection
Templates, see the R80.20 Performance Tuning Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_Performanc
eTuning_AdminGuide/html_frameset.htm.
7. On the General page, configure the same settings as in the existing synchronized service.
8. On the Advanced page:
a) Configure the same settings as in the existing synchronized service.
b) In the Cluster and synchronization section, clear Synchronize connections if State
Synchronization is enabled on the cluster.
Important - This change applies to all policies that use this service.
9. Click OK.
10. Close the Object Explorer.
11. Use the synchronized service and the non-synchronized counterpart service in the applicable
rules in the applicable Access Control Policies.
12. Publish the session.
13. Install the Access Control Policy on the cluster object.
Sticky Connections
Introduction to Sticky Connections
A connection is considered sticky, when all of its packets are handled, in either direction, by a
single cluster member.
This is the case in High Availability mode, where all connections are routed through the same
cluster member, and hence, are sticky.
This is also the case in Load Sharing mode, when there are no VPN peers, Static NAT rules, or SIP
traffic.
In Load Sharing mode, there are cases, where it is necessary to ensure that a connection that
starts on a specific cluster member will continue to be processed by the same cluster member in
both directions. Certain connections can be made sticky by enabling the Sticky Decision Function
in the cluster object in SmartConsole.
The Check Point Cluster Correction Layer (CCL) deals with asymmetric connections in Check
Point cluster.
Item Description
1 Internal network
5 Internet
In this scenario:
• A third-party peer (gateway or client) attempts to create a VPN tunnel.
• Cluster Members "A" and "B" belong to a ClusterXL in Load Sharing mode.
The third-party peers, lacking the ability to store more than one set of SAs, cannot negotiate a
VPN tunnel with multiple cluster members, and therefore the cluster member cannot complete
the routing transaction.
This issue is resolved for certain third-party peers or gateways that can save only one set of SAs
by making the connection sticky. The Cluster Correction Layer (CCL) makes sure that a single
cluster member processes all VPN sessions, initiated by the same third-party gateway.
Item Description
1 Security Gateway - Cluster member A
4 Internet
5 Gateway - Spoke A
6 Gateway - Spoke B
Spokes A and B must be set to always communicate using the same cluster member. The Cluster
Correction Layer (CCL) solves half of this problem, in that a single cluster member processes all
VPN sessions initiated by either third-party gateway.
To make sure that all communications between Spokes A and B are always using the same cluster
member, you must make some changes to the applicable user.def file (see sk98239
http://supportcontent.checkpoint.com/solutions?id=sk98239). This second step ensures that both
third-party gateways always connect to the same cluster member (see "Configuring a Third-Party
Gateway in a Hub and Spoke VPN Deployment" on page 51).
Non-Sticky Connections
A connection is called sticky if a single cluster member handles all packets of that connection. In
a non-sticky connection, the response packet of a connection returns through a different cluster
member than the original request packet.
The cluster synchronization mechanism knows how to handle non-sticky connections properly. In
a non-sticky connection, a cluster member can receive an out-of-state packet, which Firewall
normally drops because it poses a security risk.
In Load Sharing configurations, all cluster members are active. In Static NAT and encrypted
connections, the source and destination IP addresses change. Therefore, Static NAT and
encrypted connections through a Load Sharing cluster may be non-sticky. Non-stickiness may
also occur with Hide NAT, but ClusterXL has a mechanism to make it sticky.
In High Availability configurations, all packets reach the Active member, so all connections are
sticky. If failover occurs during connection establishment, the connection is lost, but
synchronization can be performed later.
If the other members do not know about a non-sticky connection, the packet will be out-of-state,
and the connection will be dropped for security reasons. However, the Synchronization
mechanism knows how to inform other members of the connection. The Synchronization
mechanism thereby prevents out-of-state packets in valid, but non-sticky connections, so that
these non-sticky connections are allowed.
Non-sticky connections will also occur if the network Administrator has configured asymmetric
routing, where a reply packet returns through a different Security Gateway than the original
packet.
Item Description
1 Client
4 Server
The client initiates a connection by sending a SYN packet to the server. The SYN passes through
cluster member A, but the SYN/ACK reply returns through cluster member B. This is a non-sticky
connection, because the reply packet returns through a different Security Gateway than the
original packet.
The synchronization network notifies cluster member B. If cluster member B is updated before
the SYN/ACK packet sent by the server reaches it, the connection is handled normally. If, however,
synchronization is delayed, and the SYN/ACK packet is received on cluster member B before the
SYN flag has been updated, then the Security Gateway treats the SYN/ACK packet as out-of-state,
and drops the connection.
You can configure enhanced 3-Way TCP Handshake (see "Enhanced 3-Way TCP Handshake
Enforcement" on page 132) enforcement to address this issue.
Configuring ClusterXL
In This Section:
Installing Cluster Members .........................................................................................55
Configuring Routing for Client Computers .................................................................56
Selecting the CCP Transport Mode on the Cluster Members ....................................57
Configuring the Cluster Object and Members ............................................................59
Configuring a ClusterXL in Bridge Mode .....................................................................67
This procedure describes how to configure the Load Sharing Multicast, Load Sharing Unicast, and
High Availability modes from scratch. Their configuration is identical, apart from the mode
selection in SmartConsole Cluster object or Cluster creation wizard.
Mode Description
Automatic The CCP mode changes automatically between Multicast, Broadcast, and Unicast
to find the optimized CCP mode according to network state.
These CCP mode changes are independent for individual interfaces.
This CCP mode prevents unnecessary cluster failovers and interface state
changes when CCP packets are not received because of networking issues.
Multicast In this CCP mode, the CCP packets are sent to a multicast Layer 2 destination
MAC address (01:00:5E:xx:yy:zz). See sk25977
http://supportcontent.checkpoint.com/solutions?id=sk25977.
This is the default CCP mode for non-Sync interfaces.
Broadcast In this CCP mode, the CCP packets are sent to a broadcast Layer 2 MAC address
(FF:FF:FF:FF:FF:FF). See sk25977
http://supportcontent.checkpoint.com/solutions?id=sk25977 and sk36644
http://supportcontent.checkpoint.com/solutions?id=sk36644.
This is the default CCP mode for Sync interface.
This is the only supported CCP mode on Bridge interfaces.
Use this CCP mode if the connecting switches do not pass CCP multicast packets.
Example output:
[Expert@Member2:0]# cphaprob -a if
[Expert@Member2:0]#
[Expert@Member2:0]# cphaconf set_ccp auto
[Expert@Member2:0]#
[Expert@Member2:0]# cat $FW_BOOT_DIR/ha_boot.conf
ha_installed 1
ccp_mode automatic
[Expert@Member2:0]#
[Expert@Member2:0]# cphaconf set_ccp broadcast
[Expert@Member2:0]#
[Expert@Member2:0]# cphaprob -a if
[Expert@Member2:0]#
[Expert@Member2:0]# cat $FW_BOOT_DIR/ha_boot.conf
ha_installed 1
ccp_mode broadcast
[Expert@Member2:0]#
[Expert@Member2:0]# grep ccp /config/active
cluster:ccp:broadcast t
[Expert@Member2:0]#
Gaia VRRP
4. In the Cluster member's properties window do these steps for each cluster member and click
Next:
We assume you create a new cluster object from the scratch.
a) Click Add > New Cluster Member to configure each cluster member.
b) In the Cluster Name field, enter unique name for the cluster member object.
c) In the Cluster IPv4 Address, enter the unique Cluster Virtual IPv4 addresses for this
cluster member. This is the main IPv4 address of the cluster member object.
d) In the Cluster IPv6 Address, enter the unique Cluster Virtual IPv6 addresses for this
cluster member. This is the main IPv6 address of the cluster member object.
Important - You must define a corresponding IPv4 address for every IPv6 address. This
release does not support pure IPv6 addresses.
e) In the Activation Key and Confirm Activation Key fields, enter a one-time password that
you entered in First Time Configuration Wizard during the installation of this cluster
member.
f) Click Initialize.
Management Server will try to establish SIC with each cluster member. The Trust State
field should show Trust established.
g) Click OK.
5. In the Cluster Topology window, define a network type (network role) for each cluster
interface and define the Cluster Virtual IP addresses.
The wizard automatically calculates the subnet for each cluster network and assigns it to the
applicable interface on each cluster member. The calculated subnet shows in the upper
section of the window.
The available network objectives are:
• Cluster Interface - A cluster interface that connects to an internal or external network.
Enter the Cluster Virtual IP addresses for each network (internal or external). Also see
Cluster IP Addresses on Different Subnets (on page 133).
• Cluster Sync Interface - A cluster Synchronization interface. In Load Sharing mode, you
must define a synchronization interface.
For Check Point Appliances (for example, 23800) or Open Servers: The Synchronization
Network is supported only on the lowest VLAN tag of a VLAN interface.
For Small Office appliances (for example, 1100, 1200R, 1400): You can only select 1st
sync and only for the LAN2/SYNC interface. You cannot configure VLANs on the
Synchronization interface.
Important Note - Make sure that you do not define IPv6 address for sync interfaces. The
wizard does not let you define an interface with an IPv6 address as a sync interface.
• Private - An interface that is not part of the cluster. Cluster does not monitor this interface.
Cluster failover does not occur if a fault occurs on this interface. This option is
recommended for the dedicated management interface.
Click Next.
6. In the Cluster Definition Wizard Complete window, click Finish.
After you complete the wizard, we recommend that you open the cluster object and complete the
configuration:
• Define Anti-Spoofing properties for each interface
• Change the Topology settings for each interface, if necessary
• Define the Network Type
• Configure other Software Blades, features and properties as necessary
Configuration Steps
Configuring General Properties...................................................................................62
Adding a New Cluster Member to the Cluster Object ................................................62
Adding an Existing Security Gateway as a Cluster Member to the Cluster Object ...63
Deleting a Cluster Member from Cluster Object ........................................................64
Working with Cluster Topology ....................................................................................64
Completing the Cluster Definition ...............................................................................65
Changing the Synchronization Interface .....................................................................65
Private An interface that is not part of the cluster. ClusterXL does not
monitor the state of this interface. As a result, there is no cluster
failover if a fault occurs with this interface. This option is
recommended for the management interface.
Virtual IPv4 - Virtual IPv4 address assigned to this Cluster Virtual Interface
Virtual IPv6 - Virtual IPv6 address assigned to this Cluster Virtual Interface
Important - You must define a corresponding IPv4 address for every IPv6 address. This
release does not support the configuration of only IPv6 addresses.
b) In the Member IPs section, click Modify and configure these settings:
Physical IPv4 address and Mask Length assigned to the applicable physical interface on
each cluster member
Physical IPv6 address and Mask Length assigned to the applicable physical interface on
each cluster member
Important - You must define a corresponding IPv4 address for every IPv6 address. This
release does not support the configuration of only IPv6 addresses.
See also: Configuring Cluster Addresses on Different Subnets.
c) In the Topology section, click Modify and configure these settings:
Leads To - one of these: Internet (External), This Network (Internal)
Security Zone - one of these: User defined, According to topology (ExternalZone,
InternalZone)
Anti-Spoofing - whether to perform the Anti-Spoofing, and how to do it (Detect,
Prevent)
5. From the left navigation tree, click QoS page:
a) In the Bandwidth section, configure these settings:
Inbound Active - rate limit for inbound traffic
Outbound Active - rate limit for outbound traffic
b) In the DiffServ and Low Latency classes section, configure the applicable classes.
6. From the left navigation tree, click Advanced page:
a) In the Multicast Restrictions section, configure the applicable settings for dropping
multicast packets
b) In the Interfaces Names section, configure the names of applicable interfaces
7. Click OK.
Notes:
• This procedure applies to both Check Point Appliances and Open Servers.
• ClusterXL in Bridge Mode deployed in Active/Standby supports only two switches.
The Bridge Active/Standby mode is the preferred mode in topologies that support it.
In the Bridge Active/Standby mode, ClusterXL works in High Availability mode. For more
information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_ClusterXL_
AdminGuide/html_frameset.htm.
Example:
Item Description
1 Network, which an administrator needs to divide into two Layer 2 segments.
The ClusterXL in Bridge Mode connects between these segments.
3 Switch that connects the first network segment to one bridged slave interface (4) on the
ClusterXL in Bridge Mode.
4 One bridged slave interface (for example, eth1) on the Cluster Members in Bridge
Mode.
5 Dedicated Gaia Management Interface (for example, eth0) on the Cluster Members.
6 First Cluster Member in Bridge Mode (for example, in the Active cluster state).
7 Network that connects dedicated synchronization interfaces (for example, eth3) on the
ClusterXL in Bridge Mode.
8 Second Cluster Member in Bridge Mode (for example, in the Standby cluster state).
9 Another bridged slave interface (for example, eth2) on the Cluster Members in Bridge
Mode.
10 Switch that connects the second network segment to the other bridged slave interface
(9) on the ClusterXL in Bridge Mode.
Workflow:
1. Install the two cluster members.
2. Configure the ClusterXL in High Availability mode in SmartConsole - in Wizard Mode, or
Classic Mode.
3. Configure the applicable policy for the ClusterXL Cluster in SmartConsole.
4. Enable the Bridge Active/Standby mode on both cluster members.
Best practice - If you configure Bridge Active/Standby mode, disable STP, RSTP, and MSTP on the
adjacent switches. See the applicable documentation for your switches.
Step Description
3 During the First Time Configuration Wizard, you must configure these settings:
• In the Installation Type window, select Security Gateway and/or Security
Management.
• In the Products window:
a) In the Products section, select Security Gateway only.
b) In the Clustering section, select these two options:
Unit is a part of a cluster
ClusterXL
• In the Secure Internal Communication window, enter the desired Activation Key
(between 4 and 127 characters long).
Step Description
5A On the Cluster members' properties page, add the objects for the cluster members.
a) Click Add > New Cluster Member.
The Cluster Member Properties window opens.
b) In the Name field, enter the desired name for this cluster member object.
c) Configure the main physical IP address(es) for this cluster member object.
In the IPv4 Address and IPv6 Address fields, configure the same IPv4 and IPv6
addresses that you configured on the Management Connection page of the cluster
member's First Time Configuration Wizard. Make sure the Security Management
Server or Multi-Domain Server can connect to these IP addresses.
d) In the Activation Key and Confirm Activation Key fields, enter the same Activation
Key you entered during the cluster member's First Time Configuration Wizard.
e) Click Initialize.
f) Click OK.
g) Repeat Steps a-f to add the second cluster member, and so on.
5B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
6 On the Cluster Topology pages, configure the roles of the cluster interfaces:
a) Examine the IPv4 Network Address at the top of the page.
b) Select the applicable role:
For cluster traffic interfaces, select Representing a cluster interface and
configure the Cluster Virtual IPv4 address and its Net Mask.
For cluster synchronization interfaces, select Cluster Synchronization and
select Primary only. Check Point cluster supports only one synchronization
network. For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private use of each member (don't monitor members interfaces).
c) Click Next.
Step Description
7 On the Cluster Definition Wizard Complete page:
a) Examine the Configuration Summary.
b) Select Edit Cluster's Properties.
c) Click Finish.
The Gateway Cluster Properties window opens.
9 On the General Properties page > Platform section, select the correct options:
a) In the Hardware field:
If you install the cluster members on Check Point Appliances, select the correct
appliances series.
If you install the cluster members on Open Servers, select Open server.
b) In the Version field, select R80.20.
c) In the OS field, select Gaia.
Step Description
11A On the Cluster Members page:
a) Click Add > New Cluster Member.
The Cluster Member Properties window opens.
b) In the Name field, enter the desired name for this cluster member object.
c) Configure the main physical IP address(es) for this cluster member object.
In the IPv4 Address and IPv6 Address fields, configure the same IPv4 and IPv6
addresses that you configured on the Management Connection page of the cluster
member's First Time Configuration Wizard. Make sure the Security Management
Server or Multi-Domain Server can connect to these IP addresses.
d) Click Communication.
e) In the One-time password and Confirm one-time password fields, enter the same
Activation Key you entered during the cluster member's First Time Configuration
Wizard.
f) Click Initialize.
g) Click Close.
h) Click OK.
i) Repeat Steps a-h to add the second cluster member, and so on.
11B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
Step Description
12 On the ClusterXL and VRRP page:
a) In the Select the cluster mode and configuration section, select High Availability
and ClusterXL.
b) In the Tracking section, select the desired option.
c) In the Advanced Settings section:
Optional: Select Use State Synchronization. We recommend to select this
option.
Optional: Select Use Virtual MAC (for more information, see sk50840
http://supportcontent.checkpoint.com/solutions?id=sk50840).
Select the High Availability recovery - Maintain current active Cluster Member,
or Switch to higher priority Cluster Member. For more information, see the
R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
Step Description
13 On the Network Management page:
a) Select each interface and click Edit. The Network: <Name of Interface> window
opens.
b) From the left navigation tree, click the General page.
c) In the General section, in the Network Type field, select the applicable type:
For cluster traffic interfaces, select Cluster. Make sure the Cluster Virtual IPv4
address and its Net Mask are correct.
For cluster synchronization interfaces, select Sync or Cluster+Sync (we do not
recommend this configuration). Check Point cluster supports only one
synchronization network. For more information, see the R80.20 ClusterXL
Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private.
d) In the Member IPs section, make sure the IPv4 address and its Net Mask are
correct on each Cluster Member.
Note - For cluster traffic interfaces, you can configure the Cluster Virtual IP address
to be on a different network than the physical IP addresses of the cluster members.
In this case, you must configure the required static routes on the cluster members.
For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.2
0_ClusterXL_AdminGuide/html_frameset.htm.
e) In the Topology section:
Make sure the settings are correct in the Leads To and Security Zone fields.
Make sure to enable the Anti-Spoofing.
Important:
• Make sure the Bridge interface and Bridge slave interfaces are not in the Topology.
• You cannot define the Topology of the Bridge interface. It is External by default.
14 Configure other applicable settings for this ClusterXL object.
15 Click OK.
Step Description
2 Create a new Cluster object in one of these ways:
• From the top toolbar, click the New ( ) > Cluster > Cluster.
• In the top left corner, click Objects menu > More object types > Network Object >
Gateways and Servers > Cluster > New Cluster.
• In the top right corner, click Objects Pane > New > More > Network Object > Gateways
and Servers > Cluster > Cluster.
3 In the Check Point Security Gateway Creation window, click Classic Mode.
The Gateway Cluster Properties window opens.
5 On the General Properties page > Platform section, select the correct options:
a) In the Hardware field:
If you install the cluster members on Check Point Appliances, select the correct
appliances series.
If you install the cluster members on Open Servers, select Open server.
b) In the Version field, select R80.20.
c) In the OS field, select Gaia.
Step Description
7A On the Cluster Members page:
a) Click Add > New Cluster Member.
The Cluster Member Properties window opens.
b) In the Name field, enter the desired name for this cluster member object.
c) Configure the main physical IP address(es) for this cluster member object.
In the IPv4 Address and IPv6 Address fields, configure the same IPv4 and IPv6
addresses that you configured on the Management Connection page of the cluster
member's First Time Configuration Wizard. Make sure the Security Management
Server or Multi-Domain Server can connect to these IP addresses.
d) Click Communication.
e) In the One-time password and Confirm one-time password fields, enter the same
Activation Key you entered during the cluster member's First Time Configuration
Wizard.
f) Click Initialize.
g) Click Close.
h) Click OK.
i) Repeat Steps a-h to add the second cluster member, and so on.
7B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
Step Description
8 On the ClusterXL and VRRP page:
a) In the Select the cluster mode and configuration section, select High Availability
and ClusterXL.
b) In the Tracking section, select the desired option.
c) In the Advanced Settings section:
Optional: Select Use State Synchronization. We recommend to select this
option.
Optional: Select Use Virtual MAC (for more information, see sk50840
http://supportcontent.checkpoint.com/solutions?id=sk50840).
Select the High Availability recovery - Maintain current active Cluster Member,
or Switch to higher priority Cluster Member. For more information, see the
R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
Step Description
9 On the Network Management page:
a) Select each interface and click Edit. The Network: <Name of Interface> window
opens.
b) From the left navigation tree, click the General page.
c) In the General section, in the Network Type field, select the applicable type:
For cluster traffic interfaces, select Cluster. Make sure the Cluster Virtual IPv4
address and its Net Mask are correct.
For cluster synchronization interfaces, select Sync or Cluster+Sync (we do not
recommend this configuration). Check Point cluster supports only one
synchronization network. For more information, see the R80.20 ClusterXL
Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private.
d) In the Member IPs section, make sure the IPv4 address and its Net Mask are
correct on each Cluster Member.
Note - For cluster traffic interfaces, you can configure the Cluster Virtual IP address
to be on a different network than the physical IP addresses of the cluster members.
In this case, you must configure the required static routes on the cluster members.
For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.2
0_ClusterXL_AdminGuide/html_frameset.htm.
e) In the Topology section:
Make sure the settings are correct in the Leads To and Security Zone fields.
Make sure to enable the Anti-Spoofing.
Important:
• Make sure the Bridge interface and Bridge slave interfaces are not in the Topology.
• You cannot define the Topology of the Bridge interface. It is External by default.
10 Configure other applicable settings for this ClusterXL object.
11 Click OK.
Step Description
3 Create a new policy and configure the applicable layers:
a) At the top, click the + tab (or press CTRL T).
b) On the Manage Policies tab, click Manage policies and layers.
c) In the Manage policies and layers window, create a new policy and configure the
applicable layers.
d) Click Close.
e) On the Manage Policies tab, click the new policy you created.
2 Run: cpconfig
4 Enter y to confirm.
Item Description
8 Examine the cluster configuration on each cluster member.
• In Gaia Clish, run:
show cluster state
• In Expert mode, run:
cphaprob state
Example output:
Cluster Mode: High Availability (Active Up, Bridge Mode) with IGMP Membership
Number Unique Address Firewall State (*)
1 (local) 2.2.2.3 Active
2 2.2.2.2 Standby
When you define a Bridge interface on a cluster member, Bridge Active/Active mode is enabled by
default.
Notes:
• This procedure applies to both Check Point Appliances and Open Servers.
• This procedure describes ClusterXL in Bridge Active/Active Mode deployed with two switches.
In the Bridge Active/Active mode, ClusterXL works in Load Sharing mode. For more information,
see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_ClusterXL_
AdminGuide/html_frameset.htm.
Example:
Item Description
1 Network, which an administrator needs to divide into two Layer 2 segments.
The ClusterXL in Bridge Mode connects between these segments.
3 Switch that connects the first network segment to one bridged slave interface (4) on the
ClusterXL in Bridge Mode.
4 One bridged slave interface (for example, eth1) on the Cluster Members in Bridge
Mode.
5 Dedicated Gaia Management Interface (for example, eth0) on the Cluster Members.
6 First Cluster Member in Bridge Mode (in the Active cluster state).
7 Network that connects dedicated synchronization interfaces (for example, eth3) on the
ClusterXL in Bridge Mode.
8 Second Cluster Member in Bridge Mode (in the Active cluster state).
9 Another bridged slave interface (for example, eth2) on the Cluster Members in Bridge
Mode.
10 Switch that connects the second network segment to the other bridged slave interface
(9) on the ClusterXL in Bridge Mode.
Item Description
11 Second network segment.
Workflow:
1. Install the two cluster members.
2. Configure the Bridge interface on both cluster members - in Gaia Portal, or Gaia Clish.
3. Configure the ClusterXL in High Availability mode in SmartConsole - in Wizard Mode, or
Classic Mode.
4. Configure the applicable policy for the ClusterXL Cluster in SmartConsole.
3 During the First Time Configuration Wizard, you must configure these settings:
• In the Installation Type window, select Security Gateway and/or Security
Management.
• In the Products window:
a) In the Products section, select Security Gateway only.
b) In the Clustering section, select these two options:
Unit is a part of a cluster
ClusterXL
• In the Secure Internal Communication window, enter the desired Activation Key
(between 4 and 127 characters long).
Step Description
1 In your web browser, connect to the Gaia Portal on the Security Gateway.
2 In the left navigation tree, click Network Management > Network Interfaces.
3 Make sure that the slave interfaces, which you wish to add to the Bridge interface, do not
have IP addresses.
Step Description
5 On the Bridge tab, enter or select a Bridge Group ID (unique integer between 1 and 1024).
6 Select the interfaces from the Available Interfaces list and then click Add.
Notes:
• A Bridge interface in Gaia can contain only two slave interfaces.
• Do not select the interface that you configured as Gaia Management Interface.
7 On the IPv4 tab, do not enter the IPv4 address.
9 Click OK.
Step Description
1 Connect to the command line on the Security Gateway.
3 Make sure that the slave interfaces, which you wish to add to the Bridge interface, do not
have IP addresses. Run:
show interface <Name of Interface> ipv4-address
show interface <Name of Interface> ipv6-address
Step Description
2 Create a new Cluster object in one of these ways:
• From the top toolbar, click the New ( ) > Cluster > Cluster.
• In the top left corner, click Objects menu > More object types > Network Object >
Gateways and Servers > Cluster > New Cluster.
• In the top right corner, click Objects Pane > New > More > Network Object > Gateways
and Servers > Cluster > Cluster.
3 In the Check Point Security Gateway Cluster Creation window, click Wizard Mode.
5A On the Cluster members' properties page, add the objects for the cluster members.
a) Click Add > New Cluster Member.
The Cluster Member Properties window opens.
b) In the Name field, enter the desired name for this cluster member object.
c) Configure the main physical IP address(es) for this cluster member object.
In the IPv4 Address and IPv6 Address fields, configure the same IPv4 and IPv6
addresses that you configured on the Management Connection page of the cluster
member's First Time Configuration Wizard. Make sure the Security Management
Server or Multi-Domain Server can connect to these IP addresses.
d) In the Activation Key and Confirm Activation Key fields, enter the same Activation
Key you entered during the cluster member's First Time Configuration Wizard.
e) Click Initialize.
f) Click OK.
g) Repeat Steps a-f to add the second cluster member, and so on.
Step Description
5B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
6 On the Cluster Topology page, configure the roles of the cluster interfaces:
a) Examine the IPv4 Network Address at the top of the page.
b) Select the applicable role:
For cluster traffic interfaces, select Representing a cluster interface and
configure the Cluster Virtual IPv4 address and its Net Mask.
For cluster synchronization interfaces, select Cluster Synchronization and
select Primary only. Check Point cluster supports only one synchronization
network. For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private use of each member (don't monitor members interfaces).
c) Click Next.
Step Description
9 On the General Properties page > Platform section, select the correct options:
a) In the Hardware field:
If you install the cluster members on Check Point Appliances, select the correct
appliances series.
If you install the cluster members on Open Servers, select Open server.
b) In the Version field, select R80.20.
c) In the OS field, select Gaia.
Step Description
11B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
Step Description
13 On the Network Management page:
a) Select each interface and click Edit. The Network: <Name of Interface> window
opens.
b) From the left navigation tree, click the General page.
c) In the General section, in the Network Type field, select the applicable type:
For cluster traffic interfaces, select Cluster. Make sure the Cluster Virtual IPv4
address and its Net Mask are correct.
For cluster synchronization interfaces, select Sync or Cluster+Sync (we do not
recommend this configuration). Check Point cluster supports only one
synchronization network. For more information, see the R80.20 ClusterXL
Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private.
d) In the Member IPs section, make sure the IPv4 address and its Net Mask are
correct on each Cluster Member.
Note - For cluster traffic interfaces, you can configure the Cluster Virtual IP address
to be on a different network than the physical IP addresses of the cluster members.
In this case, you must configure the required static routes on the cluster members.
For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.2
0_ClusterXL_AdminGuide/html_frameset.htm.
e) In the Topology section:
Make sure the settings are correct in the Leads To and Security Zone fields.
Make sure to enable the Anti-Spoofing.
Important:
• Make sure the Bridge interface and Bridge slave interfaces are not in the Topology.
• You cannot define the Topology of the Bridge interface. It is External by default.
14 Configure other applicable settings for this ClusterXL object.
15 Click OK.
Step Description
2 Create a new Cluster object in one of these ways:
• From the top toolbar, click the New ( ) > Cluster > Cluster.
• In the top left corner, click Objects menu > More object types > Network Object >
Gateways and Servers > Cluster > New Cluster.
• In the top right corner, click Objects Pane > New > More > Network Object > Gateways
and Servers > Cluster > Cluster.
3 In the Check Point Security Gateway Creation window, click Classic Mode.
The Gateway Cluster Properties window opens.
5 On the General Properties page > Platform section, select the correct options:
a) In the Hardware field:
If you install the cluster members on Check Point Appliances, select the correct
appliances series.
If you install the cluster members on Open Servers, select Open server.
b) In the Version field, select R80.20.
c) In the OS field, select Gaia.
Step Description
7A On the Cluster Members page:
a) Click Add > New Cluster Member.
The Cluster Member Properties window opens.
b) In the Name field, enter the desired name for this cluster member object.
c) Configure the main physical IP address(es) for this cluster member object.
In the IPv4 Address and IPv6 Address fields, configure the same IPv4 and IPv6
addresses that you configured on the Management Connection page of the cluster
member's First Time Configuration Wizard. Make sure the Security Management
Server or Multi-Domain Server can connect to these IP addresses.
d) Click Communication.
e) In the One-time password and Confirm one-time password fields, enter the same
Activation Key you entered during the cluster member's First Time Configuration
Wizard.
f) Click Initialize.
g) Click Close.
h) Click OK.
i) Repeat Steps a-h to add the second cluster member, and so on.
7B If the Trust State field does not show Established, perform these steps:
a) Connect to the command line on the cluster member.
b) Make sure there is a physical connectivity between the cluster member and the
Management Server (for example, pings can pass).
c) Run: cpconfig
d) Enter the number of this option: Secure Internal Communication.
e) Follow the instructions on the screen to change the Activation Key.
f) In the SmartConsole, click Reset.
g) Enter the same Activation Key you entered in the cpconfig menu.
h) Click Initialize.
Step Description
8 On the ClusterXL and VRRP page:
a) In the Select the cluster mode and configuration section, select High Availability
and ClusterXL.
b) In the Tracking section, select the desired option.
c) In the Advanced Settings section:
Optional: Select Use State Synchronization. We recommend to select this
option.
Optional: Select Use Virtual MAC (for more information, see sk50840
http://supportcontent.checkpoint.com/solutions?id=sk50840).
Select the High Availability recovery - Maintain current active Cluster Member,
or Switch to higher priority Cluster Member. For more information, see the
R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
Step Description
9 On the Network Management page:
a) Select each interface and click Edit. The Network: <Name of Interface> window
opens.
b) From the left navigation tree, click the General page.
c) In the General section, in the Network Type field, select the applicable type:
For cluster traffic interfaces, select Cluster. Make sure the Cluster Virtual IPv4
address and its Net Mask are correct.
For cluster synchronization interfaces, select Sync or Cluster+Sync (we do not
recommend this configuration). Check Point cluster supports only one
synchronization network. For more information, see the R80.20 ClusterXL
Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R8
0.20_ClusterXL_AdminGuide/html_frameset.htm.
For interfaces that do not pass the traffic between the connected networks,
select Private.
d) In the Member IPs section, make sure the IPv4 address and its Net Mask are
correct on each Cluster Member.
Note - For cluster traffic interfaces, you can configure the Cluster Virtual IP address
to be on a different network than the physical IP addresses of the cluster members.
In this case, you must configure the required static routes on the cluster members.
For more information, see the R80.20 ClusterXL Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.2
0_ClusterXL_AdminGuide/html_frameset.htm.
e) In the Topology section:
Make sure the settings are correct in the Leads To and Security Zone fields.
Make sure to enable the Anti-Spoofing.
Important:
• Make sure the Bridge interface and Bridge slave interfaces are not in the Topology.
• You cannot define the Topology of the Bridge interface. It is External by default.
10 Configure other applicable settings for this ClusterXL object.
11 Click OK.
Step Description
3 Create a new policy and configure the applicable layers:
a) At the top, click the + tab (or press CTRL T).
b) On the Manage Policies tab, click Manage policies and layers.
c) In the Manage policies and layers window, create a new policy and configure the
applicable layers.
d) Click Close.
e) On the Manage Policies tab, click the new policy you created.
Example:
Item Description
1 Network, which an administrator needs to divide into two Layer 2 segments.
The ClusterXL in Bridge Mode connects between these segments.
3 Switch that connects the first network segment to one bridged slave interface (6) on the
ClusterXL in Bridge Mode.
4 Switch that connects between one switch (that directly connects to the first network
segment) and one bridged slave interface (6) on the ClusterXL in Bridge Mode.
5 Dedicated Gaia Management Interface (for example, eth0) on the Cluster Members.
6 One bridged slave interface (for example, eth1) on the Cluster Members in Bridge
Mode.
7 First Cluster Member in Bridge Mode (in the Active cluster state).
8 Network that connects dedicated synchronization interfaces (for example, eth3) on the
ClusterXL in Bridge Mode.
9 Second Cluster Member in Bridge Mode (in the Active cluster state).
10 Another bridged slave interface (for example, eth2) on the Cluster Members in Bridge
Mode.
Item Description
11 Switch that connects the second network segment to the other bridged slave interface
(10) on the ClusterXL in Bridge Mode.
12 Switch that connects between one switch (that directly connects to the second network
segment) and the other bridged slave interface (10) on the ClusterXL in Bridge Mode.
Workflow:
1. Install the two cluster members.
2. Configure the Bridge interface on both cluster members - in Gaia Portal, or Gaia Clish.
3. Configure the ClusterXL in High Availability mode in SmartConsole - in Wizard Mode, or
Classic Mode.
4. Configure the applicable policy for the ClusterXL Cluster in SmartConsole.
The workflow and detailed instructions are the same as in the Configuring ClusterXL in Bridge
Mode - Active/Active with Two Switches (on page 80).
2. In the Cluster Member page of the Cluster Properties window, edit the Cluster Member
object.
3. In the Cluster Member Properties window, click the NAT tab.
4. Configure Static or Hide NAT as desired.
VLAN interfaces Monitors only lowest VLAN ID VSX High Availability (non-VSLS):
configured on a physical interface
Monitors only lowest and highest
VLAN IDs configured on a physical
interface
Only the lowest and This is the default behavior VSX High Availability (non-VSLS):
highest VLAN IDs This is the default behavior
All VLAN IDs Not supported Virtual System Load Sharing: This is
the default behavior
(*) In a VSX cluster, to disable the default VLAN monitoring behavior, you need to set the value of
How the Destination Cluster MAC Address is Assigned in Load Sharing Multicast Mode
When a member that is outside the cluster wishes to communicate with the cluster, it sends an
ARP query with the cluster (virtual) IP address. The cluster replies to the ARP request with a
multicast MAC address, even though the IP address is a unicast address.
This destination multicast MAC address of the cluster is based on the unicast IP address of the
cluster. The upper three bytes are 01.00.5E, and they identify a Multicast MAC in the standard way.
The lower three bytes are the same as the lower three bytes of the IP address. An example MAC
address based on the IP address 10.0.10.11 is shown below.
For example, it is OK for the cluster interface of one of the clusters connected to the VLAN to have
the address 10.0.10.11, and the cluster interface of a second cluster to have the address
10.0.10.12. However, the following addresses for the interfaces of the first and second clusters
will cause complications: 10.0.10.11 and 20.0.10.11.
254 is the default value and should already be set. If duplicate Source MAC addresses of
CCP packets appear on the network even though automatic mode is set, then enter unique
values for each cluster (manual mode).
• To work in manual mode (only if instructed so by Check Point Support), enter an integer
value between 1 and 253.
Enter a unique value for each cluster in the domain.
9. Repeat steps 4-8 for each cluster.
10. Save the changes: go to File menu > click Save All.
11. Close the GuiDBedit Tool.
12. Connect with SmartConsole to Security Management Server or Domain Management Server.
13. Install the policy on the Cluster object.
14. Examine the MAC magic value on each cluster member:
cphaprob mmagic
All cluster members of the same cluster should have the save MAC magic value.
Example:
[Expert@MemberB:0]# cphaprob mmagic
To change the MAC magic ID during a Connectivity Upgrade (R80.10 and higher):
Before the upgrade, find out the current configuration mode.
1. Connect to one of the cluster members.
2. Run cphaprob mmagic
The Configuration Mode field will show manual or automatic.
If the configuration field is automatic, upgrade the cluster. The upgraded cluster member
will learn the MAC magic value from a member that has not yet been upgraded (if a value
exists). Select a value if no previous value exists.
Note - If the configuration field shows manual, and you want to continue to use manual
configuration, the same MAC magic value must be reused.
3. Before the upgrade, find the configured MAC magic value:
• For R80.10 and higher, run: cphaprob mmagic
• For R77.30, run: cphaconf cluster_id get
• For R77.20 and below, run: fw ctl get int fwha_mac_magic
4. If you are working in manual mode, connect to the management server using GuiDBedit Tool:
a) Connect with GuiDBedit Tool (see sk13009
http://supportcontent.checkpoint.com/solutions?id=sk13009) to Security Management
Server or Domain Management Server.
b) In the upper left pane, go to Table > Network Objects > network_objects.
c) In the upper right pane, select the applicable Cluster object.
d) Press CTRL+F (or go to Search menu > Find) > paste cluster_magic > click Find Next.
ClusterXL Administration Guide R80.20 | 104
Advanced Features and Procedures
1A Interface 1
1B interface 2
2 Bond Interface
3 Router
A bond interface (also known as a bonding group or bond) is identified by its Bond ID (for
example: bond1) and is assigned an IP address. The physical interfaces included in the bond are
called slaves and do not have IP addresses.
You can define a bond interface to use one of these functional strategies:
• High Availability (Active/Backup): Gives redundancy when there is an interface or a link
failure. This strategy also supports switch redundancy. Bond High Availability works in
Active/Backup mode - interface Active/Standby mode. When an Active slave interface is down,
the connection automatically fails over to the primary slave interface. If the primary slave
interface is not available, the connection fails over to a different slave interface.
• Load Sharing (Active/Active): All slave interfaces in the UP state are used simultaneously.
Traffic is distributed among the slave interfaces to maximize throughput. Bond Load Sharing
does not support switch redundancy.
Note - Bonding Load Sharing mode requires SecureXL to be enabled on Security Gateway or
each cluster member.
You can configure Bond Load Sharing to use one of these modes:
• Round Robin - Selects the Active slave interfaces sequentially.
• 802.3ad - Dynamically uses Active slave interfaces to share the traffic load. This mode uses
the LACP protocol, which fully monitors the interface link between the Check Point Security
Gateway and a switch.
• XOR - All slave interfaces in the UP state are Active for Load Sharing. Traffic is assigned to
Active slave interfaces based on the transmit hash policy: Layer 2 information (XOR of
hardware MAC addresses), or Layer 3+4 information (IP addresses and Ports).
For Bonding High Availability mode and for Bonding Load Sharing mode:
• The number of bond interfaces that can be defined is limited by the maximal number of
interfaces supported by each platform. See the R80.20 Release Notes
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_RN/html
_frameset.htm.
• Up to 8 slave interfaces can be configured in a single High Availability or Load Sharing bond
interface.
When dealing with mission-critical applications, an enterprise requires its network to be highly
available.
Clustering provides redundancy, and thus, High Availability, at the Security Gateway level. Without
Bonding, redundancy of Network Interface Cards (NICs) or of the switches on either side of the
Security Gateway are only possible in a cluster, and only by failover of the Security Gateway to
another cluster member.
Item Description
1 Cluster member GW1 with interfaces connected to the external switches (items 5 and
6)
2 Cluster member GW2 with interfaces connected to the external switches (items 5 and
6)
3 Interconnecting network C1
4 Interconnecting network C2
5 Switch S1
6 Switch S2
If cluster member GW1 (item 1), its NIC, or switch S1 (item 5) fails, cluster member GW2 (item 2)
becomes the only Active member, connecting to switch S2 (item 6) over network C2 (item 4). If any
component fails (cluster member, NIC, or switch), the result of the failover is that no further
redundancy exists. A further failure of any active component completely stops network traffic
through this cluster.
Item Description
1 Cluster member GW1 with interfaces connected to the external switches (items 4 and
5)
2 Cluster member GW2 with interfaces connected to the external switches (items 4 and
5)
3 Interconnecting network
4 Switch S1
5 Switch S2
In this scenario:
• GW1 and GW2are cluster members in the High Availability mode, each connected to the two
external switches
• S1 and S2 are external switches
• Item 3 are the network connections
If any of the interfaces on a cluster member that connect to an external switch fails, the other
interface continues to provide the connectivity.
If any cluster member, its NIC, or switch fails, the other cluster member, connecting to switch S2
over network C2. If any component fails (cluster member , NIC, or switch), the result of the
failover is that no further redundancy exists. A further failure of any active component completely
stops network traffic.
Bonding provides High Availability of NICs. If one fails, the other can function in its place.
Sync Redundancy
The use of more than one physical synchronization interface (1st sync, 2nd sync, 3rd sync) for
synchronization redundancy is not supported.
For synchronization redundancy, you can use bond interfaces.
To set the value of the kernel parameter temporarily (does not survive reboot):
Step Description
1 Connect to the command line on each VRRP cluster member.
4 Make sure the value of the kernel parameter fwha_bond_enhanced_enable was set
to 1:
[Expert@Member_HostName:0]# fw ctl get int fwha_bond_enhanced_enable
5 Add this line to the file (spaces and comments are not allowed):
fwha_bond_enhanced_enable=1
Step Description
8 Make sure the value of the kernel parameter fwha_bond_enhanced_enable was set
to 1:
[Expert@Member_HostName:0]# fw ctl get int fwha_bond_enhanced_enable
Important - If you change your cluster configuration from VRRP to ClusterXL, you must remove
the kernel parameter configuration from each cluster member.
interface (when n-2 slave interfaces remain up) will cause the entire bond interface to be
considered down, even if the bond contains more than two slave interfaces.
If a smaller number of slave interfaces will be able to handle the expected traffic, you can increase
redundancy by explicitly defining the critical minimal number of slave interfaces. Divide your
maximal expected traffic speed by the speed of your slave interfaces and round up to a whole
number to determine an appropriate number of critical slave interfaces.
To define the critical number of slave interfaces explicitly, create and edit the following file:
$FWDIR/conf/cpha_bond_ls_config.conf
Each line of the file should be written in the following syntax:
<Name_of_Bond> <critical_minimal number_of_slave>
For example, if bond0 has 7 slave interfaces, and bond1 has 6 slave interfaces, file contents
could be:
bond0 5
bond1 3
In this example:
• bond0 would be considered down when 3 of its slave interfaces have failed.
• bond1 would be considered down when 4 of its slave interfaces have failed.
Group of Bonds
Introduction
Group of Bonds, which is a logical group of existing Bond interfaces, provides additional link
redundancy.
Example topology with Group of Bonds:
• There is one router - R
• There are two switches that connect to the router R - SW-1 and SW-2
• There are two cluster members GW-A (Active) and GW-B (Standby)
• There are two Bond interfaces on each cluster member - Bond-1 and Bond-2
• On the cluster member GW-A:
• Bond-1 interface connects to the switch SW-1
• Bond-2 interface connects to the switch SW-2
• On the cluster member GW-B:
• Bond-1 interface connects to the switch SW-2
• Bond-2 interface connects to the switch SW-1
3. On the cluster member GW-A, the Critical Device Interface Active Check reports its state as
"problem".
4. The cluster member GW-A changes its cluster state from Active to Down.
5. The cluster fails over - the cluster member GW-B changes its cluster state from Standby to
Active.
This is not the desired behavior, because the cluster member GW-A connects not only to the
switch SW-1, but also to the switch SW-2. In our example topology, there is no actual reason to
fail-over from the cluster member GW-A to the cluster member GW-B.
In order to overcome this problem, cluster members use the Group of Bonds consisting of Bond-1
and Bond-2. The Group of Bonds fails only when both Bond interfaces fail on the cluster member.
Only then the cluster fails over.
Chain of events with configured Group of Bonds:
1. The cluster member GW-A is the Active and the cluster member GW-B is the Standby.
2. On the cluster member GW-A, the Bond-1 interface fails.
3. On the cluster member GW-A, the Critical Device Interface Active Check reports its state as
"problem".
4. The cluster member GW-A does not change its cluster state from Active to Down.
5. On the cluster member GW-A, the Bond-2 interface fails as well.
6. The cluster member GW-A changes its cluster state from Active to Down.
7. The cluster fails over - the cluster member GW-B changes its cluster state from Standby to
Active.
Notes:
• The apostrophe characters are mandatory part of the syntax.
• Spaces are not allowed in the value of the kernel parameter
fwha_group_of_bonds_str.
Example:
fw ctl set str fwha_group_of_bonds_str 'GoB0:bond0,bond1;GoB1:bond2,bond3,bond4'
9. Make sure the cluster member accepted the new configuration:
[Expert@Member:0]# fw ctl get str fwha_group_of_bonds_str
10. In SmartConsole, install the Access Control Policy on the cluster object.
Monitoring
To see the configured Groups of Bonds, run the cphaprob show_bond_groups ("Monitoring
Bond Interfaces" on page 187) command.
Logs
Cluster members generate some applicable logs, which you can see in:
• In non-VSX Cluster: In the /var/log/messages files and output of the dmesg command.
• In VSX Cluster: In the $FWDIR/log/fwk.elg file in the context of the applicable Virtual
System.
To see the full logs, you need to enable the kernel debug:
• In the kernel module fw, debug flags error and ioctl
• Kernel module cluster, debug flag if
The full logs show:
• A physical slave interface goes down/up.
• A Bond interface goes down/up (regardless of a change in the cluster member's state).
• A Group of Bonds goes down/up.
Note - These logs are generated only once per event.
Limitations
• The maximal length on the text string <Name for Group of Bonds> is 16 characters.
• The maximal length on this text string is 1024 characters:
<Name for Group of Bonds>:<List of all Bonds in this Group separated by the comma>
• You can configure the maximum of five Groups of Bonds on a cluster member or Virtual
System.
• You can configure the maximum of five Bond interfaces in each Groups of Bonds.
• Group of Bonds feature does support Virtual Switches and Virtual Routers. Meaning, do not
configure Groups of Bonds in the context of these Virtual Devices.
• Group of Bonds feature supports only Bond interfaces that belong to the same Virtual System.
You cannot configure bonds that belong to different Virtual Systems into the same Group of
Bonds.
You must perform all configuration in the context of the applicable Virtual System.
• Group of Bonds feature does support Sync interfaces (an interface on a cluster member,
whose Network Type was set as Sync or Cluster+Sync in SmartConsole in cluster object).
• Group of Bonds feature does support Bridge interfaces.
• If a Bond interface goes down on one cluster member, the cphaprob show_bond_groups
("Monitoring Bond Interfaces" on page 187) command on the peer cluster members also
shows the same Bond interace as DOWN.
This is because the peer cluster members stop receiving the CCP packets on that Bond
interface and cannot probe the local network to determine that their Bond interface is really
working.
• After you add a Bond interface to the existing Group of Bonds, you must install the Access
Control Policy on the cluster object.
• After you remove a Bond interface from the existing Group of Bonds, you must install the
Access Control Policy on the cluster object.
Setting Affinities
If you are running SecureXL in a multi-core system, after you define bonds, set affinities manually.
Use the sim affinity -s command.
Note - The sim affinity commands take effect only if the SecureXL is enabled and actually
running. SecureXL begins running when you install a Policy for the first time.
eth1 eth4
eth2 eth5
Two of the CPU cores will need to handle two interfaces each. An optimal configuration can be:
bond0 bond1
eth0 core 0 eth3 core 0
eth2 core 2
eth5 core 3
Troubleshooting Workflow
1. Examine the status of the bond ("Verifying that the Bond is Functioning Properly" on page 112).
2. If there is a problem, see if the physical link is down:
a) Run this command:
In Gaia Clish:
show cluster bond name <bond_name>
In Expert mode:
cphaprob show_bond <bond_name>
b) Look for a slave interface that reports the status of the link as no.
c) Examine the cable connections and other hardware.
d) Examine the port configuration on the switch.
3. See if a cluster member is down:
• In Gaia Clish:
show cluster state
• In Expert mode:
cphaprob state
If any of the cluster members has a firewall State other than Active, continue with the
cphaprob state troubleshooting ("ClusterXL Monitoring Commands" on page 167).
4. View the logs in SmartConsole > Logs & Monitor > Logs.
In a VSX cluster member, reboot is needed after these actions on a bond interface:
1. Changing a bond mode.
2. Adding a slave interface into a bond.
Note - Removing a slave interface does not require reboot.
Summary table:
ICMP_CONN_ALLOWED 1
Load Sharing Uncast MAC address of the Pivot cluster member's interface
The use of different subnets with cluster objects has some limitations (see "Limitations of Cluster
Addresses on Different Subnets" on page 136).
If you do not define the static routes correctly, it will not be possible to connect to the cluster
members and pass traffic through them.
Note - It is not necessary to configure static routes manually on VSX cluster members. This is
done automatically when you configure routes in SmartConsole.
For configuration instructions, see the R80.20 Gaia Administration Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_Gaia_Admin
Guide/html_frameset.htm - Chapter Network Management - Sections IPv4 Static Routes and IPv6
Static Routes.
5. Click OK.
6. Install the Access Control Policy on this cluster object.
Configuring Anti-Spoofing
1. In SmartConsole, define a Group object, which contains the objects of both the external
network and the internal network.
In the example ("Example of Cluster IP Addresses on Different Subnets" on page 135), suppose
Side "A" is the external network, and Side "B" is the internal network.
You must define a Group object, which contains both network 172.16.4.0 and network
192.168.2.0.
2. Open the cluster object.
3. Go to Network Management pane.
4. Select the cluster interface and click Edit.
5. On the General page, in the Topology section, click Modify.
6. Select Override.
7. Select This Network (Internal).
8. Select Specific and select the Group object that contains the objects of both the external
network and the internal network.
9. Click OK.
10. Install the Access Control Policy on this cluster object.
Important - The new Security Management Server must have the same host name as the
existing Standalone Computer.
1. Do a clean Security Management Server installation based on the procedures in the R80.20
Installation and Upgrade Guide
https://sc1.checkpoint.com/documents/R80.20_GA/WebAdminGuides/EN/CP_R80.20_Installat
ion_and_Upgrade_Guide/html_frameset.htm. Make sure that you only select Management
Server options.
Make sure that you install all Hotfixes and plug-ins that were installed in the existing
Standalone computer.
2. Close all of the Expert mode shells. Log into the regular shell.
3. Copy the exported database files to a temporary folder on the new Security Management
Server.
4. Import the management databases, from the Expert mode, run:
# cd $FWDIR/bin/upgrade_tools/
# ./upgrade_import /<path_to>/<export_file_name>
Important - If the import fails with the Database migration between standalone and
management only machines is not supported error, see sk61681
http://supportcontent.checkpoint.com/solutions?id=sk61681 for a workaround.
5. Connect with SmartConsole to the new Security Management Server and make sure that all
settings are correct.
6. Close SmartConsole and reboot the computer.
Make sure that you only select Network Security tab options.
Make sure that you install all Hotfixes and plug-ins that were installed in the existing
Standalone Computer.
2. In SmartConsole, create and configure the Security Gateway object.
Make sure that you establish SIC trust.
3. In SmartConsole, install the Access Control Policy on this Security Gateway object.
4. Connect the systems to the network.
5. Thoroughly test and debug the deployment.
Make sure that the rules for all Software Blades work correctly.
This Security Gateway will become the Source Member for the new ClusterXL cluster.
2. In the Cluster Members page, click Add > Add Existing Gateway.
3. Connect to computer 'B', and define its topology.
4. Define the Synchronization networks for the cluster.
5. Define the cluster topology. To avoid reconfiguring network devices, the cluster IP addresses
should be on the same subnet as the IP addresses of computer 'A', on its proposed cluster
interfaces.
6. Install the Access Control Policy on this cluster object, currently including member 'B' only.
2 VLAN switches
Item Description
5 Router that connects to ISP A (192.168.1.1)
7 Internet
You can configure the ISP preference to be for Load Sharing or Primary/Backup.
Load Sharing - Uses the two links with a distributed load of connections going out from the
Security Gateway. Connections coming in are alternated. You can configure best relative loads for
the links (set a faster link to handle more load). New connections are randomly assigned to a link.
If one link fails, the other takes the load.
Primary/Backup - Uses one link for connections going out from the Security Gateway and coming
in. It switches to the backup if the primary link fails. When the primary link is restored, new
connections are assigned to it. Existing connections continue on the backup link until they are
complete.
Note: ISP Redundancy settings override VPN Link Selection settings.
6. Configure Static NAT to translate the public IP addresses to the real server's IP address.
External clients use one of the two IP addresses.
Note - If the servers use different services (for example, HTTP and FTP), you can use NAT for
only two public IP addresses.
7. Define an Access Control Policy rule: allow DNS traffic through the Security Gateway or
cluster using the domain_udp service.
If you have a public IP address from each ISP for each publicly reachable server (in addition to
the Security Gateway), define NAT rules:
a) Give each server a private IP address.
b) Use the public IP addresses in the Original Destination.
c) Use the private IP address in the Translated Destination.
d) Select Any as the Original Service.
Note - If using Manual NAT, automatic ARP does not work for the NATed IP addresses. You need
to configure the local.arp file as described in sk30197
http://supportcontent.checkpoint.com/solutions?id=sk30197.
When done, install the Access Control Policy on this cluster object.
Failure Recovery
Dynamic Routing on ClusterXL avoids creating a ripple effect upon failover by informing the
neighboring routers that the router has exited a maintenance mode. The neighboring routers then
reestablish their relationships to the cluster, without informing the other routers in the network.
These restart protocols are widely adopted by all major networking vendors. The following table
lists the RFC and drafts compliant with Check Point Dynamic Routing:
Important - We do not recommend that you run these commands. These commands
must be run automatically only by the Security Gateway or the Check Point Support. The
only exception to this rule is to changing the CCP mode, as described below.
Important - You must configure all the cluster members in the same way.
Syntax
Notes:
• In Gaia Clish:
Enter the set cluster<ESC><ESC> to see all the available commands.
• In Expert mode:
Run the cphaconf command see all the available commands.
You can run the cphaconf commands only from the Expert mode.
• Syntax legend:
a) Curly brackets or braces {}:
Enclose a list of available commands or parameters, separated by the vertical bar |, from
which user can enter only one.
b) Angle brackets <>:
Enclose a variable - a supported value user needs to specify explicitly.
c) Square brackets or brackets []:
Enclose an optional command or parameter, which user can also enter.
• You can can include these commands in scripts to run them automatically.
The meaning of each command is explained in the next sections.
Description Command in Command in
of Command Gaia Clish Expert Mode
set cluster member idmode cphaconf mem_id_mode
Configure how to show the id id
cluster member in local name name
ClusterXL logs - by its
Member ID or its Member
Name ("Configuring the
Cluster Member ID Mode
in Local Logs" on page
157)
cphaconf stop
cphaconf clear-secured
cphaconf clear-non-monitored
cphaconf debug_data
Description
You can configure how to show the cluster member in the local ClusterXL logs - by its Member ID
(default) or its Member Name.
This configuration affects these local logs:
• /var/log/messages
• dmesg
• $FWDIR/log/fwd.elg
Note - See Viewing the Cluster Member ID Mode in Local Logs (on page 205).
Syntax
Shell Command
set cluster member idmode
Gaia Clish id
name
cphaconf mem_id_mode
Expert mode id
name
Example
[Expert@Member2:0]# cphaprob names
Current member print mode in local logs is set to: ID
[Expert@Member2:0]#
Description
You can add a user-defined critical device to the default list of critical devices. Use this command
to register <device> as a critical process, and add it to the list of devices that must run for the
cluster member to be considered active. If <device> fails, then the cluster member is seen as
failed.
If a Critical Device fails to report its state to the cluster member in the defined timeout, the
Critical Device, and by design the cluster member, are seen as failed.
Define the status of the Critical Device that is reported to ClusterXL upon registration. This initial
status can be one of these:
• ok - Critical Device is alive.
• init - Critical Device is initializing. The cluster member is Down. In this state, the cluster
member cannot become Active.
• problem - Critical Device failed. If this state is reported to ClusterXL, the cluster member
immediately goes Down. This causes a failover.
Syntax
Shell Command
Gaia Clish N/A
Notes:
• The name of the Critical Device must have no more than 15 characters, and must not include
white spaces.
• For no timeout, use the value 0.
• The -p flag makes these changes permanent. After you reboot the cluster member, the status
of critical devices that were registered with this flag is saved.
• The -g flag applies the command to all configured Virtual Systems.
Restrictions:
• Total number of critical devices (pnotes) on cluster member is limited to 16.
• Name of any critical device (pnote) on cluster member is limited to 16 characters.
Description
Unregistering a user-defined Critical Device (Pnote) means that this device is no longer
considered critical. If a Critical Device was registered with a state problem, before you ran this
command, then after you run this command, the status of the cluster member depends only on
the states of the remaining Critical Devices.
Syntax
Shell Command
Gaia Clish N/A
Notes:
• The -p flag makes these changes permanent. This means that after you reboot, these Critical
Devices remain unregistered.
• The -g flag applies the command to all configured Virtual Systems.
Description
Use this command to report (change) the state of a Critical Device to ClusterXL.
The reported state can be one of these:
• ok - Critical Device is alive.
• init - Critical Device is initializing. The cluster member is Down. In this state, the cluster
member cannot become Active.
• problem - Critical Device failed. If this state is reported to ClusterXL, the cluster member
immediately goes Down. This causes a failover.
If a Critical Device fails to report its state to the cluster member within the defined timeout, the
Critical Device, and by design the cluster member, are seen as failed. This is true only for Critical
Devices with timeouts. If a Critical Device is registered with the -t 0 parameter, there is no
timeout. Until the Critical Device reports otherwise, the state of the Critical Device is considered to
be the last reported state.
Syntax
Shell Command
Gaia Clish N/A
Notes:
• The -g flag applies the command to all configured Virtual Systems.
• If the <Name of Critical Device> reports its state as "problem", then the cluster member
reports its state as failed.
Description
Register all the user-defined Critical Devices listed in the specified file.
This file must be an ASCII file, with each Critical Device defined on a separate line.
Each definition much contain three parameters, which must be separated by a space or a tab
character:
<device> <timeout> <status>
Where:
Parameter Description
<status> The Critical Device <device> reports one of these statuses to the cluster
member:
• ok - Critical Device is alive.
• init - Critical Device is initializing. The cluster member is Down. In this
state, the cluster member cannot become Active.
• problem - Critical Device failed. If this state is reported to ClusterXL, the
cluster member immediately goes Down. This causes a failover.
Syntax
Shell Command
Gaia Clish N/A
Note - The -g flag applies the command to all configured Virtual Systems.
Description
Unregisters all critical devices from the cluster member.
Syntax
Shell Command
Gaia Clish N/A
Notes:
• The -a flag specifies that all Pnotes must be unregistered
• The -g flag applies the command to all configured Virtual Systems
Description
You can configure the Cluster Control Protocol (CCP) mode ("Selecting the CCP Transport Mode
on the Cluster Members" on page 57) on the cluster members
Syntax
Shell Command
set cluster member ccp
Gaia Clish auto
broadcast
multicast
unicast
cphaconf set_ccp
Expert mode auto
unicast
multicast
broadcast
Syntax
Shell Command
set cluster member admin
Gaia Clish down
up
clusterXL_admin
Expert mode down
up
Example
Member2> show cluster state
Member2>
Member2>
Member2>
Syntax
Notes:
• In Gaia Clish:
Enter the show cluster<ESC><ESC> to see all the available commands.
• In Expert mode:
Run the cphaprob command see all the available commands.
You can run the cphaprob commands from Gaia Clish as well.
• Syntax legend:
a) Curly brackets or braces {}:
Enclose a list of available commands or parameters, separated by the vertical bar |, from
which user can enter only one.
b) Angle brackets <>:
Enclose a variable - a supported value user needs to specify explicitly.
c) Square brackets or brackets []:
Enclose an optional command or parameter, which user can also enter.
• You can can include these commands in scripts to run them automatically.
The meaning of each command is explained in the next sections.
Description Command in Command in
of Command Gaia Clish Expert Mode
show cluster state cphaprob [-vs <VSID>]
Show states of cluster state
members and their names
("Monitoring Cluster State"
on page 171)
show cluster members pnotes cphaprob [-l] [-ia] [-e]
Show Critical Devices all list
(Pnotes) and their states problem
on the cluster member
("Monitoring Critical
Devices" on page 176)
members
idmode
igmp
interfaces
all
secured
virtual
vlans
ips
pnotes
all
problem
mmagic
roles
state
statistics
sync [reset]
transport [reset]
Syntax
Shell Command
Gaia Clish 1. set virtual-system <VSID>
2. show cluster state
Example
MEM2> cphaprob state
MEM2>
Field Description
Cluster Mode Can be one of these:
• Load Sharing (Multicast).
• Load Sharing (Unicast).
• High Availability (Primary Up).
• High Availability (Active Up).
• Virtual System Load Sharing
• For third-party clustering products: Service, refer to Clustering
Definitions and Terms, for more information.
Field Description
ID • In the High Availability mode - indicates the cluster member
priority, as configured in the cluster object in SmartConsole.
• In Load Sharing mode - indicates the cluster member ID, as
configured in the cluster object in SmartConsole.
Unique Address Usually, shows the IP addresses of the Sync interfaces.
In some cases, can show IP addresses of other cluster interfaces.
Assigned Load • In the ClusterXL High Availability mode - shows the Active
cluster member with 100% load, and all other Standby cluster
members with 0% load.
• In ClusterXL Load Sharing modes (Unicast and Multicast) -
shows allActive cluster members with 100% load.
State • In the ClusterXL High Availability mode, only one cluster
member in a fully-functioning cluster must be ACTIVE, and the
other cluster members must be in the STANDBY state.
• In the ClusterXL Load Sharing modes (Unicast and Multicast),
all cluster members in a fully-functioning cluster must be
ACTIVE.
• In 3rd-party clustering configuration, all cluster members in a
fully-functioning cluster must be ACTIVE. This is because this
command only reports the status of the Full Synchronization
process.
See the summary table below.
Active PNOTEs Shows the Critical Devices ("Monitoring Critical Devices" on page
176) that report theirs states as "problem".
Last member state change Shows information about the last time this cluster member
event changed its cluster state.
State change Shows the previous cluster state and the new cluster state of this
cluster member.
Reason for state change Shows the reason why this cluster member changed its cluster
state.
Event time Shows the date and the time when this cluster member changed
its cluster state.
Last cluster failover event Shows information about the last time a cluster failover occurred.
Field Description
Transition to new ACTIVE Shows which cluster member became the new Active.
Event time Shows the date and the time of the last cluster failover.
Failover counter Shows the number of cluster failovers since the boot.
Notes:
• This value survives reboot.
• This counter is synchronized between cluster members.
Time of counter reset Shows the date and the time of the last counter reset, and the
reset initiator.
When you examine the state of the cluster member, consider whether it forwards packets, and
whether it has a problem that prevents it from forwarding packets. Each state reflects the result
of a test on critical devices. This table shows the possible cluster states, and whether or not they
represent a problem.
ACTIVE(!) A problem was detected, but the cluster member Yes Yes
still forwards packets, because it is the only member
ACTIVE(!F)
in the cluster, or because there are no other Active
ACTIVE(!P) members in the cluster. In any other situation, the
ACTIVE(!FP) state of the member is Down.
• ACTIVE(!) - See above.
• ACTIVE(!F) - See above. Cluster member is in
the freeze state.
• ACTIVE(!P) - See above. This is the Pivot
cluster member in Load Sharing Unicast mode.
• ACTIVE(!FP) - See above. This is the Pivot
cluster member in Load Sharing Unicast mode
and it is in the freeze state.
DOWN One of the Critical Devices ("Monitoring Critical No Yes
Devices" on page 176) reports its state as
"problem".
Interface Active Monitors the state of All cluster interfaces At least one of the
Check cluster interfaces. on this cluster cluster interfaces on
member are up (CCP this cluster member is
packets are sent and down (CCP packets are
received on all cluster not sent and/or
interfaces). received on time).
Fullsync Monitors if Full Sync This cluster member This cluster member
on this cluster completed Full Sync was not able to
member completed successfully. complete Full Sync.
successfully.
Policy Monitors if the Security This cluster member Security Policy is not
Policy is installed. successfully installed currently installed on
Security Policy. this cluster member.
fwd Monitors the Security fwd daemon on this fwd daemon on this
Gateway process cluster member cluster member did
called fwd. reported its state on not report its state on
time. time.
cvpnd Monitors the Mobile cvpnd daemon on this cvpnd daemon on this
Access back-end cluster member cluster member did
process called cvpnd. reported its state on not report its state on
This pnote appears if time. time.
Mobile Access
Software Blade is
enabled.
ted Monitors the Threat ted daemon on this ted daemon on this
Emulation process cluster member cluster member did
called ted. reported its state on not report its state on
time. time.
a name of a user space User executed the All monitored user At least one of the
process (except fwd, $FWDIR/bin/cluste space processes on monitored user space
routed, cvpnd, ted) rXL_monitor_proce this cluster member on this cluster
ss script. are running. member processes is
See Appendix C - The not running.
clusterXL_monitor_pro
cess Script (on page
224).
Syntax
Shell Command
Gaia Clish show cluster pnotes {all | problem}
Where:
Command Description
show cluster pnotes all Shows cluster full list of Critical Devices
show cluster pnotes problem Prints the list of all the "Built-in Devices" and the
"Registered Devices"
cphaprob -l Prints the list of all the "Built-in Devices" and the
"Registered Devices"
cphaprob -i list When there are no issues on the cluster member, shows:
There are no pnotes in problem state
When a Critical Device reports a problem, prints only the
Critical Device that reports its state as "problem".
cphaprob -ia list When there are no issues on the cluster member, shows:
There are no pnotes in problem state
When a Critical Device reports a problem, prints the
Critical Device "Problem Notification" and the
Critical Device that reports its state as "problem"
Command Description
cphaprob -e list When there are no issues on the cluster member, shows:
There are no pnotes in problem state
When a Critical Device reports a problem, prints only the
Critical Device that reports its state as "problem"
Example
Critical Device fwd reports its state as problem because the fwd process is not up.
[Expert@Member2:0]# cphaprob -l list
Built-in Devices:
Registered Devices:
[Expert@Member2:0]#
Syntax
Shell Command
Gaia Clish 1. set virtual-system <VSID>
2. show cluster members interfaces {all | secured |
virtual | vlans}
Where:
Command Description
show cluster members interfaces all Shows full list of all cluster interfaces:
• including the number of required
interfaces
• including Network Objective
• including VLAN monitoring mode, or
list of monitored VLAN interfaces
show cluster members interfaces secured Shows only cluster interfaces (Cluster and
Sync) and their states:
• without Network Objective
• without VLAN monitoring mode
• without monitored VLAN interfaces
Command Description
show cluster members interfaces virtual Shows full list of cluster virtual interfaces
and their states:
• including the number of required
interfaces
• including Network Objective
• without VLAN monitoring mode
• without monitored VLAN interfaces
show cluster members interfaces vlans Shows only monitored VLAN interfaces
Output
The output of these commands must be identical to the configuration in the cluster object's
Network Management page.
Example
Member2> show cluster members interfaces all
eth3 192.168.151.7
eth4 192.168.1.5
Member2>
Required secured interfaces Shows the total number of the required Sync interfaces.
This number is based on the configuration of the cluster object >
Network Management page.
Non-Monitored This means that cluster member does not monitor the state of
this interface.
In <tp_con, in the cluster object > Network Management page,
administrator configured the Network Type Private for this
interface.
non sync(non secured) This interface role means that this interface does not transfer
Delta Sync packets.
In <tp_con, in the cluster object > Network Management page,
administrator configured the Network Type Cluster for this
interface.
Virtual cluster interfaces Shows the total number of the configure virtual cluster
interfaces.
This number is based on the configuration of the cluster object >
Network Management page.
No VLANs are monitored on Shows the VLAN monitoring mode - there are no VLAN
the member interfaces configured on the cluster interfaces.
For more information, seeVLAN Support in ClusterXL (on page
100).
Monitoring mode is Monitor all Shows the VLAN monitoring mode - there are some VLAN
VLANs: All VLANs are interfaces configured on the cluster interfaces, and cluster
monitored member monitors all VLAN IDs.
For more information, seeVLAN Support in ClusterXL (on page
100).
Monitoring mode is Monitor Shows the VLAN monitoring mode - there are some VLAN
specific VLAN: Only specified interfaces configured on the cluster interfaces, and cluster
VLANs are monitored member monitors only specific VLAN IDs.
For more information, seeVLAN Support in ClusterXL (on page
100).
Syntax
Shell Command
Gaia Clish 1. show cluster bond {all | name <bond_name>}
2. show bonding groups
Where:
Command Description
show cluster bond all Shows configuration of all configured bond
show bonding groups interfaces
cphaprob show_bond
show cluster bond name <bond_name> Shows configuration of the specified bond
interface
cphaprob show_bond <bond_name>
Example 1
[Expert@Member2:0]# cphaprob show_bond
Legend:
-------
UP! - Bond interface state is UP, yet attention is required
Slaves configured - number of slave interfaces configured on the bond
Slaves link up - number of operational slaves
Slaves required - minimal number of operational slaves required for bond to be UP
[Expert@Member2:0]#
Field Description
Bond name Name of the Gaia bonding group.
Example 2
[Expert@Member2:0]# cphaprob show_bond bond1
[Expert@Member2:0]#
Description of the output fields for the "cphaprob show_bond <bond_name>" and
"show cluster bond name <bond_name>" commands:
Field Description
Bond name Name of the Gaia bonding group.
Bond mode Bonding mode of this Gaia bonding group. One of these:
• High Availability
• Load Sharing
Field Description
Bond status Status of the Gaia bonding group. One of these:
• UP - Bond interface is fully operational
• UP! - Bond interface state is UP, yet attention is required
• DOWN - Bond interface failed
Configured slave interfaces Total number of physical slave interfaces configured in this Gaia
bonding group.
In use slave interfaces Number of operational physical slave interfaces in this Gaia
bonding group.
Example 3
[Expert@Member2:0]# cphaprob show_bond_groups
Legend:
---------
Bonds in group - a list of the bonds in the bond group
Required active bonds - number of required active bonds
[Expert@Member2:0]#
Field Description
Group of bonds name Name of the Group of Bonds (on page 118).
Bonds in group Names of the Gaia bond interfaces configured in this Group of
Bonds.
Syntax
Shell Command
Gaia Clish show cluster failover
Example
[Expert@Member2:0]# cphaprob show_failover
[Expert@Member2:0]#
[Expert@Member2:0]# clusterXL_admin down
Setting member to administratively down state ...
Member current state is Down
[Expert@Member2:0]#
[Expert@Member2:0]# cphaprob show_failover
[Expert@Member2:0]#
Syntax
Shell Command
Gaia Clish 1. set virtual-system <VSID>
2. show cluster mmagic
Example 1
[Expert@Member2:0]# cphaprob mmagic
MAC magic: 1
MAC forward magic: 254
[Expert@Member2:0]#
Example 2
[Expert@Member2:0]# cphaprob mmagic
MAC magic: 2
MAC forward magic: 1
[Expert@Member2:0]#
5. Examine the Delta Sync statistics to see if the problem is solved (see "Output example" on
page 194).
6. Solve any identified problem (see "Synchronization Troubleshooting Options" on page 199).
Output example
This section describes and explains the output parameters of the show cluster statistics
sync and cphaprob syncstat commands.
Example output from a cluster member:
Delta Sync Statistics
Sync status: OK
Drops:
Lost updates................................. 0
Lost bulk update events...................... 0
Oversized updates not sent................... 0
Sync at risk:
Sent reject notifications.................... 0
Received reject notifications................ 0
Sent updates:
Total generated sync messages................ 12316
Sent retransmission requests................. 0
Sent retransmission updates.................. 0
Peak fragments per update.................... 1
Received updates:
Total received updates....................... 12
Received retransmission requests............. 0
Timers:
Delta Sync interval (ms)..................... 100
Field Description
Lost updates Shows how many Delta Sync updates this cluster member considers as lost
(based on sequence numbers in CCP packets).
If this counter shows a value greater than 0, this cluster member lost Delta
Sync updates.
Possible mitigation:
Increase the size of the Sending Queue and the size of the Receiving Queue:
• Increase the size of the Sending Queue (see "Increasing the Size of the
Sending Queue" on page 199), if the counter Received reject notification is
increasing.
• Increase the size of the Receiving Queue (see "Increasing the Size of the
Receiving Queue" on page 199), if the counter Received reject notification
is not increasing.
Field Description
Lost bulk update Shows how many times this cluster member missed Delta Sync updates.
events
(bulk update = twice the size of the local receiving queue)
This counter increases when this cluster member receives a Delta Sync
update with a sequence number much greater than expected. This probably
indicates some networking issues that cause massive packet drops.
This counter increases when the amount of missed Delta Sync updates is
more than twice the local Receiving Queue Size.
Possible mitigation:
• If the counter's value is steady, this might indicate a one time
synchronization problem that can be resolved by running manual Full
Sync. See sk37029
http://supportcontent.checkpoint.com/solutions?id=sk37029.
• If the counter's value keeps increasing, probable there are some
networking issues. Increase the sizes of both the Receiving Queue (see
"Increasing the Size of the Receiving Queue" on page 199) and Sending
Queue (see "Increasing the Size of the Sending Queue" on page 199).
Oversized Shows how many oversized Delta Sync updates were discarded before
updates not sent sending them.
This counter increases when Delta Sync update is larger than the local
Fragments Queue Size.
Possible mitigation:
• If the counter's value is steady, increase the size of the Sending Queue
(see "Increasing the Size of the Sending Queue" on page 199).
• If the counter's value keeps increasing, contact Check Point Support
https://www.checkpoint.com/support-services/contact-support/.
Field Description
Sent reject Shows how many times this cluster member rejected Delta Sync
notifications retransmission requests from its peer cluster members, because this cluster
member does not hold the requested Delta Sync update anymore.
Received reject Shows how many reject notifications this cluster member received from its
notification peer cluster members.
Field Description
Total generated Shows how many Delta Sync updates were generated. This counts the Delta
sync messages Sync updates, Retransmission Requests, Retransmission Acknowledgments,
and so on).
Sent Shows how many times this cluster member asked its peer cluster members
retransmission to retransmit specific Delta Sync update(s).
requests
Retransmission requests are sent when certain Delta Sync updates (with a
specified sequence number) are missing, while the sending cluster member
already received Delta Sync updates with advanced sequences.
Note - Compare the number of Sent retransmission requests to the Total
generated sync messages of the other cluster members.
A large counter's value can imply connectivity problems. If the counter's value
is unreasonably high (more than 30% of the Total generated sync messages
of other cluster members), contact Check Point Support
https://www.checkpoint.com/support-services/contact-support/ equipped
with the entire output and a detailed description of the network topology and
configuration.
Sent Shows how many times this cluster member retransmitted specific Delta
retransmission Sync update(s) at the requests from its peer cluster members.
updates
Peak fragments Shows the peak amount of fragments in the Fragments Queue on this cluster
per update member (usually, should be 1).
Field Description
Total received Shows the total number of Delta Sync updates this cluster member received
updates from its peer cluster members.
This counts only Delta Sync updates (not Retransmission Requests,
Retransmission Acknowledgments, and others).
Received Shows how many retransmission requests this cluster member received from
retransmission its peer cluster members.
requests
A large counter's value can imply connectivity problems. If the counter's value
is unreasonably high (more than 30% of the Total generated sync messages
on this cluster member), contact Check Point Support
https://www.checkpoint.com/support-services/contact-support/ equipped
with the entire output and a detailed description of the network topology and
configuration.
Field Description
Sending queue Shows the size of the cyclic queue, which buffers all the Delta Sync updates
size that were already sent until it receives an acknowledgment from the peer
cluster members.
This queue is needed for retransmitting the requested Delta Sync updates.
Each cluster member has one Sending Queue.
Default: 512 Delta Sync updates, which is also the minimal value.
Receiving queue Shows the size of the cyclic queue, which buffers the received Delta Sync
size updates in two cases:
• When Delta Sync updates are missing, this queue is used to hold the
remaining received Delta Sync updates until the lost Delta Sync updates
are retransmitted (cluster members must keep the order, in which they
save the Delta Sync updates in the kernel tables).
• This queue is used to re-assemble a fragmented Delta Sync update.
Each cluster member has one Receiving Queue.
Default: 256 Delta Sync updates, which is also the minimal value.
Fragments Shows the size of the queue, which is used to prepare a Delta Sync update
queue size before moving it to the Sending Queue.
Notes:
• This queue must be smaller than the Sending Queue.
• This queue must be significantly smaller than the Receiving Queue.
Default: 50 Delta Sync updates, which is also the minimal value.
Field Description
Delta Sync Shows the interval at which this cluster member sends the Delta Sync
interval (ms) updates from its Sending Queue.
The base time unit is 100ms (or 1 tick).
Default: 100 ms, which is also the minimum value.
See Increasing the Sync Timer (on page 199).
To make sure that the new Sync Timer interval value survives boot, see How to Configure a Kernel
Parameter to Survive Reboot (see "How to Configure Reboot Survival" on page 127).
By default, fwha_timer_sync_res has a value of 1, meaning that the Sync Timer operates every
base time unit (every 100ms). If you configure the value of this kernel parameter to N, the Sync
Timer operates every N*100ms.
Syntax
Shell Command
Gaia Clish show cluster members igmp
Example
Member2> show cluster members igmp
IGMP Membership: Enabled
Supported Version: 2
Report Interval [sec]: 60
Syntax
Shell Command
Gaia Clish show cluster statistics transport [reset]
The reset flag resets the kernel statistics, which were collected since the last reboot or reset.
Example
Member2> show cluster statistics transport
Operand Calls Bytes Average Ratio %
----------------------------------------------------------
ERROR 0 0 0 0
SET 2035 106444 52 99
RENAME 0 0 0 0
REFRESH 0 0 0 0
DELETE 0 0 0 0
SLINK 1 64 64 0
UNLINK 0 0 0 0
MODIFYFIELDS 0 0 0 0
RECORD DATA CONN 0 0 0 0
COMPLETE DATA CONN 0 0 0 0
Syntax
Shell Command
Gaia Clish show cluster members ips
Example
Member1> show cluster members ips
(Local)
0 1 172.23.88.176
0 2 1.0.0.176
0 3 2.0.0.176
0 4 3.0.0.176
1 2 1.0.0.177
1 3 2.0.0.177
1 4 3.0.0.177
------------------------------------------
Member1>
(Local)
1 1 172.23.88.177
1 2 1.0.0.177
1 3 2.0.0.177
1 4 3.0.0.177
------------------------------------------
Member2>
Syntax
Shell Command
Gaia Clish show cluster members idmode
Example
[Expert@Member2:0]# cphaprob names
[Expert@Member2:0]#
Syntax
Shell Command
Gaia Clish show ospf interfaces [detailed]
Example 1
[Expert@Member2:0]# cphaprob routedifcs
[Expert@Member2:0]#
Example 2
[Expert@Member2:0]# cphaprob routedifcs
eth0
[Expert@Member2:0]#
Syntax
Shell Command
Gaia Clish show cluster role
Example
[Expert@Member2:0]# cphaprob roles
ID Role
1 Non-Master
2 (local) Master
[Expert@Member2:0]#
Syntax
Shell Command
Gaia Clish N/A
cphaprob corr
Expert mode cphaprob -c {a | d |f}
Where:
Command Description
cphaprob corr Shows Cluster Correction Statistics for all traffic.
cphaprob -c d Shows Cluster Correction Statistics for CoreXL SND corrections only.
cphaprob -c f Shows Cluster Correction Statistics for CoreXL Firewall instances and
SND.
Example
[Expert@Member2:0]# cphaprob corr
[Expert@Member2:0]# cphaprob -c a
[Expert@Member2:0]# cphaprob -c d
[Expert@Member2:0]# cphaprob -c f
General logs
Starting <ClusterXL|State Synchronization>.
Indicates that ClusterXL (or State Synchronization, for 3rd party clusters) was successfully started
on the reporting member. This message is usually issued after a member boots, or after an
explicit call to cphastart.
Stopping <ClusterXL|State Synchronization>.
Informs that ClusterXL (or State Synchronization) was deactivated on this member. The member
will no longer be a part of the cluster (even if configured to be so), until ClusterXL is restarted.
Unconfigured cluster Computers changed their MAC Addresses. Please reboot
the cluster so that the changes take affect.
This message is usually issued when a member is shut down, or after an explicit call to
cphastop.
State logs
Mode inconsistency detected: member [ID] ([IP]) will change its mode to
[MODE]. Please re-install the security policy on the cluster.
This message should rarely happen. It indicates that another cluster member has reported a
different cluster mode than is known to the local member. This is usually the result of a failure to
install the Access Control Policy on all cluster members. To correct this problem, install the
Access Control Policy again.
Note - The cluster will continue to operate after a mode inconsistency has been detected, by
altering the mode of the reporting member to match the other cluster members. However, it is
highly recommended that the policy will be re-installed as soon as possible.
State change of member [ID] ([IP]) from [STATE] to [STATE] was cancelled,
since all other members are down. Member remains [STATE].
When a member needs to change its state (for example, when an active member encounters a
problem and needs to bring itself down), it first queries the other members for their state. If all
other members are down, this member cannot change its state to a non-active one (or else all
members will be down, and the cluster will not function). Thus, the reporting cluster member
continues to function, despite its problem (and will usually report its state as "Active Attention").
member [ID] ([IP]) <is active|is down|is stand-by|is initializing>
([REASON]).
This message is issued whenever a cluster member changes its state. The log text specifies the
new state of the member.
Pnote logs
Pnote log messages are issued when a pnote device changes its state.
• [DEVICE] on member [ID] ([IP]) status OK ([REASON]).
The pnote device is working normally.
• [DEVICE] on member [ID] ([IP]) detected a problem ([REASON]).
Either an error was detected by the pnote device, or the device has not reported its state for a
number of seconds (as set by the "timeout" option of the pnote)
• [DEVICE] on member [ID] ([IP]) is initializing ([REASON]).
Indicates that the device has registered itself with the pnote mechanism, but has not yet
determined its state.
• [DEVICE] on member [ID] ([IP]) is in an unknown state ([STATE ID]) ([REASON]).
This message should not normally appear. Contact Check Point Support.
Interface logs
• interface [INTERFACE NAME] of member [ID] ([IP]) is up.
Indicates that this interface is working normally, meaning that it is able to receive and transmit
packets on the expected subnet.
• interface [INTERFACE NAME] of member [ID] ([IP]) is down (receive <up|down>, transmit
<up|down>).
This message is issued whenever an interface encounters a problem, either in receiving or
transmitting packets. Note that in this case the interface may still be working properly, as far
as the OS is concerned, but is unable to communicate with other cluster members due to a
faulty cluster configuration.
• interface [INTERFACE NAME] of member [ID] ([IP]) was added.
Notifies users that a new interface was registered with the Security Gateway (meaning that
packets arriving on this interface are filtered by the firewall). Usually this message is the
result of activating an interface (such as issuing an ifconfig up command on Unix systems).
The interface is now included in the ClusterXL reports (in the output of the Gaia Clish
command show cluster interfaces all, or Expert mode command cphaprob -am if).
Note that the interface may still be reported as "Disconnected", in case it was configured as
such for ClusterXL.
SecureXL logs
• SecureXL device was deactivated since it does not support CPLS.
This message is the result of an attempt to configure a ClusterXL in Load Sharing Multicast
mode over Security Gateways using an acceleration device that does not support Load Sharing.
As a result, acceleration will be turned off, but the cluster will work in Check Point Load
Sharing mode (CPLS).
Reason Strings
• member [ID] ([IP]) reports more interfaces up.
This text can be included in a pnote log message describing the reasons for a problem report:
Another member has more interfaces reported to be working, than the local member does.
This means that the local member has a faulty interface, and that its counterpart can do a
better job as a cluster member. The local member will therefore go down, leaving the member
specified in the message to handle traffic.
• member [ID] ([IP]) has more interfaces - check your disconnected interfaces configuration
in the <discntd.if file|registry>.
This message is issued when members in the same cluster have a different number of
interfaces. A member having less interfaces than the maximal number in the cluster (the
reporting member) may not be working properly, as it is missing an interface required to
operate against a cluster IP address, or a synchronization network. If some of the interfaces on
the other cluster member are redundant, and should not be monitored by ClusterXL, they
should be explicitly designated as "Disconnected". This is done using the file
$FWDIR/conf/discntd.if file.
• [NUMBER] interfaces required, only [NUMBER] up.
ClusterXL has detected a problem with one or more of the monitored interfaces. This does not
necessarily mean that the member will go down, as the other members may have less
operational interfaces. In such a condition, the member with the highest number of operational
interfaces will remain up, while the others will go down.
To see all defined SNMP traps, run threshold_config and select (8) View thresholds overview
from the Threshold Engine Configuration Options menu.
You can download the most recent Check Point MIBs file
http://supportcontent.checkpoint.com/solutions?id=sk90470 from the Support Center.
Method 1 (recommended)
Change in the Command in Gaia Clish Command in Expert Notes
state of the mode
cluster member
Change the state of the set cluster member clusterXL_admin • Does not disable
cluster member to admin down down Delta Sync.
DOWN
• See Appendix A -
The
clusterXL_admin
Script (on page
220).
Change the state of the set cluster member clusterXL_admin • Does not initiate
cluster member to UP admin up up Full Sync.
• See Appendix A -
The
clusterXL_admin
Script (on page
220).
Change the state of the 1. cphaconf set_pnote -d <Name of Does not initiate Full
cluster member to UP Critical Device> -s ok report Sync.
2. cphaconf set_pnote -d <Name of
Critical Device> unregister
Notes:
• In Load Sharing mode, the cluster distributes the traffic load between the remaining Active
members.
• In High Availability mode, the cluster fails over to a Standby cluster member with the highest
priority.
3. Make sure that Firewall rules do not block traffic on TCP port 2010 between the cluster
members.
4. Make sure that the RouteD daemon is running on the Active member.
5. Look for a router-id mismatch in the OSPF configuration.
6. Make sure that the OSPF interface is up on the Standby member.
For advanced troubleshooting procedures and more information, see sk92787
http://supportcontent.checkpoint.com/solutions?id=sk92787.
For troubleshooting OSPF and the RouteD daemon, see sk84520
http://supportcontent.checkpoint.com/solutions?id=sk84520.
Y Shows the ID or the NAME of the local Cluster Member that generated this
log message.
See Configuring the Cluster Member ID Mode in Local Logs (on page 157).
if ( $1 == "up" ) then
echo "Setting member to normal operation ..."
$FWDIR/bin/cphaconf set_pnote -d admin_down $PERSISTENT unregister > & /dev/null
if ( `uname` == 'IPSO' ) then
sleep 5
else
sleep 1
endif
if ( $1 == "down" ) then
echo "Setting member to administratively down state ..."
$FWDIR/bin/cphaconf set_pnote -d admin_down -t 0 -s problem $PERSISTENT register > & /dev/null
sleep 1
Appendix B - The
clusterXL_monitor_ips Script
You can use the clusterXL_monitor_ips script to ping a list of predefined IP addresses and
change the state of the cluster member to Down or Up based on the replies to these pings. For this
script to work, you must write the IP addresses in the $FWDIR/conf/cpha_hosts file - each IP
address on a separate line. This file does not support comments or spaces.
Location of this script on your cluster members is: $FWDIR/bin/clusterXL_monitor_ips
This shell script does these:
1. Registers a Critical Device called host_monitor with status ok.
2. Starts to send pings to the list of predefined IP addresses in the $FWDIR/conf/cpha_hosts
file.
3. While the script receives responses to its pings, it does not change the status of that Critical
Device.
4. If the script does not receive a response to even one ping, it reports the state of that Critical
Device as problem. This gracefully changes the state of the cluster member to Down. If the
script receives responses to its pings again, it changes the status of that Critical Device to ok
again.
For more information, see sk35780 http://supportcontent.checkpoint.com/solutions?id=sk35780.
Important - You must do these changes on all cluster members.
Example:
#!/bin/sh
#
# The script tries to ping the hosts written in the file $FWDIR/conf/cpha_hosts. The names (must be
resolveable) ot the IPs of the hosrs must be written in seperate lines.
# the file must not contain anything else.
# We ping the given hosts every number of seconds given as parameter to the script.
# USAGE:
# cpha_monitor_ips X silent
# where X is the number of seconds between loops over the IPs.
# if silent is set to 1, no messages will appear on the console
#
# We initially register a pnote named "host_monitor" in the problem notification mechanism
# when we detect that a host is not responding we report the pnote to be in "problem" state.
# when ping succeeds again - we report the pnote is OK.
silent=0
if [ -n "$2" ]; then
if [ $2 -le 1 ]; then
silent=$2
fi
fi
hostfile=$FWDIR/conf/cpha_hosts
arch=`uname -s`
if [ $arch = "Linux" ]
then
#system is linux
ping="ping -c 1 -w 1"
else
ping="ping"
fi
$FWDIR/bin/cphaconf set_pnote -d host_monitor -t 0 -s ok register
TRUE=1
while [ "$TRUE" ]
do
result=1
Appendix C - The
clusterXL_monitor_process Script
You can use the clusterXL_monitor_process script to monitor if the specified user space
processes run, and cause cluster fail-over if these processes do not run. For this script to work,
you must write the correct case-sensitive names of the monitored processes in the
$FWDIR/conf/cpha_proc_list file - each process name on a separate line. This file does not
support comments or spaces.
Location of this script on your cluster members is:
$FWDIR/bin/clusterXL_monitor_process
This shell script does these:
1. Registers Critical Devices (with status ok) called as the names of the processes you specified
in the $FWDIR/conf/cpha_proc_list file.
2. While the script detects that the specified process runs, it does not change the status of the
corresponding Critical Device.
3. If the script detects that the specified process do not run anymore, it reports the state of the
corresponding Critical Device as problem. This gracefully changes the state of the cluster
member to Down. If the script detects that the specified process runs again, it changes the
status of the corresponding Critical Device to ok again.
For more information, see sk92904 http://supportcontent.checkpoint.com/solutions?id=sk92904.
Important - You must do these changes on all cluster members.
Example:
#!/bin/sh
#
# This script monitors the existance of processes in the system. The process names should be written
# in the $FWDIR/conf/cpha_proc_list file one every line.
#
# USAGE :
# cpha_monitor_process X silent
# where X is the number of seconds between process probings.
# if silent is set to 1, no messages will appear on the console.
#
#
# We initially register a pnote for each of the monitored processes
# (process name must be up to 15 charachters) in the problem notification mechanism.
# when we detect that a process is missing we report the pnote to be in "problem" state.
# when the process is up again - we report the pnote is OK.
if [ "$2" -le 1 ]
then
silent=$2
else
silent=0
fi
if [ -f $FWDIR/conf/cpha_proc_list ]
then
procfile=$FWDIR/conf/cpha_proc_list
else
echo "No process file in $FWDIR/conf/cpha_proc_list "
exit 0
fi
arch=`uname -s`
done
while [ 1 ]
do
result=1
status=$?
if [ $status = 0 ]
then
if [ $silent = 0 ]
then
echo " $process is alive"
fi
# echo "3, $FWDIR/bin/cphaconf set_pnote -d $process -s ok report"
$FWDIR/bin/cphaconf set_pnote -d $process -s ok report
else
if [ $silent = 0 ]
then
echo " $process is down"
fi
done
if [ $result = 0 ]
then
if [ $silent = 0 ]
then
echo " One of the monitored processes is down!"
fi
else
if [ $silent = 0 ]
then
echo " All monitored processes are up "
fi
fi
if [ "$silent" = 0 ]
then
echo "sleeping"
fi
sleep $1
done