Este contenido no está disponible en el idioma seleccionado.
Cluster Administration
Configuring and Managing a Red Hat Cluster
Edition 5
Abstract
Introduction
- Red Hat Enterprise Linux Installation Guide — Provides information regarding installation of Red Hat Enterprise Linux 5.
- Red Hat Enterprise Linux Deployment Guide — Provides information regarding the deployment, configuration and administration of Red Hat Enterprise Linux 5.
- Red Hat Cluster Suite Overview — Provides a high level overview of the Red Hat Cluster Suite.
- Logical Volume Manager Administration — Provides a description of the Logical Volume Manager (LVM), including information on running LVM in a clustered environment.
- Global File System: Configuration and Administration — Provides information about installing, configuring, and maintaining Red Hat GFS (Red Hat Global File System).
- Global File System 2: Configuration and Administration — Provides information about installing, configuring, and maintaining Red Hat GFS2 (Red Hat Global File System 2).
- Using Device-Mapper Multipath — Provides information about using the Device-Mapper Multipath feature of Red Hat Enterprise Linux 5.
- Using GNBD with Global File System — Provides an overview on using Global Network Block Device (GNBD) with Red Hat GFS.
- Linux Virtual Server Administration — Provides information on configuring high-performance systems and services with the Linux Virtual Server (LVS).
- Red Hat Cluster Suite Release Notes — Provides information about the current release of Red Hat Cluster Suite.
1. Feedback
Cluster_Administration(EN)-5 (2014-6-30T15:52)
Chapter 1. Red Hat Cluster Configuration and Management Overview
1.1. Configuration Basics
Note
- Setting up hardware. Refer to Section 1.1.1, “Setting Up Hardware”.
- Installing Red Hat Cluster software. Refer to Section 1.1.2, “Installing Red Hat Cluster software”.
- Configuring Red Hat Cluster Software. Refer to Section 1.1.3, “Configuring Red Hat Cluster Software”.
1.1.1. Setting Up Hardware
- Cluster nodes — Computers that are capable of running Red Hat Enterprise Linux 5 software, with at least 1GB of RAM. The maximum number of nodes supported in a Red Hat Cluster is 16.
- Ethernet switch or hub for public network — This is required for client access to the cluster.
- Ethernet switch or hub for private network — This is required for communication among the cluster nodes and other cluster hardware such as network power switches and Fibre Channel switches.
- Network power switch — A network power switch is recommended to perform fencing in an enterprise-level cluster.
- Fibre Channel switch — A Fibre Channel switch provides access to Fibre Channel storage. Other options are available for storage according to the type of storage interface; for example, iSCSI or GNBD. A Fibre Channel switch can be configured to perform fencing.
- Storage — Some type of storage is required for a cluster. The type required depends on the purpose of the cluster.
Figure 1.1. Red Hat Cluster Hardware Overview
1.1.2. Installing Red Hat Cluster software
1.1.2.1. Upgrading the Cluster Software
- Shut down all cluster services on a single cluster node. For instructions on stopping cluster software on a node, refer to Section 6.1, “Starting and Stopping the Cluster Software”. It may be desirable to manually relocate cluster-managed services and virtual machines off of the host prior to stopping rgmanager.
- Execute the yum update command to install the new RPMs. For example:
yum update -y openais cman rgmanager lvm2-cluster gfs2-utils
- Reboot the cluster node or restart the cluster services manually. For instructions on starting cluster software on a node, refer to Section 6.1, “Starting and Stopping the Cluster Software”.
1.1.3. Configuring Red Hat Cluster Software
Figure 1.2. Cluster Configuration Structure
- Conga — This is a comprehensive user interface for installing, configuring, and managing Red Hat clusters, computers, and storage attached to clusters and computers.
system-config-cluster
— This is a user interface for configuring and managing a Red Hat cluster.- Command line tools — This is a set of command line tools for configuring and managing a Red Hat cluster.
system-config-cluster
is provided in subsequent chapters of this document. Information about the command line tools is available in the man pages for the tools.
1.2. Conga
- One Web interface for managing cluster and storage
- Automated Deployment of Cluster Data and Supporting Packages
- Easy Integration with Existing Clusters
- No Need to Re-Authenticate
- Integration of Cluster Status and Logs
- Fine-Grained Control over User Permissions
- luci lists clusters that have been set up with that luci. A system administrator can administer all clusters listed on this tab. Other users can administer only clusters that the user has permission to manage (granted by an administrator).— Provides tools for creating and configuring clusters. Each instance of
Figure 1.3. luci Tab
Figure 1.4. luci Tab
Figure 1.5. luci Tab
1.3. system-config-cluster
Cluster Administration GUI
system-config-cluster
. It is for use with the cluster infrastructure and the high-availability service management components. system-config-cluster
consists of two major functions: the Cluster Configuration Tool and the Cluster Status Tool. The Cluster Configuration Tool provides the capability to create, edit, and propagate the cluster configuration file (/etc/cluster/cluster.conf
). The Cluster Status Tool provides the capability to manage high-availability services. The following sections summarize those functions.
Note
system-config-cluster
provides several convenient tools for configuring and managing a Red Hat Cluster, the newer, more comprehensive tool, Conga, provides more convenience and flexibility than system-config-cluster
.
1.3.1. Cluster Configuration Tool
Figure 1.6. Cluster Configuration Tool
/etc/cluster/cluster.conf
) with a hierarchical graphical display in the left panel. A triangle icon to the left of a component name indicates that the component has one or more subordinate components assigned to it. Clicking the triangle icon expands and collapses the portion of the tree below a component. The components displayed in the GUI are summarized as follows:
- Cluster Nodes — Displays cluster nodes. Nodes are represented by name as subordinate elements under Cluster Nodes. Using configuration buttons at the bottom of the right frame (below Properties), you can add nodes, delete nodes, edit node properties, and configure fencing methods for each node.
- Fence Devices — Displays fence devices. Fence devices are represented as subordinate elements under Fence Devices. Using configuration buttons at the bottom of the right frame (below Properties), you can add fence devices, delete fence devices, and edit fence-device properties. Fence devices must be defined before you can configure fencing (with the button) for each node.
- Managed Resources — Displays failover domains, resources, and services.
- Failover Domains — For configuring one or more subsets of cluster nodes used to run a high-availability service in the event of a node failure. Failover domains are represented as subordinate elements under Failover Domains. Using configuration buttons at the bottom of the right frame (below Properties), you can create failover domains (when Failover Domains is selected) or edit failover domain properties (when a failover domain is selected).
- Resources — For configuring shared resources to be used by high-availability services. Shared resources consist of file systems, IP addresses, NFS mounts and exports, and user-created scripts that are available to any high-availability service in the cluster. Resources are represented as subordinate elements under Resources. Using configuration buttons at the bottom of the right frame (below Properties), you can create resources (when Resources is selected) or edit resource properties (when a resource is selected).
Note
The Cluster Configuration Tool provides the capability to configure private resources, also. A private resource is a resource that is configured for use with only one service. You can configure a private resource within a Service component in the GUI. - Services — For creating and configuring high-availability services. A service is configured by assigning resources (shared or private), assigning a failover domain, and defining a recovery policy for the service. Services are represented as subordinate elements under Services. Using configuration buttons at the bottom of the right frame (below Properties), you can create services (when Services is selected) or edit service properties (when a service is selected).
1.3.2. Cluster Status Tool
Figure 1.7. Cluster Status Tool
/etc/cluster/cluster.conf
). You can use the Cluster Status Tool to enable, disable, restart, or relocate a high-availability service.
1.4. Command Line Administration Tools
system-config-cluster
Cluster Administration GUI, command line tools are available for administering the cluster infrastructure and the high-availability service management components. The command line tools are used by the Cluster Administration GUI and init scripts supplied by Red Hat. Table 1.1, “Command Line Tools” summarizes the command line tools.
Command Line Tool | Used With | Purpose |
---|---|---|
ccs_tool — Cluster Configuration System Tool | Cluster Infrastructure | ccs_tool is a program for making online updates to the cluster configuration file. It provides the capability to create and modify cluster infrastructure components (for example, creating a cluster, adding and removing a node). For more information about this tool, refer to the ccs_tool(8) man page. |
cman_tool — Cluster Management Tool | Cluster Infrastructure | cman_tool is a program that manages the CMAN cluster manager. It provides the capability to join a cluster, leave a cluster, kill a node, or change the expected quorum votes of a node in a cluster. For more information about this tool, refer to the cman_tool(8) man page. |
fence_tool — Fence Tool | Cluster Infrastructure | fence_tool is a program used to join or leave the default fence domain. Specifically, it starts the fence daemon (fenced ) to join the domain and kills fenced to leave the domain. For more information about this tool, refer to the fence_tool(8) man page. |
clustat — Cluster Status Utility | High-availability Service Management Components | The clustat command displays the status of the cluster. It shows membership information, quorum view, and the state of all configured user services. For more information about this tool, refer to the clustat(8) man page. |
clusvcadm — Cluster User Service Administration Utility | High-availability Service Management Components | The clusvcadm command allows you to enable, disable, relocate, and restart high-availability services in a cluster. For more information about this tool, refer to the clusvcadm(8) man page. |
Chapter 2. Before Configuring a Red Hat Cluster
Important
2.1. General Configuration Considerations
- Number of cluster nodes supported
- The maximum number of nodes supported in a Red Hat Cluster is 16.
- GFS/GFS2
- Although a GFS/GFS2 file system can be implemented in a standalone system or as part of a cluster configuration, for the RHEL 5.5 release and later, Red Hat does not support the use of GFS/GFS2 as a single-node file system. Red Hat does support a number of high-performance single-node file systems that are optimized for single node, and thus have generally lower overhead than a cluster file system. Red Hat recommends using those file systems in preference to GFS/GFS2 in cases where only a single node needs to mount the file system. Red Hat will continue to support single-node GFS/GFS2 file systems for existing customers.When you configure a GFS/GFS2 file system as a cluster file system, you must ensure that all nodes in the cluster have access to the shared file system. Asymmetric cluster configurations in which some nodes have access to the file system and others do not are not supported.This does not require that all nodes actually mount the GFS/GFS2 file system itself.
- No-single-point-of-failure hardware configuration
- Clusters can include a dual-controller RAID array, multiple bonded network channels, multiple paths between cluster members and storage, and redundant un-interruptible power supply (UPS) systems to ensure that no single failure results in application down time or loss of data.Alternatively, a low-cost cluster can be set up to provide less availability than a no-single-point-of-failure cluster. For example, you can set up a cluster with a single-controller RAID array and only a single Ethernet channel.Certain low-cost alternatives, such as host RAID controllers, software RAID without cluster support, and multi-initiator parallel SCSI configurations are not compatible or appropriate for use as shared cluster storage.
- Data integrity assurance
- To ensure data integrity, only one node can run a cluster service and access cluster-service data at a time. The use of power switches in the cluster hardware configuration enables a node to power-cycle another node before restarting that node's HA services during a failover process. This prevents two nodes from simultaneously accessing the same data and corrupting it. It is strongly recommended that fence devices (hardware or software solutions that remotely power, shutdown, and reboot cluster nodes) are used to guarantee data integrity under all failure conditions. Watchdog timers provide an alternative way to to ensure correct operation of HA service failover.
- Ethernet channel bonding
- Cluster quorum and node health is determined by communication of messages among cluster nodes via Ethernet. In addition, cluster nodes use Ethernet for a variety of other critical cluster functions (for example, fencing). With Ethernet channel bonding, multiple Ethernet interfaces are configured to behave as one, reducing the risk of a single-point-of-failure in the typical switched Ethernet connection among cluster nodes and other cluster hardware.Red Hat Enterprise Linux 5 supports bonding mode 1 only. It is recommended that you wire each node's slaves to the switches in a consistent manner, with each node's primary device wired to switch 1 and each node's backup device wired to switch 2.
2.2. Compatible Hardware
2.3. Enabling IP Ports
2.3.1. Enabling IP Ports on Cluster Nodes
Note
IP Port Number | Protocol | Component |
---|---|---|
5404, 5405 | UDP | cman (Cluster Manager) |
11111 | TCP | ricci (part of Conga remote agent) |
14567 | TCP | gnbd (Global Network Block Device) |
16851 | TCP | modclusterd (part of Conga remote agent) |
21064 | TCP | dlm (Distributed Lock Manager) |
50006, 50008, 50009 | TCP | ccsd (Cluster Configuration System daemon) |
50007 | UDP | ccsd (Cluster Configuration System daemon) |
Note
rgmanager
. For Red Hat Enterprise Linux 5.1 and later, rgmanager
does not use TCP or UDP sockets.
2.3.2. Enabling IP Ports on Computers That Run luci
Note
IP Port Number | Protocol | Component |
---|---|---|
8084 | TCP | luci (Conga user interface server) |
11111 | TCP | ricci (Conga remote agent) |
LUCI_HTTPS_PORT
parameter in the /etc/sysconfig/luci
file as follows:
LUCI_HTTPS_PORT=10.10.10.10:8084
2.4. Configuring ACPI For Use with Integrated Fence Devices
Note
shutdown -h now
). Otherwise, if ACPI Soft-Off is enabled, an integrated fence device can take four or more seconds to turn off a node (refer to note that follows). In addition, if ACPI Soft-Off is enabled and a node panics or freezes during shutdown, an integrated fence device may not be able to turn off the node. Under those circumstances, fencing is delayed or unsuccessful. Consequently, when a node is fenced with an integrated fence device and ACPI Soft-Off is enabled, a cluster recovers slowly or requires administrative intervention to recover.
Note
chkconfig
management and verify that the node turns off immediately when fenced. The preferred way to disable ACPI Soft-Off is with chkconfig
management: however, if that method is not satisfactory for your cluster, you can disable ACPI Soft-Off with one of the following alternate methods:
- Changing the BIOS setting to "instant-off" or an equivalent setting that turns off the node without delay
Note
Disabling ACPI Soft-Off with the BIOS may not be possible with some computers. - Appending
acpi=off
to the kernel boot command line of the/boot/grub/grub.conf
fileImportant
This method completely disables ACPI; some computers do not boot correctly if ACPI is completely disabled. Use this method only if the other methods are not effective for your cluster.
- Section 2.4.1, “Disabling ACPI Soft-Off with
chkconfig
Management” — Preferred method - Section 2.4.2, “Disabling ACPI Soft-Off with the BIOS” — First alternate method
- Section 2.4.3, “Disabling ACPI Completely in the
grub.conf
File” — Second alternate method
2.4.1. Disabling ACPI Soft-Off with chkconfig
Management
chkconfig
management to disable ACPI Soft-Off either by removing the ACPI daemon (acpid
) from chkconfig
management or by turning off acpid
.
Note
chkconfig
management at each cluster node as follows:
- Run either of the following commands:
chkconfig --del acpid
— This command removesacpid
fromchkconfig
management.— OR —chkconfig --level 2345 acpid off
— This command turns offacpid
.
- Reboot the node.
- When the cluster is configured and running, verify that the node turns off immediately when fenced.
Note
You can fence the node with thefence_node
command or Conga.
2.4.2. Disabling ACPI Soft-Off with the BIOS
chkconfig
management (Section 2.4.1, “Disabling ACPI Soft-Off with chkconfig
Management”). However, if the preferred method is not effective for your cluster, follow the procedure in this section.
Note
- Reboot the node and start the
BIOS CMOS Setup Utility
program. - Navigate to themenu (or equivalent power management menu).
- At the Example 2.1, “menu, set the function (or equivalent) to (or the equivalent setting that turns off the node via the power button without delay).
BIOS CMOS Setup Utility
: set to ” shows a menu with set to and set to .Note
The equivalents to, , and may vary among computers. However, the objective of this procedure is to configure the BIOS so that the computer is turned off via the power button without delay. - Exit the
BIOS CMOS Setup Utility
program, saving the BIOS configuration. - When the cluster is configured and running, verify that the node turns off immediately when fenced.
Note
You can fence the node with thefence_node
command or Conga.
Example 2.1. BIOS CMOS Setup Utility
: set to
+---------------------------------------------|-------------------+ | ACPI Function [Enabled] | Item Help | | ACPI Suspend Type [S1(POS)] |-------------------| | x Run VGABIOS if S3 Resume Auto | Menu Level * | | Suspend Mode [Disabled] | | | HDD Power Down [Disabled] | | | Soft-Off by PWR-BTTN [Instant-Off | | | CPU THRM-Throttling [50.0%] | | | Wake-Up by PCI card [Enabled] | | | Power On by Ring [Enabled] | | | Wake Up On LAN [Enabled] | | | x USB KB Wake-Up From S3 Disabled | | | Resume by Alarm [Disabled] | | | x Date(of Month) Alarm 0 | | | x Time(hh:mm:ss) Alarm 0 : 0 : | | | POWER ON Function [BUTTON ONLY | | | x KB Power ON Password Enter | | | x Hot Key Power ON Ctrl-F1 | | | | | | | | +---------------------------------------------|-------------------+
2.4.3. Disabling ACPI Completely in the grub.conf
File
chkconfig
management (Section 2.4.1, “Disabling ACPI Soft-Off with chkconfig
Management”). If the preferred method is not effective for your cluster, you can disable ACPI Soft-Off with the BIOS power management (Section 2.4.2, “Disabling ACPI Soft-Off with the BIOS”). If neither of those methods is effective for your cluster, you can disable ACPI completely by appending acpi=off
to the kernel boot command line in the grub.conf
file.
Important
grub.conf
file of each cluster node as follows:
- Open
/boot/grub/grub.conf
with a text editor. - Append
acpi=off
to the kernel boot command line in/boot/grub/grub.conf
(refer to Example 2.2, “Kernel Boot Command Line withacpi=off
Appended to It”). - Reboot the node.
- When the cluster is configured and running, verify that the node turns off immediately when fenced.
Note
You can fence the node with thefence_node
command or Conga.
Example 2.2. Kernel Boot Command Line with acpi=off
Appended to It
# grub.conf generated by anaconda # # Note that you do not have to rerun grub after making changes to this file # NOTICE: You have a /boot partition. This means that # all kernel and initrd paths are relative to /boot/, eg. # root (hd0,0) # kernel /vmlinuz-version ro root=/dev/VolGroup00/LogVol00 # initrd /initrd-version.img #boot=/dev/hda default=0 timeout=5 serial --unit=0 --speed=115200 terminal --timeout=5 serial console title Red Hat Enterprise Linux Server (2.6.18-36.el5) root (hd0,0) kernel /vmlinuz-2.6.18-36.el5 ro root=/dev/VolGroup00/LogVol00 console=ttyS0,115200n8 acpi=off initrd /initrd-2.6.18-36.el5.img
acpi=off
has been appended to the kernel boot command line — the line starting with "kernel /vmlinuz-2.6.18-36.el5".
2.5. Considerations for Configuring HA Services
rgmanager
, implements cold failover for off-the-shelf applications. In a Red Hat cluster, an application is configured with other cluster resources to form an HA service that can fail over from one cluster node to another with no apparent interruption to cluster clients. HA-service failover can occur if a cluster node fails or if a cluster system administrator moves the service from one cluster node to another (for example, for a planned outage of a cluster node).
- IP address resource — IP address 10.10.10.201.
- An application resource named "httpd-content" — a web server application init script
/etc/init.d/httpd
(specifyinghttpd
). - A file system resource — Red Hat GFS named "gfs-content-webserver".
Figure 2.1. Web Server Cluster Service Example
Note
system-config-cluster
).
/etc/cluster/cluster.conf
(in each cluster
node). In the cluster configuration file, each resource tree is an XML
representation that specifies each resource, its attributes, and its
relationship among other resources in the resource tree (parent,
child, and sibling relationships).Note
- Apache
- Application (Script)
- LVM (HA LVM)
- MySQL
- NFS
- Open LDAP
- Oracle
- PostgreSQL 8
- Samba
Note
Red Hat Enterprise Linux 5 does not support running Clustered Samba in an active/active configuration, in which Samba serves the same shared file system from multiple nodes. Red Hat Enterprise Linux 5 does support running Samba in a cluster in active/passive mode, with failover from one node to the other nodes in a cluster. Note that if failover occurs, locking states are lost and active connections to Samba are severed so that the clients must reconnect. - SAP
- Tomcat 5
- The types of resources needed to create a service
- Parent, child, and sibling relationships among resources
2.6. Configuring max_luns
max_luns
in Red Hat Enterprise Linux 5.
max_luns
(or max_scsi_luns
for 2.4 kernels) in the /etc/modprobe.conf
file of each node. In Red Hat Enterprise Linux 5, cluster nodes detect multiple LUNs without intervention required; it is not necessary to configure max_luns
to detect multiple LUNs.
2.7. Considerations for Using Quorum Disk
qdiskd
, that provides supplemental heuristics to determine node fitness. With heuristics you can determine factors that are important to the operation of the node in the event of a network partition. For example, in a four-node cluster with a 3:1 split, ordinarily, the three nodes automatically "win" because of the three-to-one majority. Under those circumstances, the one node is fenced. With qdiskd
however, you can set up heuristics that allow the one node to win based on access to a critical resource (for example, a critical network path). If your cluster requires additional methods of determining node health, then you should configure qdiskd
to meet those needs.
Note
qdiskd
is not required unless you have special requirements for node health. An example of a special requirement is an "all-but-one" configuration. In an all-but-one configuration, qdiskd
is configured to provide enough quorum votes to maintain quorum even though only one node is working.
Important
qdiskd
parameters for your Red Hat Cluster depend on the site environment and special requirements needed. To understand the use of heuristics and other qdiskd
parameters, refer to the qdisk(5) man page. If you require assistance understanding and using qdiskd
for your site, contact an authorized Red Hat support representative.
qdiskd
, you should take into account the following considerations:
- Cluster node votes
- Each cluster node should have the same number of votes.
- CMAN membership timeout value
- The CMAN membership timeout value (the time a node needs to be unresponsive before CMAN considers that node to be dead, and not a member) should be at least two times that of the
qdiskd
membership timeout value. The reason is because the quorum daemon must detect failed nodes on its own, and can take much longer to do so than CMAN. The default value for CMAN membership timeout is 10 seconds. Other site-specific conditions may affect the relationship between the membership timeout values of CMAN andqdiskd
. For assistance with adjusting the CMAN membership timeout value, contact an authorized Red Hat support representative. - Fencing
- To ensure reliable fencing when using
qdiskd
, use power fencing. While other types of fencing (such as watchdog timers and software-based solutions to reboot a node internally) can be reliable for clusters not configured withqdiskd
, they are not reliable for a cluster configured withqdiskd
. - Maximum nodes
- A cluster configured with
qdiskd
supports a maximum of 16 nodes. The reason for the limit is because of scalability; increasing the node count increases the amount of synchronous I/O contention on the shared quorum disk device. - Quorum disk device
- A quorum disk device should be a shared block device with concurrent read/write access by all nodes in a cluster. The minimum size of the block device is 10 Megabytes. Examples of shared block devices that can be used by
qdiskd
are a multi-port SCSI RAID array, a Fibre Channel RAID SAN, or a RAID-configured iSCSI target. You can create a quorum disk device withmkqdisk
, the Cluster Quorum Disk Utility. For information about using the utility refer to the mkqdisk(8) man page.Note
Using JBOD as a quorum disk is not recommended. A JBOD cannot provide dependable performance and therefore may not allow a node to write to it quickly enough. If a node is unable to write to a quorum disk device quickly enough, the node is falsely evicted from a cluster.
2.8. Red Hat Cluster Suite and SELinux
- Red Hat Enterprise Linux 5.4 and earlier —
disabled
state only. - Red Hat Enterprise Linux 5.5 and later —
enforcing
orpermissive
state with the SELinux policy type set totargeted
(or with thestate
set todisabled
).
Note
fenced_can_network_connect
is persistently set to on
. This allows the fence_xvm
fencing agent to work properly, enabling the system to fence virtual machines.
2.9. Multicast Addresses
Note
Note
2.10. Configuring the iptables Firewall to Allow Cluster Components
iptables
firewall for the various cluster components.
openais
, use the following filtering. Port 5405 is used to receive multicast traffic.
iptables -I INPUT -p udp -m state --state NEW -m multiport --dports 5404,5405 -j ACCEPT
ricci
:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 11111 -j ACCEPT
modcluster
:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 16851 -j ACCEPT
gnbd
:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 14567 -j ACCEPT
luci
:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 8084 -j ACCEPT
DLM
:
iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 21064 -j ACCEPT
ccsd
:
iptables -I INPUT -p udp -m state --state NEW -m multiport --dports 50007 -j ACCEPT iptables -I INPUT -p tcp -m state --state NEW -m multiport --dports 50008 -j ACCEPT
service iptables save ; service iptables restart
rgmanager
does not access the network directly; rgmanager
communication happens by means of openais
network transport. Enabling openais
allows rgmanager
(or any openais
clients) to work automatically.
2.11. Considerations for Using Conga
2.12. Configuring Virtual Machines in a Clustered Environment
rgmanager
tools to start and stop the virtual machines. Using xm
or virsh
to start the machine can result in the virtual machine running in more than one place, which can cause data corruption in the virtual machine.
- Ensure that you are using the
rgmanager 2.0.52-1.el5_4.3
or later package release. - Store the virtual machine configuration files in a non-default location.
xm
or virsh
, as the configuration file will be unknown out of the box to libvirt
or the xm
tool.
path
attribute of a virtual machine resource. Note that the path
attribute is a directory or set of directories separated by the colon ':' character, not a path to a specific file.
Chapter 3. Configuring Red Hat Cluster With Conga
3.1. Configuration Tasks
- Configuring and running the Conga configuration user interface — the luci server. Refer to Section 3.2, “Starting luci and ricci”.
- Creating a cluster. Refer to Section 3.3, “Creating A Cluster”.
- Configuring global cluster properties. Refer to Section 3.4, “Global Cluster Properties”.
- Configuring fence devices. Refer to Section 3.5, “Configuring Fence Devices”.
- Configuring cluster members. Refer to Section 3.6, “Configuring Cluster Members”.
- Creating failover domains. Refer to Section 3.7, “Configuring a Failover Domain”.
- Creating resources. Refer to Section 3.8, “Adding Cluster Resources”.
- Creating cluster services. Refer to Section 3.9, “Adding a Cluster Service to the Cluster”.
- Configuring storage. Refer to Section 3.10, “Configuring Cluster Storage”.
3.2. Starting luci and ricci
- At each node to be administered by Conga, install the ricci agent. For example:
#
yum install ricci
- At each node to be administered by Conga, start ricci. For example:
#
service ricci start
Starting ricci: [ OK ] - Select a computer to host luci and install the luci software on that computer. For example:
#
yum install luci
Note
Typically, a computer in a server cage or a data center hosts luci; however, a cluster computer can host luci. - At the computer running luci, initialize the luci server using the
luci_admin init
command. For example:#
luci_admin init
Initializing the Luci server Creating the 'admin' user Enter password: <Type password and press ENTER.> Confirm password: <Re-type password and press ENTER.> Please wait... The admin password has been successfully set. Generating SSL certificates... Luci server has been successfully initialized Restart the Luci server for changes to take effect eg. service luci restart - Start luci using
service luci restart
. For example:#
service luci restart
Shutting down luci: [ OK ] Starting luci: generating https SSL certificates... done [ OK ] Please, point your web browser to https://nano-01:8084 to access luci - At a Web browser, place the URL of the luci server into the URL address box and click Go (or the equivalent). The URL syntax for the luci server is
https://luci_server_hostname:8084
. The first time you access luci, two SSL certificate dialog boxes are displayed. Upon acknowledging the dialog boxes, your Web browser displays the luci login page.
3.3. Creating A Cluster
- As administrator of luci, select the tab.
- Click.
- At thetext box, enter a cluster name. The cluster name cannot exceed 15 characters. Add the node name and password for each cluster node. Enter the node name for each node in the column; enter the root password for each node in the column. Check the checkbox if clustered storage is required.
- Click. Clicking causes the following actions:
- Cluster software packages to be downloaded onto each cluster node.
- Cluster software to be installed onto each cluster node.
- Cluster configuration file to be created and propagated to each node in the cluster.
- Starting the cluster.
A progress page shows the progress of those actions for each node in the cluster.When the process of creating a new cluster is complete, a page is displayed providing a configuration interface for the newly created cluster.
3.4. Global Cluster Properties
- Thetext box displays the cluster name; it does not accept a cluster name change. You cannot change the cluster name. The only way to change the name of a Red Hat cluster is to create a new cluster configuration with the new name.
- Thevalue is set to
1
by default and is automatically incremented each time you modify your cluster configuration. However, if you need to set it to another value, you can specify it at the text box. - You can enter advanced cluster properties by clicking. Clicking reveals a list of advanced properties. You can click any advanced property for online help about the property.
Enter the values required and clickfor changes to take effect.- Theparameter is the number of seconds the fence daemon (
fenced
) waits before fencing a node (a member of the fence domain) after the node has failed. The default value is0
. Its value may be varied to suit cluster and network performance. - Theparameter is the number of seconds the fence daemon (
fenced
) waits before fencing a node after the node joins the fence domain. The default value is3
. A typical setting for is between 20 and 30 seconds, but can vary according to cluster and network performance.
Enter values required and Clickfor changes to take effect.Note
For more information about fenced(8) man page.and , refer to theNote
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.If you do not specify a multicast address, the Red Hat Cluster software (specifically,cman
, the Cluster Manager) creates one. It forms the upper 16 bits of the multicast address with 239.192 and forms the lower 16 bits based on the cluster ID.Note
The cluster ID is a unique identifier thatcman
generates for each cluster. To view the cluster ID, run thecman_tool status
command on a cluster node.If you do specify a multicast address, you should use the 239.192.x.x series thatcman
uses. Otherwise, using a multicast address outside that range may cause unpredictable results. For example, using 224.0.0.x (which is "All hosts on the network") may not be routed correctly, or even routed at all by some hardware.Note
If you specify a multicast address, make sure that you check the configuration of routers that cluster packets pass through. Some routers may take a long time to learn addresses, seriously impacting cluster performance.- Table 3.1, “Quorum-Disk Parameters” describes the parameters. If you need to use a quorum disk, click , enter quorum disk parameters, click , and restart the cluster for the changes to take effect.tab — This tab provides an interface for configuring these parameters: , , , , , , , , and . The parameter is enabled by default.
Important
Quorum-disk parameters and heuristics depend on the site environment and the special requirements needed. To understand the use of quorum-disk parameters and heuristics, refer to the qdisk(5) man page. If you require assistance understanding and using quorum disk, contact an authorized Red Hat support representative.Note
Clickingon the tab propagates changes to the cluster configuration file (/etc/cluster/cluster.conf
) in each cluster node. However, for the quorum disk to operate, you must restart the cluster (refer to Section 4.1, “Starting, Stopping, and Deleting Clusters”).
Parameter | Description | |||
---|---|---|---|---|
Disables quorum partition. Disables quorum-disk parameters in the | tab.||||
Enables quorum partition. Enables quorum-disk parameters in the | tab.||||
The frequency of read/write cycles, in seconds. | ||||
The number of votes the quorum daemon advertises to CMAN when it has a high enough score. | ||||
The number of cycles a node must miss to be declared dead. | ||||
The minimum score for a node to be considered "alive". If omitted or set to 0, the default function, floor((n+1)/2) , is used, where n is the sum of the heuristics scores. The value must never exceed the sum of the heuristic scores; otherwise, the quorum disk cannot be available. | ||||
The storage device the quorum daemon uses. The device must be the same on all nodes. | ||||
Specifies the quorum disk label created by the mkqdisk utility. If this field contains an entry, the label overrides the field. If this field is used, the quorum daemon reads /proc/partitions and checks for qdisk signatures on every block device found, comparing the label against the specified label. This is useful in configurations where the quorum device name differs among nodes. | ||||
| ||||
Propagates the changes to the cluster configuration file (/etc/cluster/cluster.conf ) in each cluster node. |
3.5. Configuring Fence Devices
Note
- Creating shared fence devices — Refer to Section 3.5.1, “Creating a Shared Fence Device”. The procedures apply only to creating shared fence devices. You can create non-shared (and shared) fence devices while configuring nodes (refer to Section 3.6, “Configuring Cluster Members”).
- Modifying or deleting fence devices — Refer to Section 3.5.2, “Modifying or Deleting a Fence Device”. The procedures apply to both shared and non-shared fence devices.
3.5.1. Creating a Shared Fence Device
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of the fence devices for a cluster and causes the display of menu items for fence device configuration: and .
Note
If this is an initial cluster configuration, no fence devices have been created, and therefore none are displayed. - Click Add a Sharable Fence Device page to be displayed (refer to Figure 3.1, “Fence Device Configuration”).. Clicking causes the
Figure 3.1. Fence Device Configuration
- At the Add a Sharable Fence Device page, click the drop-down box under and select the type of fence device to configure.
- Specify the information in the Fencing Type dialog box according to the type of fence device. Refer to Appendix B, Fence Device Parameters for more information about fence device parameters.
- Click.Clickingcauses a progress page to be displayed temporarily. After the fence device has been added, the detailed cluster properties menu is updated with the fence device under .
3.5.2. Modifying or Deleting a Fence Device
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of the fence devices for a cluster and causes the display of menu items for fence device configuration: and .
- Click. Clicking causes the display of a list of fence devices under .
- Click a fence device in the list. Clicking a fence device in the list causes the display of a Fence Device Form page for the fence device selected from the list.
- Either modify or delete the fence device as follows:
- To modify the fence device, enter changes to the parameters displayed. Refer to Appendix B, Fence Device Parameters for more information about fence device parameters. Click and wait for the configuration to be updated.
- To delete the fence device, clickand wait for the configuration to be updated.
Note
You can create shared fence devices on the node configuration page, also. However, you can only modify or delete a shared fence device viaat the detailed menu for the cluster (below the menu).
3.6. Configuring Cluster Members
3.6.1. Initially Configuring Members
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of an element and a element with a list of the nodes already configured in the cluster.
- Click a link for a node at either the list in the center of the page or in the list in the detailed menu under themenu. Clicking a link for a node causes a page to be displayed for that link showing how that node is configured.
- At the bottom of the page, under, click .
- Select a fence device and provide parameters for the fence device (for example port number).
Note
You can choose from an existing fence device or create a new fence device. - Clickand wait for the change to take effect.
3.6.2. Adding a Member to a Running Cluster
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of an element and a element with a list of the nodes already configured in the cluster. (In addition, a list of the cluster nodes is displayed in the center of the page.)
- Click Add a node to cluster name page.. Clicking causes the display of the
- At that page, enter the node name in thetext box; enter the root password in the text box. Check the checkbox if clustered storage is required. If you want to add more nodes, click and enter node name and password for the each additional node.
- Click. Clicking causes the following actions:
- Cluster software packages to be downloaded onto the added node.
- Cluster software to be installed (or verification that the appropriate software packages are installed) onto the added node.
- Cluster configuration file to be updated and propagated to each node in the cluster — including the added node.
- Joining the added node to cluster.
A progress page shows the progress of those actions for each added node. - When the process of adding a node is complete, a page is displayed providing a configuration interface for the cluster.
- At the detailed menu for the cluster (below themenu), click . Clicking causes the following displays:
- A list of cluster nodes in the center of the page
- Theelement and the element with a list of the nodes configured in the cluster at the detailed menu for the cluster (below the menu)
- Click the link for an added node at either the list in the center of the page or in the list in the detailed menu under themenu. Clicking the link for the added node causes a page to be displayed for that link showing how that node is configured.
- At the bottom of the page, under, click .
- Select a fence device and provide parameters for the fence device (for example port number).
Note
You can choose from an existing fence device or create a new fence device. - Clickand wait for the change to take effect.
3.6.3. Deleting a Member from a Cluster
- Click the link of the node to be deleted. Clicking the link of the node to be deleted causes a page to be displayed for that link showing how that node is configured.
Note
To allow services running on a node to fail over when the node is deleted, skip the next step. - Disable or relocate each service that is running on the node to be deleted:
Note
Repeat this step for each service that needs to be disabled or started on another node.- Under, click the link for a service. Clicking that link cause a configuration page for that service to be displayed.
- On that page, at thedrop-down box, choose to either disable the service are start it on another node and click .
- Upon confirmation that the service has been disabled or started on another node, click the Choose a cluster to administer page to be displayed.tab. Clicking the tab causes the
- At the Choose a cluster to administer page, click the link of the node to be deleted. Clicking the link of the node to be deleted causes a page to be displayed for that link showing how that node is configured.
- On that page, at thedrop-down box, choose and click . When the node is deleted, a page is displayed that lists the nodes in the cluster. Check the list to make sure that the node has been deleted.
3.7. Configuring a Failover Domain
- Unrestricted — Allows you to specify that a subset of members are preferred, but that a cluster service assigned to this domain can run on any available member.
- Restricted — Allows you to restrict the members that can run a particular cluster service. If none of the members in a restricted failover domain are available, the cluster service cannot be started (either manually or by the cluster software).
- Unordered — When a cluster service is assigned to an unordered failover domain, the member on which the cluster service runs is chosen from the available failover domain members with no priority ordering.
- Ordered — Allows you to specify a preference order among the members of a failover domain. The member at the top of the list is the most preferred, followed by the second member in the list, and so on.
- Failback — Allows you to specify whether a service in the failover domain should fail back to the node that it was originally running on before that node failed. Configuring this characteristic is useful in circumstances where a node repeatedly fails and is part of an ordered failover domain. In that circumstance, if a node is the preferred node in a failover domain, it is possible for a service to fail over and fail back repeatedly between the preferred node and another node, causing severe impact on performance.
Note
The failback characteristic is applicable only if ordered failover is configured.
Note
Note
httpd
), which requires you to set up the configuration identically on all members that run the cluster service). Instead of setting up the entire cluster to run the cluster service, you must set up only the members in the restricted failover domain that you associate with the cluster service.
Note
3.7.1. Adding a Failover Domain
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of failover domains with related services and the display of menu items for failover domains: and .
- Click Add a Failover Domain page.. Clicking causes the display of the
- At the Add a Failover Domain page, specify a failover domain name at the text box.
Note
The name should be descriptive enough to distinguish its purpose relative to other names used in your cluster. - To enable setting failover priority of the members in the failover domain, click thecheckbox. With checked, you can set the priority value, , for each node selected as members of the failover domain.
- To restrict failover to members in this failover domain, click the checkbox next to. With checked, services assigned to this failover domain fail over only to nodes in this failover domain.
- To specify that a node does not fail back in this failover domain, click the checkbox next to. With checked, if a service fails over from a preferred node, the service does not fail back to the original node once it has recovered.
- Configure members for this failover domain. Under, click the checkbox for each node that is to be a member of the failover domain. If is checked, set the priority in the text box for each member of the failover domain.
- Click Failover Domain Form page. That page displays the added resource and includes the failover domain in the cluster menu to the left under .. Clicking causes a progress page to be displayed followed by the display of the
- To make additional changes to the failover domain, continue modifications at the Failover Domain Form page and click when you are done.
3.7.2. Modifying a Failover Domain
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of failover domains with related services and the display of menu items for failover domains: and .
- Click. Clicking causes the display of failover domains under at the detailed menu for the cluster (below the menu).
- At the detailed menu for the cluster (below the Failover Domain Form page. At the Failover Domain Form page, you can modify the failover domain name, prioritize failover, restrict failover to this domain, and modify failover domain membership.menu), click the failover domain to modify. Clicking the failover domain causes the display of the
- Modifying failover name — To change the failover domain name, modify the text at thetext box.
Note
The name should be descriptive enough to distinguish its purpose relative to other names used in your cluster. - Failover priority — To enable or disable prioritized failover in this failover domain, click the not checked, setting priority levels is disabled for this failover domain.checkbox. With checked, you can set the priority value, , for each node selected as members of the failover domain. With
- Restricted failover — To enable or disable restricted failover for members in this failover domain, click the checkbox next to not checked, services assigned to this failover domain can fail over to nodes outside this failover domain.. With checked, services assigned to this failover domain fail over only to nodes in this failover domain. With
- Failback — To enable or disable failback in a failover domain, click the checkbox next to. With checked, if a service fails over from a preferred node, the service does not fail back to the original node once it has recovered.
- Modifying failover domain membership — Under, click the checkbox for each node that is to be a member of the failover domain. A checked box for a node means that the node is a member of the failover domain. If is checked, you can adjust the priority in the text box for each member of the failover domain.
- Click Failover Domain Form page. That page displays the added resource and includes the failover domain in the cluster menu to the left under .. Clicking causes a progress page to be displayed followed by the display of the
- To make additional changes to the failover domain, continue modifications at the Failover Domain Form page and click when you are done.
3.8. Adding Cluster Resources
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of resources in the center of the page and causes the display of menu items for resource configuration: and .
- Click Add a Resource page to be displayed.. Clicking causes the
- At the Add a Resource page, click the drop-down box under and select the type of resource to configure. Appendix C, HA Resource Parameters describes resource parameters.
- Click Resources forcluster name page. That page displays the added resource (and other resources).. Clicking causes a progress page to be displayed followed by the display of
3.9. Adding a Cluster Service to the Cluster
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of services in the center of the page and causes the display of menu items for services configuration: , , and .
- To configure any service other than a virtual machine service, Click Add a Service page to be displayed.. Clicking causes the
- On the Add a Service page, at the text box, type the name of the service.
Note
Use a descriptive name that clearly distinguishes the service from other services in the cluster. - Below thetext box, enter the following parameters for this service.
- not checked, the service must be started manually any time the cluster comes up from the stopped state.— When the checkbox is checked, the service is started automatically when a cluster is started and running. If the checkbox is
- Table C.3, “File System”, and the table of GFS resource parameters, Table C.4, “GFS”.— Setting this option will release NFS locks on a file system in a soft attempt to unmount a file system, which may be necessary if your filesystem is exported via NFS and occasionally fails to unmount (either during shutdown or service relocation). You can also enable NFS daemon and lock workarounds for individual file system resources, which will result in a hard attempt to unmount a file system, as described in the table of file system resource parameters,
Run exclusive
. You can override this option by manual start or relocate operations.- — List of cluster members to try in the event that a service fails. For information on configuring a failover domain with Conga, refer to
- In addition, you can specify theand the .
- Add a resource to the service; click only to this service. The process of adding a local resource is the same as adding a global resource described in Section 3.8, “Adding Cluster Resources”. Adding a global resource adds a resource that has been previously added as a global resource (refer to Section 3.8, “Adding Cluster Resources”).. Clicking causes the display of two drop-down boxes: and . Adding a new local resource adds a resource that is available
- At the drop-down box of either Section 3.8, “Adding Cluster Resources”.)or , select the resource to add and configure it according to the options presented. (The options are the same as described in
Note
If you are adding a Samba-service resource, connect a Samba-service resource directly to the service, not to a resource within a service. - If you want to add resources to that resource, click. Clicking causes the display of additional options to local and global resources. You can continue adding children resources to the resource to suit your requirements. To view children resources, click the triangle icon to the left of .
- When you have completed adding resources to the service, and have completed adding children resources to resources, click. Clicking causes a progress page to be displayed followed by a page displaying the added service (and other services).
Note
/sbin/ip addr list
command on a cluster node. The following output shows the /sbin/ip addr list
command executed on a node running a cluster service:
1: lo: <LOOPBACK,UP> mtu 16436 qdisc noqueue link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: <BROADCAST,MULTICAST,UP> mtu 1356 qdisc pfifo_fast qlen 1000 link/ether 00:05:5d:9a:d8:91 brd ff:ff:ff:ff:ff:ff inet 10.11.4.31/22 brd 10.11.7.255 scope global eth0 inet6 fe80::205:5dff:fe9a:d891/64 scope link inet 10.11.4.240/22 scope global secondary eth0 valid_lft forever preferred_lft forever
Note
3.10. Configuring Cluster Storage
- A computer is unreachable via the network.
- A computer has been re-imaged and the luci server admin must re-authenticate with the ricci agent on the computer.
Note
clvmd
) or the High Availability Logical Volume Management agents (HA-LVM). If you are not able to use either the clvmd
daemon or HA-LVM for operational reasons or because you do not have the correct entitlements, you must not use single-instance LVM on the shared disk as this may result in data corruption. If you have any concerns please contact your Red Hat service representative.
Chapter 4. Managing Red Hat Cluster With Conga
4.1. Starting, Stopping, and Deleting Clusters
- Restart a cluster.
- Start a cluster.
- Stop a cluster.
- Delete a cluster.
- At the right of the Choose a cluster to administer page is a drop-down box. By default, the drop-down box is set to . Clicking the drop-down box box reveals all the selections available: , / , and . The actions of each function are summarized as follows:for each cluster listed on the
- Selectingshuts down cluster software in all cluster nodes.Selectingstarts cluster software.
- Select one of the functions and click.
- Clickingcauses a progress page to be displayed. When the action is complete, a page is displayed showing either of the following pages according to the action selected:
- Forand / — Displays a page with the list of nodes for the cluster.
- For Choose a cluster to administer page in the tab, showing a list of clusters.— Displays the
4.2. Managing Cluster Nodes
- Make a node leave or join a cluster.
- Fence a node.
- Reboot a node.
- Delete a node.
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of nodes in the center of the page and causes the display of an element and a element with a list of the nodes already configured in the cluster.
- At the right of each node listed on the page displayed from the preceding step, click thedrop-down box. Clicking drop-down box reveals the following selections: / , , , and . The actions of each function are summarized as follows:
- Selectingshuts down cluster software and makes the node leave the cluster. Making a node leave a cluster prevents the node from automatically joining the cluster when it is rebooted.Selectingstarts cluster software and makes the node join the cluster. Making a node join a cluster allows the node to automatically join the cluster when it is rebooted.
cluster.conf
file from/etc/cluster/
.
- Select one of the functions and click.
- Clickingcauses a progress page to be displayed. When the action is complete, a page is displayed showing the list of nodes for the cluster.
4.3. Managing High-Availability Services
- Configure a service.
- Stop or start a service.
- Restart a service.
- Delete a service
- At the detailed menu for the cluster (below themenu), click . Clicking causes the display of services for the cluster in the center of the page.
- At the right of each service listed on the page, click thedrop-down box. Clicking drop-down box reveals the following selections depending on if the service is running:
- If service is running —, , and .
- If service is not running —, , and .
The actions of each function are summarized as follows:- Section 3.8, “Adding Cluster Resources” and Section 3.9, “Adding a Cluster Service to the Cluster”.) In addition, a drop-down box on the page provides other functions depending on if the service is running.— is available when the service is running or not running. Selecting causes the services configuration page for the service to be displayed. On that page, you can change the configuration of the service. For example, you can add a resource to the service. (For more information about adding resources and services, refer toWhen a service is running, the drop-down box provides the following functions: restarting, disabling, and relocating the service.When a service is not running, the drop-down box on the configuration page provides the following functions: enabling and deleting the service.If you are making configuration changes, save the changes by clicking. Clicking causes a progress page to be displayed. When the change is complete, another page is displayed showing a list of services for the cluster.If you have selected one of the functions in the drop-down box on the configuration page, click. Clicking causes a progress page to be displayed. When the change is complete, another page is displayed showing a list of services for the cluster.
4.4. Backing Up and Restoring the luci Configuration
/var/lib/luci/var/Data.fs
file. This is not the cluster configuration itself, which is stored in the cluster.conf
file. Instead, it contains the list of users and clusters and related properties that luci maintains.
- Execute
service luci stop
. - Execute
luci_admin backup [backup_xml_file_path
].Specifying the backup_xml_file_path is optional. If you do not specify a file path, the backup file will be written to/var/lib/luci/var/luci_backup.xml
. - Execute
service luci start
.
- Execute
service luci stop
. - Execute
luci_admin restore -r backup_xml_file_path
.If you do not specify a backup path argument, the command uses/var/lib/luci/var/luci_backup.xml
.Specifying the-r
option indicates that you will replace all configuration with the configuration specified in the backup file. If you do not specify this option, the default behavior (which you can also specify with the-u
) is to merge any additional configuration information from the XML backup into the current database. - Execute
service luci start
.
Chapter 5. Configuring Red Hat Cluster With system-config-cluster
system-config-cluster
, and consists of the following sections:
Note
system-config-cluster
provides several convenient tools for configuring and managing a Red Hat Cluster, the newer, more comprehensive tool, Conga, provides more convenience and flexibility than system-config-cluster
. You may want to consider using Conga instead (refer to Chapter 3, Configuring Red Hat Cluster With Conga and Chapter 4, Managing Red Hat Cluster With Conga).
5.1. Configuration Tasks
system-config-cluster
consists of the following steps:
- Starting the Cluster Configuration Tool,
system-config-cluster
. Refer to Section 5.2, “Starting the Cluster Configuration Tool”. - Configuring cluster properties. Refer to Section 5.3, “Configuring Cluster Properties”.
- Creating fence devices. Refer to Section 5.4, “Configuring Fence Devices”.
- Creating cluster members. Refer to Section 5.5, “Adding and Deleting Members”.
- Creating failover domains. Refer to Section 5.6, “Configuring a Failover Domain”.
- Creating resources. Refer to Section 5.7, “Adding Cluster Resources”.
- Creating cluster services.
- Propagating the configuration file to the other nodes in the cluster.
- Starting the cluster software. Refer to Section 5.10, “Starting the Cluster Software”.
5.2. Starting the Cluster Configuration Tool
ssh -Y
command and issuing the system-config-cluster
command. For example, to start the Cluster Configuration Tool on cluster node nano-01, do the following:
- Log in to a cluster node and run
system-config-cluster
. For example:$
ssh -Y root@nano-01
. . . #system-config-cluster
- If this is the first time you have started the Cluster Configuration Tool, the program prompts you to either open an existing configuration or create a new one. Click to start a new configuration file (refer to Figure 5.1, “Starting a New Configuration File”).
Figure 5.1. Starting a New Configuration File
Note
The Cluster Configuration Tool, exit, and restart the Red Hat Cluster Suite management GUI (tab for the Red Hat Cluster Suite management GUI is available after you save the configuration file with thesystem-config-cluster
). (The tab displays the status of the cluster service manager, cluster nodes, and resources, and shows statistics concerning cluster service operation. To manage the cluster system further, choose the tab.) - Clicking New Configuration dialog box to be displayed (refer to Figure 5.2, “Creating A New Configuration”). The New Configuration dialog box provides a text box for cluster name and the following checkboxes: and . In most circumstances you only need to configure the cluster name.causes the
Note
Choose the cluster name carefully. The only way to change the name of a Red Hat cluster is to create a new cluster configuration with the new name.Custom Configure Multicast
Red Hat Cluster software chooses a multicast address for cluster management communication among cluster nodes. If you need to use a specific multicast address, click thecheckbox and enter a multicast address in the text boxes.Note
IPV6 is not supported for Cluster Suite in Red Hat Enterprise Linux 5.If you do not specify a multicast address, the Red Hat Cluster software (specifically,cman
, the Cluster Manager) creates one. It forms the upper 16 bits of the multicast address with 239.192 and forms the lower 16 bits based on the cluster ID.Note
The cluster ID is a unique identifier thatcman
generates for each cluster. To view the cluster ID, run thecman_tool status
command on a cluster node.If you do specify a multicast address, you should use the 239.192.x.x series thatcman
uses. Otherwise, using a multicast address outside that range may cause unpredictable results. For example, using 224.0.0.x (which is "All hosts on the network") may not be routed correctly, or even routed at all by some hardware.Note
If you specify a multicast address, make sure that you check the configuration of routers that cluster packets pass through. Some routers may take a long time to learn addresses, seriously impacting cluster performance.Use a Quorum Disk
If you need to use a quorum disk, click the Table 5.1, “Quorum-Disk Parameters” describes the parameters.checkbox and enter quorum disk parameters. The following quorum-disk parameters are available in the dialog box if you enable : , , , , , , and .Important
Quorum-disk parameters and heuristics depend on the site environment and special requirements needed. To understand the use of quorum-disk parameters and heuristics, refer to the qdisk(5) man page. If you require assistance understanding and using quorum disk, contact an authorized Red Hat support representative.Note
It is probable that configuring a quorum disk requires changing quorum-disk parameters after the initial configuration. The Cluster Configuration Tool (system-config-cluster
) provides only the display of quorum-disk parameters after initial configuration. If you need to configure quorum disk, consider using Conga instead; Conga allows modification of quorum disk parameters.Overall:Whilesystem-config-cluster
provides several convenient tools for configuring and managing a Red Hat Cluster, the newer, more comprehensive tool, Conga, provides more convenience and flexibility thansystem-config-cluster
. You may want to consider using Conga instead (refer to Chapter 3, Configuring Red Hat Cluster With Conga and Chapter 4, Managing Red Hat Cluster With Conga).Figure 5.2. Creating A New Configuration
- When you have completed entering the cluster name and other parameters in the New Configuration dialog box, click . Clicking starts the Cluster Configuration Tool, displaying a graphical representation of the configuration (Figure 5.3, “The Cluster Configuration Tool”).
Figure 5.3. The Cluster Configuration Tool
Parameter | Description | |||
---|---|---|---|---|
Enables quorum disk. Enables quorum-disk parameters in the New Configuration dialog box. | ||||
The frequency of read/write cycles, in seconds. | ||||
The number of cycles a node must miss in order to be declared dead. | ||||
The number of votes the quorum daemon advertises to CMAN when it has a high enough score. | ||||
The minimum score for a node to be considered "alive". If omitted or set to 0, the default function, floor((n+1)/2) , is used, where n is the sum of the heuristics scores. The value must never exceed the sum of the heuristic scores; otherwise, the quorum disk cannot be available. | ||||
The storage device the quorum daemon uses. The device must be the same on all nodes. | ||||
Specifies the quorum disk label created by the mkqdisk utility. If this field contains an entry, the label overrides the field. If this field is used, the quorum daemon reads /proc/partitions and checks for qdisk signatures on every block device found, comparing the label against the specified label. This is useful in configurations where the quorum device name differs among nodes. | ||||
|
5.3. Configuring Cluster Properties
- At the left frame, click.
- At the bottom of the right frame (labeled Cluster Properties dialog box to be displayed. The Cluster Properties dialog box presents text boxes for , , and two parameters: and .), click the button. Clicking that button causes a
- (Optional) At the Section 5.2, “Starting the Cluster Configuration Tool”). The cluster alias should be descriptive enough to distinguish it from other clusters and systems on your network (for example,text box, specify a cluster alias for the cluster. The default cluster alias is set to the true cluster name provided when the cluster is set up (refer to
nfs_cluster
orhttpd_cluster
). The cluster alias cannot exceed 15 characters. - (Optional) Thevalue is set to
1
by default and is automatically incremented each time you save your cluster configuration. However, if you need to set it to another value, you can specify it at the text box. - Specify theparameters: and .
- Theparameter is the number of seconds the fence daemon (
fenced
) waits before fencing a node after the node joins the fence domain. The default value is3
. A typical setting for is between 20 and 30 seconds, but can vary according to cluster and network performance. - Theparameter is the number of seconds the fence daemon (
fenced
) waits before fencing a node (a member of the fence domain) after the node has failed.The default value is0
. Its value may be varied to suit cluster and network performance.
Note
For more information about fenced(8) man page.and , refer to the - Save cluster configuration changes by selecting=> .
5.4. Configuring Fence Devices
- Click Fence Device Configuration dialog box to be displayed (refer to Figure 5.4, “Fence Device Configuration”).. At the bottom of the right frame (labeled ), click the button. Clicking causes the
Figure 5.4. Fence Device Configuration
- At the Fence Device Configuration dialog box, click the drop-down box under and select the type of fence device to configure.
- Specify the information in the Fence Device Configuration dialog box according to the type of fence device. Refer to Appendix B, Fence Device Parameters for more information about fence device parameters.
- Click.
- Choose=> to save the changes to the cluster configuration.
5.5. Adding and Deleting Members
5.5.1. Adding a Member to a Cluster
- Click.
- At the bottom of the right frame (labeled Node Properties dialog box to be displayed. The Node Properties dialog box presents text boxes for and (refer to Figure 5.5, “Adding a Member to a New Cluster”).), click the button. Clicking that button causes a
Figure 5.5. Adding a Member to a New Cluster
- At thetext box, specify a node name. The entry can be a name or an IP address of the node on the cluster subnet.
Note
Each node must be on the same subnet as the node from which you are running the Cluster Configuration Tool and must be defined either in DNS or in the/etc/hosts
file of each cluster node.Note
The node on which you are running the Cluster Configuration Tool must be explicitly added as a cluster member; the node is not automatically added to the cluster configuration as a result of running the Cluster Configuration Tool. - Optionally, at thetext box, you can specify a value; however in most configurations you can leave it blank. Leaving the text box blank causes the quorum votes value for that node to be set to the default value of
1
. - Click.
- Configure fencing for the node:
- Click the node that you added in the previous step.
- At the bottom of the right frame (below Fence Configuration dialog box to be displayed.), click . Clicking causes the
- At the Fence Configuration dialog box, bottom of the right frame (below ), click . Clicking causes a fence-level element (for example, , , and so on) to be displayed below the node in the left frame of the Fence Configuration dialog box.
- Click the fence-level element.
- At the bottom of the right frame (below Fence Properties dialog box to be displayed.), click . Clicking causes the
- At the Fence Properties dialog box, click the drop-down box and select the fence device for this node. Also, provide additional information required (for example, and for an APC Power Device).
- At the Fence Properties dialog box, click . Clicking causes a fence device element to be displayed below the fence-level element.
- To create additional fence devices at this fence level, return to step 6d. Otherwise, proceed to the next step.
- To create additional fence levels, return to step 6c. Otherwise, proceed to the next step.
- If you have configured all the fence levels and fence devices for this node, click.
- Choose=> to save the changes to the cluster configuration.
5.5.2. Adding a Member to a Running Cluster
- For clusters with only two nodes —
- For clusters with more than two nodes —
5.5.2.1. Adding a Member to a Running Cluster That Contains Only Two Nodes
- Add the node and configure fencing for it as in
- Clickto propagate the updated configuration to other running nodes in the cluster.
- Use the
scp
command to send the updated/etc/cluster/cluster.conf
file from one of the existing cluster nodes to the new node. - At the Red Hat Cluster Suite management GUI Cluster Status Tool tab, disable each service listed under .
- Stop the cluster software on the two running nodes by running the following commands at each node in this order:
service rgmanager stop
service gfs stop
, if you are using Red Hat GFSservice clvmd stop
, if CLVM has been used to create clustered volumesservice cman stop
- Start cluster software on all cluster nodes (including the added one) by running the following commands in this order:
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesservice gfs start
, if you are using Red Hat GFSservice rgmanager start
- Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services are running as expected.
5.5.2.2. Adding a Member to a Running Cluster That Contains More Than Two Nodes
- Add the node and configure fencing for it as in
- Clickto propagate the updated configuration to other running nodes in the cluster.
- Use the
scp
command to send the updated/etc/cluster/cluster.conf
file from one of the existing cluster nodes to the new node. - Start cluster services on the new node by running the following commands in this order:
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesservice gfs start
, if you are using Red Hat GFSservice rgmanager start
- Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services are running as expected.
5.5.3. Deleting a Member from a Cluster
- At one of the running nodes (not to be removed), run the Red Hat Cluster Suite management GUI. At the Cluster Status Tool tab, under , disable or relocate each service that is running on the node to be deleted.
- Stop the cluster software on the node to be deleted by running the following commands at that node in this order:
service rgmanager stop
service gfs stop
, if you are using Red Hat GFSservice clvmd stop
, if CLVM has been used to create clustered volumesservice cman stop
- At the Cluster Configuration Tool (on one of the running members), delete the member as follows:
- If necessary, click the triangle icon to expand theproperty.
- Select the cluster node to be deleted. At the bottom of the right frame (labeled), click the button.
- Clicking the Figure 5.6, “Confirm Deleting a Member”).button causes a warning dialog box to be displayed requesting confirmation of the deletion (
Figure 5.6. Confirm Deleting a Member
- At that dialog box, clickto confirm deletion.
- Propagate the updated configuration by clicking thebutton. (Propagating the updated configuration automatically saves the configuration.)
- Stop the cluster software on the remaining running nodes by running the following commands at each node in this order:
service rgmanager stop
service gfs stop
, if you are using Red Hat GFSservice clvmd stop
, if CLVM has been used to create clustered volumesservice cman stop
- Start cluster software on all remaining cluster nodes by running the following commands in this order:
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesservice gfs start
, if you are using Red Hat GFSservice rgmanager start
- Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services are running as expected.
5.5.3.1. Removing a Member from a Cluster at the Command-Line
clusvcadm
commmand at a shell prompt.
- To prevent service downtime, any services running on the member to be removed must be relocated to another node on the cluster by running the following command:
clusvcadm -r cluster_service_name -m cluster_node_name
Wherecluster_service_name
is the name of the service to be relocated andcluster_member_name
is the name of the member to which the service will be relocated. - Stop the cluster software on the node to be removed by running the following commands at that node in this order:
service rgmanager stop
service gfs stop
and/orservice gfs2 stop
, if you are usinggfs
,gfs2
or bothumount -a -t gfs
and/orumount -a -t gfs2
, if you are using either (or both) in conjunction withrgmanager
service clvmd stop
, if CLVM has been used to create clustered volumesservice cman stop remove
- To ensure that the removed member does not rejoin the cluster after it reboots, run the following set of commands:
chkconfig cman off chkconfig rgmanager off chkconfig clvmd off chkconfig gfs off chkconfig gfs2 off
5.6. Configuring a Failover Domain
- Unrestricted — Allows you to specify that a subset of members are preferred, but that a cluster service assigned to this domain can run on any available member.
- Restricted — Allows you to restrict the members that can run a particular cluster service. If none of the members in a restricted failover domain are available, the cluster service cannot be started (either manually or by the cluster software).
- Unordered — When a cluster service is assigned to an unordered failover domain, the member on which the cluster service runs is chosen from the available failover domain members with no priority ordering.
- Ordered — Allows you to specify a preference order among the members of a failover domain. The member at the top of the list is the most preferred, followed by the second member in the list, and so on.
Note
Note
httpd
), which requires you to set up the configuration identically on all members that run the cluster service). Instead of setting up the entire cluster to run the cluster service, you must set up only the members in the restricted failover domain that you associate with the cluster service.
Note
5.6.1. Adding a Failover Domain
- At the left frame of the Cluster Configuration Tool, click .
- At the bottom of the right frame (labeled Add Failover Domain dialog box to be displayed.), click the button. Clicking the button causes the
- At the Add Failover Domain dialog box, specify a failover domain name at the text box and click . Clicking causes the Failover Domain Configuration dialog box to be displayed (Figure 5.7, “). : Configuring a Failover Domain”
Note
The name should be descriptive enough to distinguish its purpose relative to other names used in your cluster.Figure 5.7.
: Configuring a Failover Domain - Click thedrop-down box and select the members for this failover domain.
- To restrict failover to members in this failover domain, click (check) thecheckbox. (With checked, services assigned to this failover domain fail over only to nodes in this failover domain.)
- To prioritize the order in which the members in the failover domain assume control of a failed cluster service, follow these steps:
- Click (check) the Figure 5.8, “). Clicking : Adjusting Priority” causes the column to be displayed next to the column.checkbox (
Figure 5.8.
: Adjusting Priority - For each node that requires a priority adjustment, click the node listed in thecolumns and adjust priority by clicking one of the arrows. Priority is indicated by the position in the column and the value in the column. The node priorities are listed highest to lowest, with the highest priority node at the top of the column (having the lowest number).
- Clickto create the domain.
- At the Cluster Configuration Tool, perform one of the following actions depending on whether the configuration is for a new cluster or for one that is operational and running:
- New cluster — If this is a new cluster, choose=> to save the changes to the cluster configuration.
- Running cluster — If this cluster is operational and running, and you want to propagate the change immediately, click thebutton. Clicking automatically saves the configuration change. If you do not want to propagate the change immediately, choose => to save the changes to the cluster configuration.
5.6.2. Removing a Failover Domain
- At the left frame of the Cluster Configuration Tool, click the failover domain that you want to delete (listed under ).
- At the bottom of the right frame (labeled Cluster Configuration Tool.), click the button. Clicking the button causes a warning dialog box do be displayed asking if you want to remove the failover domain. Confirm that the failover domain identified in the warning dialog box is the one you want to delete and click . Clicking causes the failover domain to be removed from the list of failover domains under in the left frame of the
- At the Cluster Configuration Tool, perform one of the following actions depending on whether the configuration is for a new cluster or for one that is operational and running:
- New cluster — If this is a new cluster, choose=> to save the changes to the cluster configuration.
- Running cluster — If this cluster is operational and running, and you want to propagate the change immediately, click thebutton. Clicking automatically saves the configuration change. If you do not want to propagate the change immediately, choose => to save the changes to the cluster configuration.
5.6.3. Removing a Member from a Failover Domain
- At the left frame of the Cluster Configuration Tool, click the failover domain that you want to change (listed under ).
- At the bottom of the right frame (labeled Failover Domain Configuration dialog box to be displayed (Figure 5.7, “). : Configuring a Failover Domain”), click the button. Clicking the button causes the
- At the Failover Domain Configuration dialog box, in the column, click the node name that you want to delete from the failover domain and click the button. Clicking removes the node from the column. Repeat this step for each node that is to be deleted from the failover domain. (Nodes must be deleted one at a time.)
- When finished, click.
- At the Cluster Configuration Tool, perform one of the following actions depending on whether the configuration is for a new cluster or for one that is operational and running:
- New cluster — If this is a new cluster, choose=> to save the changes to the cluster configuration.
- Running cluster — If this cluster is operational and running, and you want to propagate the change immediately, click thebutton. Clicking automatically saves the configuration change. If you do not want to propagate the change immediately, choose => to save the changes to the cluster configuration.
5.7. Adding Cluster Resources
- On the Cluster Configuration Tool, click the button. Clicking the button causes the Resource Configuration dialog box to be displayed.property of the
- At the Resource Configuration dialog box, under , click the drop-down box. At the drop-down box, select a resource to configure. Appendix C, HA Resource Parameters describes resource parameters.
- When finished, click.
- Choose=> to save the change to the
/etc/cluster/cluster.conf
configuration file.
5.8. Adding a Cluster Service to the Cluster
- At the left frame, click.
- At the bottom of the right frame (labeled Add a Service dialog box to be displayed.), click the button. Clicking causes the
- At the Add a Service dialog box, type the name of the service in the text box and click . Clicking causes the Service Management dialog box to be displayed (refer to Figure 5.9, “Adding a Cluster Service”).
Note
Use a descriptive name that clearly distinguishes the service from other services in the cluster.Figure 5.9. Adding a Cluster Service
- If you want to restrict the members on which this cluster service is able to run, choose a failover domain from the Section 5.6, “Configuring a Failover Domain” for instructions on how to configure a failover domain.)drop-down box. (Refer to
- not checked, the service must be started manually any time the cluster comes up from stopped state.checkbox — This is checked by default. If is checked, the service is started automatically when a cluster is started and running. If is
- no other services running on them. For example, for a very busy web server that is clustered for high availability, it would would be advisable to keep that service on a node alone with no other services competing for his resources — that is, checked. On the other hand, services that consume few resources (like NFS and Samba), can run together on the same node without little concern over contention for resources. For those types of services you can leave the unchecked.checkbox — This sets a policy wherein the service only runs on nodes that have
Note
Circumstances that require enablingare rare. Enabling can render a service offline if the node it is running on fails and no other nodes are empty. - Select a recovery policy to specify how the resource manager should recover from a service failure. At the upper right of the Service Management dialog box, there are three options available:
- Click the Section 5.7, “Adding Cluster Resources”.button and choose the a resource listed that you have configured in
Note
If you are adding a Samba-service resource, connect a Samba-service resource directly to the service, not to a resource within a service. That is, at the Service Management dialog box, use either or ; do not use or . - If needed, you may also create a private resource that you can create that becomes a subordinate resource by clicking on the Section 5.7, “Adding Cluster Resources”. The private resource will appear as a child to the shared resource to which you associated with the shared resource. Click the triangle icon next to the shared resource to display any private resources associated.button. The process is the same as creating a shared resource described in
- When finished, click.
- Choose=> to save the changes to the cluster configuration.
Note
/sbin/ip addr list
command on a cluster node. The following output shows the /sbin/ip addr list
command executed on a node running a cluster service:
1: lo: <LOOPBACK,UP> mtu 16436 qdisc noqueue link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: <BROADCAST,MULTICAST,UP> mtu 1356 qdisc pfifo_fast qlen 1000 link/ether 00:05:5d:9a:d8:91 brd ff:ff:ff:ff:ff:ff inet 10.11.4.31/22 brd 10.11.7.255 scope global eth0 inet6 fe80::205:5dff:fe9a:d891/64 scope link inet 10.11.4.240/22 scope global secondary eth0 valid_lft forever preferred_lft forever
5.8.1. Relocating a Service in a Cluster
5.9. Propagating The Configuration File: New Cluster
- Log in to the node where you created the configuration file.
- Using the
scp
command, copy the/etc/cluster/cluster.conf
file to all nodes in the cluster.Note
Propagating the cluster configuration file this way is necessary for the first time a cluster is created. Once a cluster is installed and running, the cluster configuration file is propagated using the Red Hat cluster management GUI Section 6.3, “Modifying the Cluster Configuration”.button. For more information about propagating the cluster configuration using the GUI button, refer to
5.10. Starting the Cluster Software
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesNote
Shared storage for use in Red Hat Cluster Suite requires that you be running the cluster logical volume manager daemon (clvmd
) or the High Availability Logical Volume Management agents (HA-LVM). If you are not able to use either theclvmd
daemon or HA-LVM for operational reasons or because you do not have the correct entitlements, you must not use single-instance LVM on the shared disk as this may result in data corruption. If you have any concerns please contact your Red Hat service representative.service gfs start
, if you are using Red Hat GFSservice rgmanager start
- Start the Red Hat Cluster Suite management GUI. At the Cluster Configuration Tool tab, verify that the configuration is correct. At the Cluster Status Tool tab verify that the nodes and services are running as expected.
Chapter 6. Managing Red Hat Cluster With system-config-cluster
Note
system-config-cluster
provides several convenient tools for configuring and managing a Red Hat Cluster, the newer, more comprehensive tool, Conga, provides more convenience and flexibility than system-config-cluster
. You may want to consider using Conga instead (refer to Chapter 3, Configuring Red Hat Cluster With Conga and Chapter 4, Managing Red Hat Cluster With Conga).
6.1. Starting and Stopping the Cluster Software
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesservice gfs start
, if you are using Red Hat GFSservice rgmanager start
service rgmanager stop
service gfs stop
, if you are using Red Hat GFSservice clvmd stop
, if CLVM has been used to create clustered volumesservice cman stop
6.2. Managing High-Availability Services
Figure 6.1. Cluster Status Tool
Members Status | Description | ||
---|---|---|---|
| |||
The node is unable to participate as a cluster member. The most basic cluster software is not running on the node. |
Services Status | Description |
---|---|
The service resources are configured and available on the cluster system that owns the service. | |
The service has failed on a member and is pending start on another member. | |
The service has been disabled, and does not have an assigned owner. A disabled service is never restarted automatically by the cluster. | |
The service is not running; it is waiting for a member capable of starting the service. A service remains in the stopped state if autostart is disabled. | |
The service has failed to start on the cluster and cannot successfully stop the service. A failed service is never restarted automatically by the cluster. |
6.3. Modifying the Cluster Configuration
/etc/cluster/cluster.conf
), use the Cluster Configuration Tool. For more information about using the Cluster Configuration Tool, refer to Chapter 5, Configuring Red Hat Cluster With system-config-cluster
.
Warning
/etc/cluster/cluster.conf
file without guidance from an authorized Red Hat representative or unless you fully understand the consequences of editing the /etc/cluster/cluster.conf
file manually.
Important
1
. For more information about using the Cluster Configuration Tool, refer to Chapter 5, Configuring Red Hat Cluster With system-config-cluster
.
- Make changes to cluster elements (for example, create a service).
- Propagate the updated configuration file throughout the cluster by clicking.
Note
The Cluster Configuration Tool does not display the button if the cluster is new and has not been started yet, or if the node from which you are running the Cluster Configuration Tool is not a member of the cluster. If the button is not displayed, you can still use the Cluster Configuration Tool; however, you cannot propagate the configuration. You can still save the configuration file. For information about using the Cluster Configuration Tool for a new cluster configuration, refer to Chapter 5, Configuring Red Hat Cluster Withsystem-config-cluster
. - Clicking Warning dialog box to be displayed. Click to save and propagate the configuration.causes a
- Clicking Information dialog box to be displayed, confirming that the current configuration has been propagated to the cluster. Click .causes an
- Click thetab and verify that the changes have been propagated to the cluster members.
6.4. Backing Up and Restoring the Cluster Database
/etc/cluster/cluster.conf.bak.1
, /etc/cluster/cluster.conf.bak.2
, and /etc/cluster/cluster.conf.bak.3
. The backup file /etc/cluster/cluster.conf.bak.1
is the newest backup, /etc/cluster/cluster.conf.bak.2
is the second newest backup, and /etc/cluster/cluster.conf.bak.3
is the third newest backup.
- At the Cluster Configuration Tool tab of the Red Hat Cluster Suite management GUI, click => .
- Clicking system-config-cluster dialog box to be displayed.causes the
- At the system-config-cluster dialog box, select a backup file (for example,
/etc/cluster/cluster.conf.bak.1
). Verify the file selection in the box and click . - Increment the configuration version beyond the current working version number as follows:
- Click=> .
- At the Cluster Properties dialog box, change the value and click .
- Click=> .
- Clicking system-config-cluster dialog box to be displayed.=> causes the
- At the system-config-cluster dialog box, select
/etc/cluster/cluster.conf
and click . (Verify the file selection in the box.) - Clicking Information dialog box to be displayed. At that dialog box, click .causes an
- Propagate the updated configuration file throughout the cluster by clicking.
Note
The Cluster Configuration Tool does not display the button if the cluster is new and has not been started yet, or if the node from which you are running the Cluster Configuration Tool is not a member of the cluster. If the button is not displayed, you can still use the Cluster Configuration Tool; however, you cannot propagate the configuration. You can still save the configuration file. For information about using the Cluster Configuration Tool for a new cluster configuration, refer to Chapter 5, Configuring Red Hat Cluster Withsystem-config-cluster
. - Clicking Warning dialog box to be displayed. Click to propagate the configuration.causes a
- Click thetab and verify that the changes have been propagated to the cluster members.
6.5. Disabling Resources of a Clustered Service for Maintenance
cluster.conf
file to have hierarchical resources (similar to a dependency tree) to disable a resource in a service without disabling other resources within that service.
cluster.conf
file, a service uses a MySQL database and ext3-formatted filesystem resources.
<resources> <mysql config_file="/etc/my.cnf" name="mysql-resource" shutdown_wait="0"/> <fs device="/dev/sdb1" force_fsck="0" force_unmount="1" fsid="9349" fstype="ext3" mountpoint="/opt/db" name="SharedDisk" self_fence="0"/> </resources> <service name="ha-mysql"> <fs ref="SharedDisk"> <mysql ref="mysql-resource"/> </fs> </service>
clusvcadm -Z ha-mysql
rg_test
command:
rg_test test /etc/cluster/cluster.conf stop mysql mysql-resource
rg_test test /etc/cluster/cluster.conf start mysql mysql-resource
clusvcadm -U ha-mysql
Note
rg_test
utility will stop all instances of a resource on a given node, potentially causing undesired results if multiple services on a single node are sharing the same resource. Do not perform these steps on resources that have multiple instances within the cluster.conf
file. In such cases, it is usually necessary to disable the service for maintenance.
6.6. Disabling the Cluster Software
/sbin/chkconfig
command to stop the member from joining the cluster at boot-up as follows:
#chkconfig --level 2345 rgmanager off
#chkconfig --level 2345 gfs off
#chkconfig --level 2345 clvmd off
#chkconfig --level 2345 cman off
#chkconfig --level 2345 rgmanager on
#chkconfig --level 2345 gfs on
#chkconfig --level 2345 clvmd on
#chkconfig --level 2345 cman on
service cman start
service clvmd start
, if CLVM has been used to create clustered volumesservice gfs start
, if you are using Red Hat GFSservice rgmanager start
Appendix A. Example of Setting Up Apache HTTP Server
Note
system-config-cluster
). You can use comparable Conga functions to make an Apache HTTP Server highly available on a Red Hat Cluster.
A.1. Apache HTTP Server Setup Overview
httpd
RPM package on all cluster nodes (or on nodes in the failover domain, if used) and configuring a shared GFS shared resource for the Web content.
# chkconfig --del httpd
httpd
daemon, the cluster infrastructure initializes the service on the active cluster node. This ensures that the corresponding IP address and file system mounts are active on only one cluster node at a time.
httpd
service, a floating IP address must be assigned to the service so that the IP address will transfer from one cluster node to another in the event of failover or service relocation. The cluster infrastructure binds this IP address to the network interface on the cluster system that is currently running the Apache HTTP Server. This IP address ensures that the cluster node running httpd
is transparent to the clients accessing the service.
httpd
service is started and stopped. This prevents the cluster systems from accessing the same data simultaneously, which may result in data corruption. Therefore, do not include the file systems in the /etc/fstab
file.
A.3. Installing and Configuring the Apache HTTP Server
httpd
RPM package. For example:
rpm -Uvh httpd-<version>.<arch>.rpm
- Edit the
/etc/httpd/conf/httpd.conf
configuration file and customize the file according to your configuration. For example:- Specify the directory that contains the HTML files. Also specify this mount point when adding the service to the cluster configuration. It is only required to change this field if the mount point for the web site's content differs from the default setting of
/var/www/html/
. For example:DocumentRoot "/mnt/httpdservice/html"
- Specify a unique IP address to which the service will listen for requests. For example:
Listen 192.168.1.100:80
This IP address then must be configured as a cluster resource for the service using the Cluster Configuration Tool. - If the script directory resides in a non-standard location, specify the directory that contains the CGI programs. For example:
ScriptAlias /cgi-bin/ "/mnt/httpdservice/cgi-bin/"
- Specify the path that was used in the previous step, and set the access permissions to default to that directory. For example:
<Directory /mnt/httpdservice/cgi-bin"> AllowOverride None Options None Order allow,deny Allow from all </Directory>
Additional changes may need to be made to tune the Apache HTTP Server or add module functionality. For information on setting up other options, refer to the Red Hat Enterprise Linux System Administration Guide and the Red Hat Enterprise Linux Reference Guide.
- The standard Apache HTTP Server start script,
/etc/rc.d/init.d/httpd
is also used within the cluster framework to start and stop the Apache HTTP Server on the active cluster node. Accordingly, when configuring the service, specify this script by adding it as a resource in the Cluster Configuration Tool. - Copy the configuration file over to the other nodes of the cluster (or nodes of the failover domain, if configured).
httpd-domain
was created for this service.
- Add the init script for the Apache HTTP Server service.
- Select the Resources Configuration properties dialog box is displayed.tab and click . The
- Selectform the drop down menu.
- Enter ato be associated with the Apache HTTP Server service.
- Specify the path to the Apache HTTP Server init script (for example,
/etc/rc.d/init.d/httpd
) in the field. - Click.
- Add a device for the Apache HTTP Server content files and/or custom scripts.
- Click.
- In the Resource Configuration dialog, select from the drop-down menu.
- Enter thefor the resource (for example,
httpd-content
. - Choosefrom the drop-down menu.
- Enter the mount point in thefield (for example,
/var/www/html/
). - Enter the device special file name in thefield (for example,
/dev/sda3
).
- Add an IP address for the Apache HTTP Server service.
- Click.
- Choosefrom the drop-down menu.
- Enter theto be associated with the Apache HTTP Server service.
- Make sure that thecheckbox is left checked.
- Click.
- Click theproperty.
- Create the Apache HTTP Server service.
- Click Add a Service dialog.. Type a for the service in the
- In the Service Management dialog, select a from the drop-down menu or leave it as .
- Click thebutton. From the available list, choose each resource that you created in the previous steps. Repeat this step until all resources have been added.
- Click.
- Choose=> to save your changes.
Appendix B. Fence Device Parameters
Note
Note
/etc/cluster/cluster.conf
).
Field | Description |
---|---|
Name | A name for the APC device connected to the cluster. |
IP Address | The IP address assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The port. |
Switch (optional) | The switch number for the APC switch that connects to the node when you have multiple daisy-chained switches. |
Use SSH | (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to access the device. |
fence_apc | The fence agent for APC. |
Field | Description |
---|---|
Name | A name for the APC device connected to the cluster into which the fence daemon logs via the SNMP protocol. |
IP Address | The IP address or hostname assigned to the device. |
UDP/TCP Port | The UDP/TCP port to use for connection with the device; the default value is 161. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
SNMP version | The SNMP version to use (1, 2c, 3); the default value is 1. |
SNMP community | The SNMP community string; the default value is private . |
SNMP security level | The SNMP security level (noAuthNoPriv, authNoPriv, authPriv). |
SNMP authentication protocol | The SNMP authentication protocol (MD5, SHA). |
SNMP privacy protocol | The SNMP privacy protocol (DES, AES). |
SNMP privacy protocol password | The SNMP privacy protocol password. |
SNMP privacy protocol script | The script that supplies a password for SNMP privacy protocol. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The port. |
fence_apc_snmp | The fence agent for APC that logs into the SNP device via the SNMP protocol. |
Field | Description |
---|---|
Name | A name for the Brocade device connected to the cluster. |
IP Address | The IP address assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Port | The switch outlet number. |
fence_brocade | The fence agent for Brocade FC switches. |
Field | Description |
---|---|
Name | A name for the Bull PAP system connected to the cluster. |
IP Address | The IP address assigned to the PAP console. |
Login | The login name used to access the PAP console. |
Password | The password used to authenticate the connection to the PAP console. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Domain | Domain of the Bull PAP system to power cycle. |
fence_bullpap | The fence agent for Bull’s NovaScale machines controlled by PAP management consoles. |
Field | Description |
---|---|
Name | A name for the Cisco MDS 9000 series device with SNMP enabled. |
IP address or hostname | The IP address or hostname assigned to the device. |
UDP/TCP port (optional) | The UDP/TCP port to use for connection with the device; the default value is 161. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
SNMP version | The SNMP version to use (1, 2c, 3). |
SNMP community | The SNMP community string. |
SNMP security level | The SNMP security level (noAuthNoPriv, authNoPriv, authPriv). |
SNMP authentication protocol | The SNMP authentication protocol (MD5, SHA). |
SNMP privacy protocol | The SNMP privacy protocol (DES, AES). |
SNMP privacy protocol password | The SNMP privacy protocol password. |
SNMP privacy protocol script | The script that supplies a password for SNMP privacy protocol. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The port. |
fence_cisco_mds | The fence agent for Cisco MDS. |
Field | Description |
---|---|
Name | A name for the Cisco UCS device. |
IP Address | The IP address or hostname assigned to the device. |
IP port (optional) | The TCP port to use to connect to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSL connections | Use SSL connections to communicate with the device. |
Sub-organization | Additional path needed to access suborganization. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | Name of virtual machine. |
fence_cisco_ucs | The fence agent for Cisco UCS. |
Field | Description |
---|---|
Name | The name assigned to the DRAC. |
IP Address | The IP address assigned to the DRAC. |
Login | The login name used to access the DRAC. |
Password | The password used to authenticate the connection to the DRAC. |
Module Name | (optional) The module name for the DRAC when you have multiple DRAC modules. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSH (DRAC5 only) | (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to access the device. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
fence_drac | The fence agent for Dell Remote Access Card (DRAC). |
Field | Description |
---|---|
Name | A name for the BladeFrame device connected to the cluster. |
CServer | The hostname (and optionally the username in the form of username@hostname ) assigned to the device. Refer to the fence_egenera(8) man page for more information. |
ESH Path (optional) | The path to the esh command on the cserver (default is /opt/pan- mgr/bin/esh) |
lpan | The logical process area network (LPAN) of the device. |
pserver | The processing blade (pserver) name of the device. |
fence_egenera | The fence agent for the Egenera BladeFrame. |
Field | Description |
---|---|
Name | A name for the RSB to use as a fence device. |
Hostname | The hostname assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
fence_rsb | The fence agent for Fujitsu-Siemens RSB. |
Field | Description |
---|---|
Name | A name for the GNBD device used to fence the cluster. Note that the GFS server must be accessed via GNBD for cluster node fencing support. |
Servers | The hostname of the server to fence the client from, in either IP address or hostname form. For multiple hostnames, separate each hostname with a whitespace. |
IP Address | The cluster name of the node to be fenced. Refer to the fence_gnbd (8) man page for more information. |
fence_gnbd | The fence agent for GNBD-based GFS clusters. |
Field | Description |
---|---|
Name | A name for the server with HP iLO support. |
Hostname | The hostname assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSL connections | Use SSL connections to communicate with the device. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
fence_ilo | The fence agent for HP servers with the Integrated Light Out (iLO) PCI card. |
Field | Description |
---|---|
Name | A name for the server with HP iLO support. |
Hostname | The hostname assigned to the device. |
IP port (optional) | TCP port to use for connection with the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSH | (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to access the device. |
Path to SSH identity file | The identity file for SSH. |
Force command prompt | The command prompt to use. The default value is ’MP>’, ’hpiLO->’. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
fence_ilo_mp | The fence agent for HP iLO MP devices. |
Field | Description |
---|---|
Name | A name for the IBM BladeCenter device connected to the cluster. |
IP Address | The IP address assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Blade | The blade of the device. |
Use SSH | (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to access the device. |
fence_bladecenter | The fence agent for IBM BladeCenter. |
Field | Description |
---|---|
Name | A name for the IBM iPDU device connected to the cluster into which the fence daemon logs via the SNMP protocol. |
IP Address | The IP address or hostname assigned to the device. |
UDP/TCP Port | The UDP/TCP port to use for connection with the device; the default value is 161. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
SNMP version | The SNMP version to use (1, 2c, 3); the default value is 1. |
SNMP community | The SNMP community string; the default value is private . |
SNMP security level | The SNMP security level (noAuthNoPriv, authNoPriv, authPriv). |
SNMP authentication protocol | The SNMP authentication protocol (MD5, SHA). |
SNMP privacy protocol | The SNMP privacy protocol (DES, AES). |
SNMP privacy protocol password | The SNMP privacy protocol password. |
SNMP privacy protocol script | The script that supplies a password for SNMP privacy protocol. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The port. |
fence_ipdu | The fence agent for iPDU over SNMP. |
Field | Description |
---|---|
Name | A name for the RSA device connected to the cluster. |
Hostname | The hostname assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
fence_rsa | The fence agent for the IBM RSA II management interface. |
Field | Description |
---|---|
Name | A name for the IF MIB device connected to the cluster. |
IP address or hostname | The IP address or hostname assigned to the device. |
UDP/TCP port (optional) | The UDP/TCP port to use for connection with the device; the default value is 161. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
SNMP version | The SNMP version to use (1, 2c, 3); the default value is 1. |
SNMP community | The SNMP community string. |
SNMP security level | The SNMP security level (noAuthNoPriv, authNoPriv, authPriv). |
SNMP authentication protocol | The SNMP authentication protocol (MD5, SHA). |
SNMP privacy protocol | The SNMP privacy protocol (DES, AES). |
SNMP privacy protocol password | The SNMP privacy protocol password. |
SNMP privacy protocol script | The script that supplies a password for SNMP privacy protocol. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | Physical plug number or name of virtual machine. |
fence_ifmib | The fence agent for IF-MIB devices. |
Field | Description |
---|---|
Name | A name for the IPMI LAN device connected to the cluster. |
IP Address | The IP address assigned to the IPMI port. |
Login | The login name of a user capable of issuing power on/off commands to the given IPMI port. |
Password | The password used to authenticate the connection to the IPMI port. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Authentication Type | none , password , md2 , or md5 . |
Privilege Level | The privilege level on the IPMI device. |
Use Lanplus | True or 1 . If blank, then value is False . |
fence_ipmilan | The fence agent for machines controlled by IPMI. |
Field | Description |
---|---|
Name | A name to assign the Manual fencing agent. Refer to the fence_manual (8) man page for more information. |
Warning
Field | Description |
---|---|
Name | A name for the McData device connected to the cluster. |
IP Address | The IP address assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Port | The switch outlet number. |
fence_mcdata | The fence agent for McData FC switches. |
Field | Description |
---|---|
Name | A name for the SANBox2 device connected to the cluster. |
IP Address | The IP address assigned to the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The switch outlet number. |
fence_sanbox2 | The fence agent for QLogic SANBox2 FC switches. |
Field | Description |
---|---|
Name | Name of the RHEV-M REST API fencing device. |
IP Address | The IP address or hostname assigned to the device. |
IP port (optional) | The TCP port to use for connection with the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSL connections | Use SSL connections to communicate with the device. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | Physical plug number or name of virtual machine. |
fence_rhevm | The fence agent for RHEV-M REST API. |
Field | Description |
---|---|
Name | A name for the WTI RPS-10 power switch connected to the cluster. |
Device Name | The device name of the device the switch is connected to on the controlling host (for example, /dev/ttys2 ). |
Port | The switch outlet number. |
fence_wti | The fence agent for the WTI Network Power Switch. |
Field | Description |
---|---|
Name | A name for the SCSI fence device. |
Node name | Name of the node to be fenced. Refer to the fence_scsi (8) man page for more information. |
fence_scsi | The fence agent for SCSI persistent reservations. |
Note
- As of Red Hat Enterprise Linux 5.5 and fully-updated releases of Red Hat Enterprise Linux 5.4, SCSI fencing can be used in a 2-node cluster; previous releases did not support this feature.
- When using SCSI fencing, all nodes in the cluster must register with the same devices so that each node can remove another node's registration key from all the devices it is registered with.
- Devices used for the cluster volumes should be a complete LUN, not partitions. SCSI persistent reservations work on an entire LUN, meaning that access is controlled to each LUN, not individual partitions.
- As of Red Hat Enterprise Linux 5.5 and fully-updated releases of Red Hat Enterprise Linux 5.4, SCSI fencing can be used in conjunction with qdisk; previous releases did not support this feature. You cannot use
fence_scsi
on the LUN whereqdiskd
resides; it must be a raw LUN or raw partition of a LUN.
Field | Description |
---|---|
Name | Name of the virtual machine fencing device. |
Domain | Unique domain name of the guest to be fenced. |
Field | Description |
---|---|
Name | Name of the virtual machine fencing device. |
Hostname | The IP address or hostname assigned to the device. |
IP port (optional) | The TCP port to use for connection with the device. |
Login | The login name used to access the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Use SSL connections | Use SSL connections to communicate with the device. |
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Virtual machine name | Name of virtual machine in inventory path format (e.g., /datacenter/vm/Discovered_virtual_machine/myMachine). |
Virtual machine UUID | The UUID of the virtual machine to fence. |
fence_vmware_soap | The fence agent for VMWare over SOAP API. |
Field | Description |
---|---|
Name | A name for the Vixel switch connected to the cluster. |
IP Address | The IP address assigned to the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Port | The switch outlet number. |
fence_vixel | The fence agent for Vixel switches. |
Field | Description |
---|---|
Name | A name for the WTI power switch connected to the cluster. |
IP Address | The IP address assigned to the device. |
Password | The password used to authenticate the connection to the device. |
Password Script (optional) | The script that supplies a password for access to the fence device. Using this supersedes the | parameter.
Power wait | Number of seconds to wait after issuing a power off or power on command. |
Port | The switch outlet number. |
Use SSH | (Red Hat Enterprise Linux 5.4 and later) Indicates that system will use SSH to access the device. |
fence_wti | The fence agent for the WTI network power switch. |
Appendix C. HA Resource Parameters
system-config-cluster
, or by editing etc/cluster/cluster.conf
. Table C.1, “HA Resource Summary” lists the resources, their corresponding resource agents, and references to other tables containing parameter descriptions. To understand resource agents in more detail you can view them in /usr/share/cluster
of any cluster node.
cluster.conf
elements and attributes, refer to the cluster schema at /usr/share/system-config-cluster/misc/cluster.ng
, and the annotated schema at /usr/share/doc/system-config-cluster-X.Y.ZZ/cluster_conf.html
(for example /usr/share/doc/system-config-cluster-1.0.57/cluster_conf.html
).
Resource | Resource Agent | Reference to Parameter Description |
---|---|---|
Apache | apache.sh | Table C.2, “Apache Server” |
File System | fs.sh | Table C.3, “File System” |
GFS File System | clusterfs.sh | Table C.4, “GFS” |
IP Address | ip.sh | Table C.5, “IP Address” |
LVM | lvm.sh | Table C.6, “LVM” |
MySQL | mysql.sh | Table C.7, “MySQL” |
NFS Client | nfsclient.sh | Table C.8, “NFS Client” |
NFS Export | nfsexport.sh | Table C.9, “NFS Export” |
NFS Mount | netfs.sh | Table C.10, “NFS Mount” |
Open LDAP | openldap.sh | Table C.11, “Open LDAP” |
Oracle 10g Failover Instance | oracledb.sh | Table C.12, “Oracle 10g Failover Instance” |
Oracle DB Agent | orainstance.sh | Table C.13, “Oracle DB” |
Oracle Listener Agent | oralistener.sh | Table C.14, “Oracle Listener Agent” |
PostgreSQL 8 | postgres-8.sh | Table C.15, “PostgreSQL 8” |
SAP Database | SAPDatabase | Table C.16, “SAP Database” |
SAP Instance | SAPInstance | Table C.17, “SAP Instance” |
Samba | smb.sh | Table C.18, “Samba Service” |
Script | script.sh | Table C.19, “Script” |
Service | service.sh | Table C.20, “Service” |
Sybase ASE | ASEHAagent.sh | Table C.21, “Sybase ASE Failover Instance” |
Tomcat 5 | tomcat-5.sh | Table C.22, “Tomcat 5” |
Virtual Machine | vm.sh | Table C.23, “Virtual Machine”
NOTE: Luci displays this as a virtual service if the host cluster can support virtual machines.
|
Field | Description |
---|---|
Name | The name of the Apache Service. |
Server Root | The default value is /etc/httpd . |
Config File | Specifies the Apache configuration file. The default valuer is /etc/httpd/conf . |
httpd Options | Other command line options for httpd . |
Shutdown Wait (seconds) | Specifies the number of seconds to wait for correct end of service shutdown. |
Field | Description |
---|---|
Name | Specifies a name for the file system resource. |
File system type | If not specified, mount tries to determine the file system type. |
Mount point | Path in file system hierarchy to mount this file system. |
Device | Specifies the device associated with the file system resource. This can be a block device, file system label, or UUID of a file system. |
Options | Mount options; that is, options used when the file system is mounted. These may be file-system specific. Refer to the mount (8) man page for supported mount options. |
File system ID | Note File System ID is used only by NFS services.
When creating a new file system resource, you can leave this field blank. Leaving the field blank causes a file system ID to be assigned automatically after you commit the parameter during configuration. If you need to assign a file system ID explicitly, specify it in this field.
|
Force unmount | If enabled, forces the file system to unmount. The default setting is disabled . Force Unmount kills all processes using the mount point to free up the mount when it tries to unmount. |
Reboot host node if unmount fails | If enabled, reboots the node if unmounting this file system fails. The default setting is disabled . |
Check file system before mounting | If enabled, causes fsck to be run on the file system before mounting it. The default setting is disabled . |
Enable NFS daemon and lockd workaround | If your filesystem is exported via NFS and occasionally fails to unmount (either during shutdown or service relocation), setting this option will drop all filesystem references prior to the unmount operation. Setting this option requires that you enable the Section 3.9, “Adding a Cluster Service to the Cluster”. | option. You should set this option as a last resort only, as this is a hard attempt to unmount a file system. You can enable NFS lock workarounds in a soft attempt to unmount a file system at the level of cluster service configuration, as described in
Field | Description |
---|---|
Name | The name of the file system resource. |
Mount point | The path to which the file system resource is mounted. |
Device | The device file associated with the file system resource. |
File system type | Specify GFS or GFS2. |
Options | Mount options. |
File system ID | Note File System ID is used only by NFS services.
When creating a new GFS resource, you can leave this field blank. Leaving the field blank causes a file system ID to be assigned automatically after you commit the parameter during configuration. If you need to assign a file system ID explicitly, specify it in this field.
|
Force unmount | If enabled, forces the file system to unmount. The default setting is disabled . Force Unmount kills all processes using the mount point to free up the mount when it tries to unmount. With GFS resources, the mount point is not unmounted at service tear-down unless Force Unmount is enabled. |
Reboot host node if unmount fails | If enabled and unmounting the file system fails, the node will immediately reboot. Generally, this is used in conjunction with force-unmount support, but it is not required. |
Enable NFS daemon and lockd workaround | If your filesystem is exported via NFS and occasionally fails to unmount (either during shutdown or service relocation), setting this option will drop all filesystem references prior to the unmount operation. Setting this option requires that you enable the Section 3.9, “Adding a Cluster Service to the Cluster”. | option. You should set this option as a last resort only, as this is a hard attempt to unmount a file system. You can enable NFS lock workarounds in a soft attempt to unmount a file system at the level of cluster service configuration, as described in
Field | Description |
---|---|
IP address | The IP address for the resource. This is a virtual IP address. IPv4 and IPv6 addresses are supported, as is NIC link monitoring for each IP address. |
Monitor link | Enabling this causes the status check to fail if the link on the NIC to which this IP address is bound is not present. |
Field | Description |
---|---|
Name | A unique name for this LVM resource. |
Volume Group Name | A descriptive name of the volume group being managed. |
Logical Volume Name | Name of the logical volume being managed. This parameter is optional if there is more than one logical volume in the volume group being managed. |
Fence the node if it is unable to clean up LVM tags | Fence the node if it is unable to clean up LVM tags. |
Field | Description |
---|---|
Name | Specifies a name of the MySQL server resource. |
Config File | Specifies the configuration file. The default value is /etc/my.cnf . |
Listen Address | Specifies an IP address for MySQL server. If an IP address is not provided, the first IP address from the service is taken. |
mysqld Options | Other command line options for httpd . |
Shutdown Wait (seconds) | Specifies the number of seconds to wait for correct end of service shutdown. |
Field | Description |
---|---|
Name | This is a symbolic name of a client used to reference it in the resource tree. This is not the same thing as the Target option. |
Target | This is the server from which you are mounting. It can be specified using a hostname, a wildcard (IP address or hostname based), or a netgroup defining a host or hosts to export to. |
Options | Defines a list of options for this client — for example, additional client access rights. For more information, refer to the exports (5) man page, General Options. |
Allow Recover | Allow recovery of the NFS client. |
Field | Description |
---|---|
Name |
Descriptive name of the resource. The NFS Export resource ensures that NFS daemons are running. It is fully reusable; typically, only one NFS Export resource is needed.
Note
Name the NFS Export resource so it is clearly distinguished from other NFS resources.
|
Field | Description |
---|---|
Name |
Symbolic name for the NFS mount.
Note
This resource is required only when a cluster service is configured to be an NFS client.
|
Mount point | Path to which the file system resource is mounted. |
Host | NFS server IP address or hostname. |
Export path | NFS Export directory name. |
NFS version |
NFS protocol:
|
Options | Mount options. Specifies a list of mount options. If none are specified, the NFS file system is mounted -o sync . For more information, refer to the nfs (5) man page. |
Force unmount | If Force unmount is enabled, the cluster kills all processes using this file system when the service is stopped. Killing all processes using the file system frees up the file system. Otherwise, the unmount will fail, and the service will be restarted. |
Field | Description |
---|---|
Name | Specifies a service name for logging and other purposes. |
Config File | Specifies an absolute path to a configuration file. The default value is /etc/openldap/slapd.conf . |
URL List | The default value is ldap:/// . |
slapd Options | Other command line options for slapd . |
Shutdown Wait (seconds) | Specifies the number of seconds to wait for correct end of service shutdown. |
Field | Description |
---|---|
Instance name (SID) of Oracle instance | Instance name. |
Oracle user name | This is the user name of the Oracle user that the Oracle AS instance runs as. |
Oracle application home directory | This is the Oracle (application, not user) home directory. It is configured when you install Oracle. |
Virtual hostname (optional) | Virtual Hostname matching the installation hostname of Oracle 10g. Note that during the start/stop of an oracledb resource, your hostname is changed temporarily to this hostname. Therefore, you should configure an oracledb resource as part of an exclusive service only. |
Field | Description |
---|---|
Instance name (SID) of Oracle instance | Instance name. |
Oracle user name | This is the user name of the Oracle user that the Oracle instance runs as. |
Oracle application home directory | This is the Oracle (application, not user) home directory. It is configured when you install Oracle. |
List of Oracle listeners (optional, separated by spaces) | List of Oracle listeners which will be started with the database instance. Listener names are separated by whitespace. Defaults to empty which disables listeners. |
Path to lock file (optional) | Location for lockfile which will be used for checking if the Oracle should be running or not. Defaults to location under /tmp . |
Field | Description |
---|---|
Listener Name | Listener name. |
Oracle user name | This is the user name of the Oracle user that the Oracle instance runs as. |
Oracle application home directory | This is the Oracle (application, not user) home directory. It is configured when you install Oracle. |
Field | Description |
---|---|
Name | Specifies a service name for logging and other purposes. |
Config File | Define absolute path to configuration file. The default value is /var/lib/pgsql/data/postgresql.conf . |
Postmaster User | User who runs the database server because it cannot be run by root. The default value is postgres. |
Postmaster Options | Other command line options for postmaster. |
Startup Wait (seconds) | Specifies the number of seconds to wait for correct end of service startup. |
Shutdown Wait (seconds) | Specifies the number of seconds to wait for correct end of service shutdown. |
Field | Description |
---|---|
SAP Database Name | Specifies a unique SAP system identifier. For example, P01. |
SAP executable directory | Specifies the fully qualified path to sapstartsrv and sapcontrol . |
Database type | Specifies one of the following database types: Oracle, DB6, or ADA. |
Oracle TNS listener name | Specifies Oracle TNS listener name. |
ABAP stack is not installed, only Java stack is installed | If you do not have an ABAP stack installed in the SAP database, enable this parameter. |
Application Level Monitoring | Activates application level monitoring. |
Automatic Startup Recovery | Enable or disable automatic startup recovery. |
Path to Java SDK | Path to Java SDK. |
File name of the JDBC Driver | File name of the JDBC driver. |
Path to a pre-start script | Path to a pre-start script. |
Path to a post-start script | Path to a post-start script. |
Path to a pre-stop script | Path to a pre-stop script |
Path to a post-stop script | Path to a post-stop script |
J2EE instance bootstrap directory | The fully qualified path the J2EE instance bootstrap directory. For example, /usr/sap/P01/J00/j2ee/cluster/bootstrap . |
J2EE security store path | The fully qualified path the J2EE security store directory. For example, /usr/sap/P01/SYS/global/security/lib/tools . |
Field | Description |
---|---|
SAP Instance Name | The fully qualified SAP instance name. For example, P01_DVEBMGS00_sapp01ci. |
SAP executable directory | The fully qualified path to sapstartsrv and sapcontrol . |
Directory containing the SAP START profile | The fully qualified path to the SAP START profile. |
Name of the SAP START profile | Specifies name of the SAP START profile. |
Number of seconds to wait before checking startup status | Specifies the number of seconds to wait before checking the startup status (do not wait for J2EE-Addin). |
Enable automatic startup recovery | Enable or disable automatic startup recovery. |
Path to a pre-start script | Path to a pre-start script. |
Path to a post-start script | Path to a post-start script. |
Path to a pre-stop script | Path to a pre-stop script |
Path to a post-stop script | Path to a post-stop script |
Note
Note
Field | Description |
---|---|
Name | Specifies the name of the Samba server. |
Workgroup | Specifies a Windows workgroup name or Windows NT domain of the Samba service. |
Field | Description |
---|---|
Name | Specifies a name for the custom user script. The script resource allows a standard LSB-compliant init script to be used to start a clustered service. |
Full path to script file | Enter the path where this custom script is located (for example, /etc/init.d/userscript ). |
Field | Description |
---|---|
Service name | Name of service. This defines a collection of resources, known as a resource group or cluster service. |
Automatically start this service | If enabled, this service (or resource group) is started automatically after the cluster forms a quorum. If this parameter is disabled, this service is not started automatically after the cluster forms a quorum; the service is put into the disabled state. |
Run exclusive | If enabled, this service (resource group) can only be relocated to run on another node exclusively; that is, to run on a node that has no other services running on it. If no nodes are available for a service to run exclusively, the service is not restarted after a failure. Additionally, other services do not automatically relocate to a node running this service as Run exclusive . You can override this option by manual start or relocate operations. |
Failover Domain | Defines lists of cluster members to try in the event that a service fails. For information on configuring a failover domain with Conga, refer to Section 3.7, “Configuring a Failover Domain”. For information on configuring a failover domain with system-config-cluster , refer to Section 5.6, “Configuring a Failover Domain”. |
Recovery policy | Recovery policy provides the following options:
|
Field | Description |
---|---|
Instance Name | Specifies the instance name of the Sybase ASE resource. |
ASE server name | The ASE server name that is configured for the HA service. |
SYBASE home directory | The home directory of Sybase products. |
Login file | The full path of login file that contains the login-password pair. |
Interfaces file | The full path of the interfaces file that is used to start/access the ASE server. |
SYBASE_ASE directory name | The directory name under sybase_home where ASE products are installed. |
SYBASE_OCS directory name | The directory name under sybase_home where OCS products are installed. For example, ASE-15_0. |
Sybase user | The user who can run ASE server. |
Deep probe timeout | The maximum seconds to wait for the response of ASE server before determining that the server had no response while running deep probe. |
Field | Description |
---|---|
Name | Specifies a service name for logging and other purposes. |
Config File | Specifies the absolute path to the configuration file. The default value is /etc/tomcat5/tomcat5.conf . |
Tomcat User | User who runs the Tomcat server. The default value is tomcat. |
Catalina Options | Other command line options for Catalina. |
Catalina Base | Catalina base directory (differs for each service) The default value is /usr/share/tomcat5. |
Shutdown Wait (seconds) | Specifies the number of seconds to wait for correct end of service shutdown. The default value is 30. |
Important
rgmanager
tools to start and stop the virtual machines. Using xm
or virsh
to start the machine can result in the virtual machine running in more than one place, which can cause data corruption in the virtual machine. For information on configuring your system to reduce the chances of administrators accidentally "double-starting" virtual machines by using both cluster and non-cluster tools, refer to Section 2.12, “Configuring Virtual Machines in a Clustered Environment”.
Note
Field | Description |
---|---|
Virtual machine name | Specifies the name of the virtual machine. |
Path to VM configuration files |
A colon-delimited path specification that
xm create searches for the virtual machine configuration file. For example: /etc/xen:/guests/config_files:/var/xen/configs
Important
The path should never directly point to a virtual machine configuration file.
|
VM Migration Mapping |
Specifies an alternate interface for migration. You can specify this when, for example, the network address used for virtual machine migration on a node differs from the address of the node used for cluster communication.
Specifying the following indicates that when you migrate a virtual machine from
member to member2 , you actually migrate to target2 . Similarly, when you migrate from member2 to member , you migrate using target .
member:target,member2:target2
|
Migration type | Specifies a migration type of live or pause . The default setting is live . |
Hypervisor | Hypervisor URI (automatic, KVM, or Xen) |
Automatically start this service | If enabled, this virtual machine is started automatically after the cluster forms a quorum. If this parameter is disabled, this virtual machine service is not started automatically after the cluster forms a quorum; the virtual machine is put into the disabled state. |
Run exclusive | If enabled, this virtual machine can only be relocated to run on another node exclusively; that is, to run on a node that has no other virtual machines running on it. If no nodes are available for a virtual machine to run exclusively, the virtual machine is not restarted after a failure. Additionally, other virtual machines do not automatically relocate to a node running this virtual machine as Run exclusive . You can override this option by manual start or relocate operations. |
Failover Domain | Defines lists of cluster members to try in the event that a virtual machine fails. |
Recovery policy | Recovery policy provides the following options:
|
Maximum number of restart failures before relocating | Maximum number of restarts for an independent subtree before giving up. |
Length of time in seconds after which to forget a restart | Amount of time before a failure is forgotten for an independent subtree. |
Appendix D. HA Resource Behavior
system-config-cluster
, or by editing etc/cluster/cluster.conf
. For descriptions of HA resource parameters, refer to Appendix C, HA Resource Parameters. To understand resource agents in more detail you can view them in /usr/share/cluster
of any cluster node.
Note
/etc/cluster/cluster.conf
.
/etc/cluster/cluster.conf
(in each cluster
node). In the cluster configuration file, each resource tree is an XML
representation that specifies each resource, its attributes, and its
relationship among other resources in the resource tree (parent,
child, and sibling relationships).Note
Note
/etc/cluster/cluster.conf
, for illustration purposes only.
D.1. Parent, Child, and Sibling Relationships Among Resources
rgmanager
. All resources in a service run on the same node. From the perspective of rgmanager
, a cluster service is one entity that can be started, stopped, or relocated. Within a cluster service, however, the hierarchy of the resources determines the order in which each resource is started and stopped.The hierarchical levels consist of parent, child, and sibling.
fs:myfs
(<fs name="myfs" ...>) andip:10.1.1.2
(<ip address="10.1.1.2 .../>) are siblings.fs:myfs
(<fs name="myfs" ...>) is the parent ofscript:script_child
(<script name="script_child"/>).script:script_child
(<script name="script_child"/>) is the child offs:myfs
(<fs name="myfs" ...>).
Example D.1. Resource Hierarchy of Service foo
<service name="foo" ...> <fs name="myfs" ...> <script name="script_child"/> </fs> <ip address="10.1.1.2" .../> </service>
- Parents are started before children.
- Children must all stop cleanly before a parent may be stopped.
- For a resource to be considered in good health, all its children must be in good health.
D.2. Sibling Start Ordering and Resource Child Ordering
- Designates child-type attribute (typed child resource) — If the Service resource designates a child-type attribute for a child resource, the child resource is typed. The child-type attribute explicitly determines the start and the stop order of the child resource.
- Does not designate child-type attribute (non-typed child resource) — If the Service resource does not designate a child-type attribute for a child resource, the child resource is non-typed. The Service resource does not explicitly control the starting order and stopping order of a non-typed child resource. However, a non-typed child resource is started and stopped according to its order in
/etc/cluster/cluster.conf
In addition, non-typed child resources are started after all typed child resources have started and are stopped before any typed child resources have stopped.
Note
D.2.1. Typed Child Resource Start and Stop Ordering
service.sh
” shows the start and stop values as they appear in the Service resource agent, service.sh
. For the Service resource, all LVM children are started first, followed by all File System children, followed by all Script children, and so forth.
Resource | Child Type | Start-order Value | Stop-order Value |
---|---|---|---|
LVM | lvm | 1 | 9 |
File System | fs | 2 | 8 |
GFS File System | clusterfs | 3 | 7 |
NFS Mount | netfs | 4 | 6 |
NFS Export | nfsexport | 5 | 5 |
NFS Client | nfsclient | 6 | 4 |
IP Address | ip | 7 | 2 |
Samba | smb | 8 | 3 |
Script | script | 9 | 1 |
Example D.2. Resource Start and Stop Values: Excerpt from Service Resource Agent, service.sh
<special tag="rgmanager"> <attributes root="1" maxinstances="1"/> <child type="lvm" start="1" stop="9"/> <child type="fs" start="2" stop="8"/> <child type="clusterfs" start="3" stop="7"/> <child type="netfs" start="4" stop="6"/> <child type="nfsexport" start="5" stop="5"/> <child type="nfsclient" start="6" stop="4"/> <child type="ip" start="7" stop="2"/> <child type="smb" start="8" stop="3"/> <child type="script" start="9" stop="1"/> </special>
/etc/cluster/cluster.conf
. For example, consider the starting order and stopping order of the typed child resources in Example D.3, “Ordering Within a Resource Type”.
Example D.3. Ordering Within a Resource Type
<service name="foo"> <script name="1" .../> <lvm name="1" .../> <ip address="10.1.1.1" .../> <fs name="1" .../> <lvm name="2" .../> </service>
Typed Child Resource Starting Order
lvm:1
— This is an LVM resource. All LVM resources are started first.lvm:1
(<lvm name="1" .../>
) is the first LVM resource started among LVM resources because it is the first LVM resource listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:2
— This is an LVM resource. All LVM resources are started first.lvm:2
(<lvm name="2" .../>
) is started afterlvm:1
because it is listed afterlvm:1
in the Service foo portion of/etc/cluster/cluster.conf
.fs:1
— This is a File System resource. If there were other File System resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.ip:10.1.1.1
— This is an IP Address resource. If there were other IP Address resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.script:1
— This is a Script resource. If there were other Script resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.
Typed Child Resource Stopping Order
script:1
— This is a Script resource. If there were other Script resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.ip:10.1.1.1
— This is an IP Address resource. If there were other IP Address resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.fs:1
— This is a File System resource. If there were other File System resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:2
— This is an LVM resource. All LVM resources are stopped last.lvm:2
(<lvm name="2" .../>
) is stopped beforelvm:1
; resources within a group of a resource type are stopped in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:1
— This is an LVM resource. All LVM resources are stopped last.lvm:1
(<lvm name="1" .../>
) is stopped afterlvm:2
; resources within a group of a resource type are stopped in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.
D.2.2. Non-typed Child Resource Start and Stop Ordering
/etc/cluster/cluster.conf
. Additionally, non-typed child resources are started after all typed child resources and stopped before any typed child resources.
Example D.4. Non-typed and Typed Child Resource in a Service
<service name="foo"> <script name="1" .../> <nontypedresource name="foo"/> <lvm name="1" .../> <nontypedresourcetwo name="bar"/> <ip address="10.1.1.1" .../> <fs name="1" .../> <lvm name="2" .../> </service>
Non-typed Child Resource Starting Order
lvm:1
— This is an LVM resource. All LVM resources are started first.lvm:1
(<lvm name="1" .../>
) is the first LVM resource started among LVM resources because it is the first LVM resource listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:2
— This is an LVM resource. All LVM resources are started first.lvm:2
(<lvm name="2" .../>
) is started afterlvm:1
because it is listed afterlvm:1
in the Service foo portion of/etc/cluster/cluster.conf
.fs:1
— This is a File System resource. If there were other File System resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.ip:10.1.1.1
— This is an IP Address resource. If there were other IP Address resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.script:1
— This is a Script resource. If there were other Script resources in Service foo, they would start in the order listed in the Service foo portion of/etc/cluster/cluster.conf
.nontypedresource:foo
— This is a non-typed resource. Because it is a non-typed resource, it is started after the typed resources start. In addition, its order in the Service resource is before the other non-typed resource,nontypedresourcetwo:bar
; therefore, it is started beforenontypedresourcetwo:bar
. (Non-typed resources are started in the order that they appear in the Service resource.)nontypedresourcetwo:bar
— This is a non-typed resource. Because it is a non-typed resource, it is started after the typed resources start. In addition, its order in the Service resource is after the other non-typed resource,nontypedresource:foo
; therefore, it is started afternontypedresource:foo
. (Non-typed resources are started in the order that they appear in the Service resource.)
Non-typed Child Resource Stopping Order
nontypedresourcetwo:bar
— This is a non-typed resource. Because it is a non-typed resource, it is stopped before the typed resources are stopped. In addition, its order in the Service resource is after the other non-typed resource,nontypedresource:foo
; therefore, it is stopped beforenontypedresource:foo
. (Non-typed resources are stopped in the reverse order that they appear in the Service resource.)nontypedresource:foo
— This is a non-typed resource. Because it is a non-typed resource, it is stopped before the typed resources are stopped. In addition, its order in the Service resource is before the other non-typed resource,nontypedresourcetwo:bar
; therefore, it is stopped afternontypedresourcetwo:bar
. (Non-typed resources are stopped in the reverse order that they appear in the Service resource.)script:1
— This is a Script resource. If there were other Script resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.ip:10.1.1.1
— This is an IP Address resource. If there were other IP Address resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.fs:1
— This is a File System resource. If there were other File System resources in Service foo, they would stop in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:2
— This is an LVM resource. All LVM resources are stopped last.lvm:2
(<lvm name="2" .../>
) is stopped beforelvm:1
; resources within a group of a resource type are stopped in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.lvm:1
— This is an LVM resource. All LVM resources are stopped last.lvm:1
(<lvm name="1" .../>
) is stopped afterlvm:2
; resources within a group of a resource type are stopped in the reverse order listed in the Service foo portion of/etc/cluster/cluster.conf
.
D.3. Inheritance, the <resources> Block, and Reusing Resources
Example D.5. NFS Service Set Up for Resource Reuse and Inheritance
<resources> <nfsclient name="bob" target="bob.example.com" options="rw,no_root_squash"/> <nfsclient name="jim" target="jim.example.com" options="rw,no_root_squash"/> <nfsexport name="exports"/> </resources> <service name="foo"> <fs name="1" mountpoint="/mnt/foo" device="/dev/sdb1" fsid="12344"> <nfsexport ref="exports"> <!-- nfsexport's path and fsid attributes are inherited from the mountpoint and fsid attribute of the parent fs resource --> <nfsclient ref="bob"/> <!-- nfsclient's path is inherited from the mountpoint and the fsid is added to the options string during export --> <nfsclient ref="jim"/ > </nfsexport> </fs> <fs name="2" mountpoint="/mnt/bar" device="/dev/sdb2" fsid="12345"> <nfsexport ref="exports"> <nfsclient ref="bob"/> <!-- Because all of the critical data for this resource is either defined in the resources block or inherited, we can reference it again! --> <nfsclient ref="jim"/> </nfsexport> </fs> <ip address="10.2.13.20"/> </service>
- The service would need four nfsclient resources — one per file system (a total of two for file systems), and one per target machine (a total of two for target machines).
- The service would need to specify export path and file system ID to each nfsclient, which introduces chances for errors in the configuration.
D.4. Failure Recovery and Independent Subtrees
__independent_subtree
attribute. For example, in Example D.7, “Service foo Failure Recovery with __independent_subtree
Attribute”, the __independent_subtree
attribute is used to accomplish the following actions:
- If script:script_one fails, restart script:script_two and script:script_one.
- If script:script_two fails, restart just script:script_two.
- If script:script_three fails, restart script:script_one, script:script_two, and script:script_three.
- If script:script_four fails, restart the whole service.
Example D.6. Service foo Normal Failure Recovery
<service name="foo"> <script name="script_one" ...> <script name="script_two" .../> </script> <script name="script_three" .../> </service>
Example D.7. Service foo Failure Recovery with __independent_subtree
Attribute
<service name="foo"> <script name="script_one" __independent_subtree="1" ...> <script name="script_two" __independent_subtree="1" .../> <script name="script_three" .../> </script> <script name="script_four" .../> </service>
__independent_subtree="2"
attribute, which designates the independent subtree as non-critical.
Note
__max_restarts
configures the maximum number of tolerated restarts prior to giving up.__restart_expire_time
configures the amount of time, in seconds, after which a restart is no longer attempted.
D.5. Debugging and Testing Services and Resource Ordering
rg_test
utility. rg_test
is a command-line utility that is run from a shell or a terminal (it is not available in Conga or system-config-cluster
.) Table D.2, “rg_test
Utility Summary” summarizes the actions and syntax for the rg_test
utility.
Action | Syntax |
---|---|
Display the resource rules that rg_test understands. | rg_test rules |
Test a configuration (and /usr/share/cluster) for errors or redundant resource agents. | rg_test test /etc/cluster/cluster.conf |
Display the start and stop ordering of a service. |
Display start order:
rg_test noop /etc/cluster/cluster.conf start service
Display stop order:
rg_test noop /etc/cluster/cluster.conf stop service
|
Explicitly start or stop a service. | Important
Only do this on one node, and always disable the service in rgmanager first.
Start a service:
rg_test test /etc/cluster/cluster.conf start service
Stop a service:
rg_test test /etc/cluster/cluster.conf stop service
|
Calculate and display the resource tree delta between two cluster.conf files. | rg_test delta
For example:
rg_test delta /etc/cluster/cluster.conf.bak /etc/cluster/cluster.conf
|
Appendix E. Cluster Service Resource Check and Failover Timeout
rgmanager
monitors the status of cluster resources, and how to modify the status check interval. The appendix also describes the __enforce_timeouts
service parameter, which indicates that a timeout for an operation should cause a service to fail.
Note
/etc/cluster/cluster.conf
. For a comprehensive list and description of cluster.conf
elements and attributes, refer to the cluster schema at /usr/share/system-config-cluster/misc/cluster.ng
, and the annotated schema at /usr/share/doc/system-config-cluster-X.Y.ZZ/cluster_conf.html
(for example /usr/share/doc/system-config-cluster-1.0.57/cluster_conf.html
).
E.1. Modifying the Resource Status Check Interval
rgmanager
checks the status of individual resources, not whole services. (This is a change from clumanager
on Red Hat Enterprise Linux 3, which periodically checked the status of the whole service.) Every 10 seconds, rgmanager scans the resource tree, looking for resources that have passed their "status check" interval.
cluster.conf
file using the special <action>
tag:
<action name="status" depth="*" interval="10" />
cluster.conf
file. For example, if you had a file system resource for which you wanted to override the status check interval you could specify the file system resource in the cluster.conf
file as follows:
<fs name="test" device="/dev/sdb3"> <action name="status" depth="*" interval="10" /> <nfsexport...> </nfsexport> </fs>
depth
is set to *
, which indicates that these values should be used for all depths. The result is that the test
file system is checked at the highest-defined depth provided by the resource-agent (in this case, 20) every 10 seconds.
E.2. Enforcing Resource Timeouts
__enforce_timeouts="1"
to the reference in the cluster.conf
file.
__enforce_timeouts
attribute set for the netfs
resource. With this attribute set, then if it takes more than 30 seconds to unmount the NFS file system during a recovery process the operation will time out, causing the service to enter the failed state.
</screen> <rm> <failoverdomains/> <resources> <netfs export="/nfstest" force_unmount="1" fstype="nfs" host="10.65.48.65" mountpoint="/data/nfstest" name="nfstest_data" options="rw,sync,soft"/> </resources> <service autostart="1" exclusive="0" name="nfs_client_test" recovery="relocate"> <netfs ref="nfstest_data" __enforce_timeouts="1"/> </service> </rm>
E.3. Changing Consensus Timeout
- If configuring a cluster of 2 or less nodes, consensus will be
(token * 0.2)
, with a maximum of 2000 milliseconds and a minimum of 200 milliseconds. - If configuring a cluster of 3 or more nodes, consensus will be
(token + 2000 milliseconds)
cman
configure your consensus timeout in this fashion, realize that moving from 2 to 3 (or more) nodes will require a cluster restart, since the consensus timeout will need to change to the larger value based on the token timeout.
cluster.conf
as follows:
<totem token="X" consensus="X + 2000" />
X + 2000
automatically. An integer value must be used rather than an equation.
Note
cman
, the number of physical nodes matters and not the presence of the two_node=1
directive in cluster.conf
.
Appendix F. High Availabilty LVM (HA-LVM)
- If the applications are cluster-aware and have been tuned to run simultaneously on multiple machines at a time, then CLVM should be used. Specifically, if more than one node of your cluster will require access to your storage which is then shared among the active nodes, then you must use CLVM. CLVM allows a user to configure logical volumes on shared storage by locking access to physical storage while a logical volume is being configured, and uses clustered locking services to manage the shared storage. For information on CLVM, and on LVM configuration in general, refer to Logical Volume Manager Administration.
- If the applications run optimally in active/passive (failover) configurations where only a single node that accesses the storage is active at any one time, you should use High Availability Logical Volume Management agents (HA-LVM).
- The preferred method uses CLVM, but it will only ever activate the logical volumes exclusively. This has the advantage of easier setup and better prevention of administrative mistakes (like removing a logical volume that is in use). In order to use CLVM, the High Availability Add-On and Resilient Storage Add-On software, including the
clvmd
daemon, must be running.The procedure for configuring HA-LVM using this method is described in Section F.1, “Configuring HA-LVM Failover with CLVM (preferred, Red Hat Enterprise Linux 5.6 and later)”. - The second method uses local machine locking and LVM "tags". This method has the advantage of not requiring any LVM cluster packages; however, there are more steps involved in setting it up and it does not prevent an administrator from mistakenly removing a logical volume from a node in the cluster where it is not active. The procedure for configuring HA-LVM using this method is described in Section F.2, “Configuring HA-LVM Failover with Tagging”.
F.1. Configuring HA-LVM Failover with CLVM (preferred, Red Hat Enterprise Linux 5.6 and later)
- Ensure that your system is configured to support CLVM, which requires the following:
- The High Availability Add-On and Resilient Storage Add-On are installed, including the the
cmirror
package if the CLVM logical volumes are to be mirrored. - The
locking_type
parameter in the global section of the/etc/lvm/lvm.conf
file is set to the value '3'. - The High Availability Add-On and Resilient Storage Add-On software, including the
clvmd
daemon, must be running. For CLVM mirroring, thecmirrord
service must be started as well.
- Create the logical volume and file system using standard LVM and file system commands, as in the following example.
#
pvcreate /dev/sd[cde]1
#vgcreate -cy shared_vg /dev/sd[cde]1
#lvcreate -L 10G -n ha_lv shared_vg
#mkfs.ext3 /dev/shared_vg/ha_lv
#lvchange -an shared_vg/ha_lv
For information on creating LVM logical volumes, refer to Logical Volume Manager Administration. - Edit the
/etc/cluster/cluster.conf
file to include the newly created logical volume as a resource in one of your services. Alternately, you can use Conga or theccs
command to configure LVM and file system resources for the cluster. The following is a sample resource manager section from the/etc/cluster/cluster.conf
file that configures a CLVM logical volume as a cluster resource:<rm> <failoverdomains> <failoverdomain name="FD" ordered="1" restricted="0"> <failoverdomainnode name="neo-01" priority="1"/> <failoverdomainnode name="neo-02" priority="2"/> </failoverdomain> </failoverdomains> <resources> <lvm name="lvm" vg_name="shared_vg" lv_name="ha-lv"/> <fs name="FS" device="/dev/shared_vg/ha-lv" force_fsck="0" force_unmount="1" fsid="64050" fstype="ext3" mountpoint="/mnt" options="" self_fence="0"/> </resources> <service autostart="1" domain="FD" name="serv" recovery="relocate"> <lvm ref="lvm"/> <fs ref="FS"/> </service> </rm>
F.2. Configuring HA-LVM Failover with Tagging
/etc/lvm/lvm.conf
file, perform the following steps:
- Ensure that the
locking_type
parameter in the global section of the/etc/lvm/lvm.conf
file is set to the value '1'. - Create the logical volume and file system using standard LVM and file system commands, as in the following example.
#
pvcreate /dev/sd[cde]1
#vgcreate shared_vg /dev/sd[cde]1
#lvcreate -L 10G -n ha_lv shared_vg
#mkfs.ext3 /dev/shared_vg/ha_lv
For information on creating LVM logical volumes, refer to Logical Volume Manager Administration. - Edit the
/etc/cluster/cluster.conf
file to include the newly created logical volume as a resource in one of your services. Alternately, you can use Conga or theccs
command to configure LVM and file system resources for the cluster. The following is a sample resource manager section from the/etc/cluster/cluster.conf
file that configures a CLVM logical volume as a cluster resource:<rm> <failoverdomains> <failoverdomain name="FD" ordered="1" restricted="0"> <failoverdomainnode name="neo-01" priority="1"/> <failoverdomainnode name="neo-02" priority="2"/> </failoverdomain> </failoverdomains> <resources> <lvm name="lvm" vg_name="shared_vg" lv_name="ha_lv"/> <fs name="FS" device="/dev/shared_vg/ha_lv" force_fsck="0" force_unmount="1" fsid="64050" fstype="ext3" mountpoint="/mnt" options="" self_fence="0"/> </resources> <service autostart="1" domain="FD" name="serv" recovery="relocate"> <lvm ref="lvm"/> <fs ref="FS"/> </service> </rm>
Note
If there are multiple logical volumes in the volume group, then the logical volume name (lv_name
) in thelvm
resource should be left blank or unspecified. Also note that in an HA-LVM configuration, a volume group may be used by only a single service. - Edit the
volume_list
field in the/etc/lvm/lvm.conf
file. Include the name of your root volume group and your hostname as listed in the/etc/cluster/cluster.conf
file preceded by @. The hostname to include here is the machine on which you are editing thelvm.conf
file, not any remote hostname. Note that this string MUST match the node name given in thecluster.conf
file. Below is a sample entry from the/etc/lvm/lvm.conf
file:volume_list = [ "VolGroup00", "@neo-01" ]
This tag will be used to activate shared VGs or LVs. DO NOT include the names of any volume groups that are to be shared using HA-LVM. - Update the
initrd
device on all your cluster nodes:#
mkinitrd -f /boot/initrd-$(uname -r).img $(uname -r)
- Reboot all nodes to ensure the correct
initrd
device is in use.
Appendix G. Upgrading A Red Hat Cluster from RHEL 4 to RHEL 5
- Stop client access to cluster high-availability services.
- At each cluster node, stop the cluster software as follows:
- Stop all high-availability services.
- Run
service rgmanager stop
. - Run
service gfs stop
, if you are using Red Hat GFS. - Run
service clvmd stop
, if CLVM has been used to create clustered volumes.Note
Ifclvmd
is already stopped, an error message is displayed:#
service clvmd stop
Stopping clvm: [FAILED]The error message is the expected result when runningservice clvmd stop
afterclvmd
has stopped. - Depending on the type of cluster manager (either CMAN or GULM), run the following command or commands:
- CMAN — Run
service fenced stop; service cman stop
. - GULM — Run
service lock_gulmd stop
.
- Run
service ccsd stop
.
- Disable cluster software from starting during reboot. At each node, run
/sbin/chkconfig
as follows:#
chkconfig --level 2345 rgmanager off
#chkconfig --level 2345 gfs off
#chkconfig --level 2345 clvmd off
#chkconfig --level 2345 fenced off
#chkconfig --level 2345 cman off
#chkconfig --level 2345 ccsd off
- Edit the cluster configuration file as follows:
- At a cluster node, open
/etc/cluster/cluster.conf
with a text editor. - If your cluster is configured with GULM as the cluster manager, remove the GULM XML elements —
<gulm>
and</gulm>
— and their content from/etc/cluster/cluster.conf
. GULM is not supported in Red Hat Cluster Suite for RHEL 5. Example G.1, “GULM XML Elements and Content” shows an example of GULM XML elements and content. - At the
<clusternode>
element for each node in the configuration file, insertnodeid="number"
aftername="name"
. Use a number value unique to that node. Inserting it there follows the format convention of the<clusternode>
element in a RHEL 5 cluster configuration file.Note
Thenodeid
parameter is required in Red Hat Cluster Suite for RHEL 5. The parameter is optional in Red Hat Cluster Suite for RHEL 4. If your configuration file already containsnodeid
parameters, skip this step. - When you have completed editing
/etc/cluster/cluster.conf
, save the file and copy it to the other nodes in the cluster (for example, using thescp
command).
- If your cluster is a GULM cluster and uses Red Hat GFS, change the superblock of each GFS file system to use the DLM locking protocol. Use the
gfs_tool
command with thesb
andproto
options, specifyinglock_dlm
for the DLM locking protocol:gfs_tool sb device proto lock_dlm
For example:#
gfs_tool sb /dev/my_vg/gfs1 proto lock_dlm
You shouldn't change any of these values if the filesystem is mounted. Are you sure? [y/n]y
current lock protocol name = "lock_gulm" new lock protocol name = "lock_dlm" Done - Update the software in the cluster nodes to RHEL 5 and Red Hat Cluster Suite for RHEL 5. You can acquire and update software through Red Hat Network channels for RHEL 5 and Red Hat Cluster Suite for RHEL 5.
- Run
lvmconf --enable-cluster
. - Enable cluster software to start upon reboot. At each node run
/sbin/chkconfig
as follows:#
chkconfig --level 2345 rgmanager on
#chkconfig --level 2345 gfs on
#chkconfig --level 2345 clvmd on
#chkconfig --level 2345 cman on
- Reboot the nodes. The RHEL 5 cluster software should start while the nodes reboot. Upon verification that the Red Hat cluster is running, the upgrade is complete.
Example G.1. GULM XML Elements and Content
<gulm> <lockserver name="gulmserver1"/> <lockserver name="gulmserver2"/> <lockserver name="gulmserver3"/> </gulm>
Appendix H. Revision History
Revision History | ||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Revision 10.0-8 | Mon Sep 8 2014 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 10.0-5 | Mon Jun 30 2014 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 10.0-3 | Tue Jun 10 2014 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 10.0-2 | Wed Apr 30 2014 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 9.0-6 | Mon Sep 30 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 9.0-5 | Wed Jul 10 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 9.0-4 | Tue May 28 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 9.0-3 | Tue May 21 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 9.0-1 | Mon May 20 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 8.0-6 | Fri Jan 4 2013 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 8.0-5 | Wed Sep 26 2012 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 8.0-3 | Wed Aug 29 2012 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 8.0-1 | Tue Jul 31 2012 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 7.0-3 | Thu Feb 16 2012 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 7.0-2 | Thu Dec 15 2011 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 7.0-1 | Thu Nov 10 2011 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 6.0-1 | Thu Jul 21 2011 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 5.0-1 | Thu Dec 23 2010 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 4.0-1 | Mon Mar 15 2010 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 3.0-1 | Tue Aug 18 2009 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 2.0-1 | Tue Jan 20 2009 | |||||||||||||||||||||||
| ||||||||||||||||||||||||
Revision 1.0-1 | Wed May 21 2008 | |||||||||||||||||||||||
|
Index
A
- ACPI
- Apache HTTP Server
- httpd.conf , Installing and Configuring the Apache HTTP Server
- setting up service, Example of Setting Up Apache HTTP Server
- Apache server resource agent, HA Resource Parameters
- APC power switch, Fence Device Parameters
- APC power switch over SNMP, Fence Device Parameters
B
- behavior, HA resources, HA Resource Behavior
- Brocade fabric switch, Fence Device Parameters
- Bull PAP (Platform Administration Processor), Fence Device Parameters
C
- Cisco MDS, Fence Device Parameters
- Cisco UCS, Fence Device Parameters
- cluster
- administration, Before Configuring a Red Hat Cluster, Managing Red Hat Cluster With Conga, Managing Red Hat Cluster With system-config-cluster
- diagnosing and correcting problems, Diagnosing and Correcting Problems in a Cluster, Diagnosing and Correcting Problems in a Cluster
- disabling the cluster software, Disabling the Cluster Software
- displaying status, Cluster Status Tool, Managing High-Availability Services
- managing node, Managing Cluster Nodes
- starting, Starting the Cluster Software
- starting, stopping, restarting, and deleting, Starting, Stopping, and Deleting Clusters
- cluster administration, Before Configuring a Red Hat Cluster, Managing Red Hat Cluster With Conga, Managing Red Hat Cluster With system-config-cluster
- backing up the cluster database, Backing Up and Restoring the Cluster Database
- compatible hardware, Compatible Hardware
- configuring ACPI, Configuring ACPI For Use with Integrated Fence Devices
- configuring iptables, Enabling IP Ports
- configuring max_luns, Configuring max_luns
- Conga considerations, Considerations for Using Conga
- considerations for using qdisk, Considerations for Using Quorum Disk
- considerations for using quorum disk, Considerations for Using Quorum Disk
- diagnosing and correcting problems in a cluster, Diagnosing and Correcting Problems in a Cluster, Diagnosing and Correcting Problems in a Cluster
- disabling the cluster software, Disabling the Cluster Software
- displaying cluster and service status, Cluster Status Tool, Managing High-Availability Services
- enabling IP ports, Enabling IP Ports
- general considerations, General Configuration Considerations
- managing cluster node, Managing Cluster Nodes
- managing high-availability services, Managing High-Availability Services
- modifying the cluster configuration, Modifying the Cluster Configuration
- network switches and multicast addresses, Multicast Addresses
- restoring the cluster database, Backing Up and Restoring the Cluster Database
- SELinux, Red Hat Cluster Suite and SELinux
- starting and stopping the cluster software, Starting and Stopping the Cluster Software
- starting, stopping, restarting, and deleting a cluster, Starting, Stopping, and Deleting Clusters
- virtual machines, Configuring Virtual Machines in a Clustered Environment
- cluster configuration, Configuring Red Hat Cluster With Conga
- modifying, Modifying the Cluster Configuration
- Cluster Configuration Tool
- accessing, Cluster Configuration Tool
- cluster database
- backing up, Backing Up and Restoring the Cluster Database
- restoring, Backing Up and Restoring the Cluster Database
- cluster resource relationships, Parent, Child, and Sibling Relationships Among Resources
- cluster resource status check, Cluster Service Resource Check and Failover Timeout
- cluster resource types, Considerations for Configuring HA Services
- cluster service
- displaying status, Cluster Status Tool, Managing High-Availability Services
- cluster service managers
- cluster services, Adding a Cluster Service to the Cluster, Adding a Cluster Service to the Cluster
- (see also adding to the cluster configuration)
- Apache HTTP Server, setting up, Example of Setting Up Apache HTTP Server
- httpd.conf , Installing and Configuring the Apache HTTP Server
- cluster software
- configuration, Configuring Red Hat Cluster With Conga
- disabling, Disabling the Cluster Software
- installation and configuration, Configuring Red Hat Cluster With system-config-cluster
- starting and stopping, Starting and Stopping the Cluster Software
- cluster software installation and configuration, Configuring Red Hat Cluster With system-config-cluster
- cluster storage
- configuration, Configuring Cluster Storage
- command line tools table, Command Line Administration Tools
- configuration
- HA service, Considerations for Configuring HA Services
- configuration file
- propagation of, Propagating The Configuration File: New Cluster
- configuring cluster storage , Configuring Cluster Storage
- Configuring High Availability LVM, High Availabilty LVM (HA-LVM)
- Conga
- accessing, Configuring Red Hat Cluster Software
- considerations for cluster administration, Considerations for Using Conga
- overview, Conga
- Conga overview, Conga
D
- Dell DRAC, Fence Device Parameters
E
- Egenera SAN controller, Fence Device Parameters
F
- failover timeout, Cluster Service Resource Check and Failover Timeout
- feedback, Feedback
- fence device
- APC power switch, Fence Device Parameters
- APC power switch over SNMP, Fence Device Parameters
- Brocade fabric switch, Fence Device Parameters
- Bull PAP (Platform Administration Processor), Fence Device Parameters
- Cisco MDS, Fence Device Parameters
- Cisco UCS, Fence Device Parameters
- Dell DRAC, Fence Device Parameters
- Egenera SAN controller, Fence Device Parameters
- Fujitsu Siemens Remoteview Service Board (RSB), Fence Device Parameters
- GNBD (Global Network Block Device), Fence Device Parameters
- HP iLO (Integrated Lights Out), Fence Device Parameters
- HP iLO (Integrated Lights Out) MP, Fence Device Parameters
- IBM Blade Center, Fence Device Parameters
- IBM iPDU, Fence Device Parameters
- IBM Remote Supervisor Adapter II (RSA II), Fence Device Parameters
- IF MIB, Fence Device Parameters
- IPMI (Intelligent Platform Management Interface) LAN, Fence Device Parameters
- manual fencing, Fence Device Parameters
- McData SAN switch, Fence Device Parameters
- QLogic SANBox2 switch, Fence Device Parameters
- RHEV-M REST API, Fence Device Parameters
- RPS-10 power switch, Fence Device Parameters
- SCSI fencing, Fence Device Parameters
- virtual machine fencing, Fence Device Parameters
- Vixel SAN switch, Fence Device Parameters
- VMware (SOAP interface), Fence Device Parameters
- WTI power switch, Fence Device Parameters
- filesystem resource agent, HA Resource Parameters
- Fujitsu Siemens Remoteview Service Board (RSB), Fence Device Parameters
G
- general
- considerations for cluster administration, General Configuration Considerations
- GFS file system resource agent, HA Resource Parameters
- GNBD (Global Network Block Device), Fence Device Parameters
H
- HA service configuration
- hardware
- compatible, Compatible Hardware
- HP iLO (Integrated Lights Out), Fence Device Parameters
- HP iLO (Integrated Lights Out) MP, Fence Device Parameters
- HTTP services
- Apache HTTP Server
- httpd.conf, Installing and Configuring the Apache HTTP Server
- setting up, Example of Setting Up Apache HTTP Server
I
- IBM Blade Center, Fence Device Parameters
- IBM iPDU, Fence Device Parameters
- IBM Remote Supervisor Adapter II (RSA II), Fence Device Parameters
- IF MIB, Fence Device Parameters
- integrated fence devices
- configuring ACPI, Configuring ACPI For Use with Integrated Fence Devices
- introduction, Introduction
- other Red Hat Enterprise Linux documents, Introduction
- IP address resource agent, HA Resource Parameters
- IP ports
- enabling, Enabling IP Ports
- IPMI (Intelligent Platform Management Interface) LAN, Fence Device Parameters
- iptables
- configuring, Enabling IP Ports
- iptables firewall, Configuring the iptables Firewall to Allow Cluster Components
L
- LVM resource agent, HA Resource Parameters
- LVM, High Availability, High Availabilty LVM (HA-LVM)
M
- manual fencing, Fence Device Parameters
- max_luns
- configuring, Configuring max_luns
- McData SAN switch, Fence Device Parameters
- multicast addresses
- considerations for using with network switches and multicast addresses, Multicast Addresses
- multicast traffic, enabling, Configuring the iptables Firewall to Allow Cluster Components
- MySQL resource agent, HA Resource Parameters
N
- NFS client resource agent, HA Resource Parameters
- NFS export resource agent, HA Resource Parameters
- NFS mount resource agent, HA Resource Parameters
O
- open LDAP resource agent, HA Resource Parameters
- Oracle 10g failover instance resource agent, HA Resource Parameters
- Oracle DB resource agent, HA Resource Parameters
- Oracle listener resource agent, HA Resource Parameters
P
- parameters, fence device, Fence Device Parameters
- parameters, HA resources, HA Resource Parameters
- PostgresSQL 8 resource agent, HA Resource Parameters
Q
- qdisk
- considerations for using, Considerations for Using Quorum Disk
- QLogic SANBox2 switch, Fence Device Parameters
- quorum disk
- considerations for using, Considerations for Using Quorum Disk
R
- relationships
- cluster resource, Parent, Child, and Sibling Relationships Among Resources
- resource agent
- Apache server, HA Resource Parameters
- filesystem, HA Resource Parameters
- GFS file system, HA Resource Parameters
- IP address, HA Resource Parameters
- LVM, HA Resource Parameters
- MySQL, HA Resource Parameters
- NFS client, HA Resource Parameters
- NFS export, HA Resource Parameters
- NFS mount, HA Resource Parameters
- open LDAP, HA Resource Parameters
- Oracle 10g failover instance, HA Resource Parameters
- Oracle DB, HA Resource Parameters
- Oracle listener, HA Resource Parameters
- PostgresSQL 8, HA Resource Parameters
- Samba service, HA Resource Parameters
- SAP database, HA Resource Parameters
- SAP instance, HA Resource Parameters
- Sybase ASE failover instance, HA Resource Parameters
- Tomcat 5, HA Resource Parameters
- RHEV-M REST API, Fence Device Parameters
- RPS-10 power switch, Fence Device Parameters
S
- Samba service resource agent, HA Resource Parameters
- SAP database resource agent, HA Resource Parameters
- SAP instance resource agent, HA Resource Parameters
- SCSI fencing, Fence Device Parameters
- SELinux
- configuring, Red Hat Cluster Suite and SELinux
- starting the cluster software, Starting the Cluster Software
- status check, cluster resource, Cluster Service Resource Check and Failover Timeout
- Sybase ASE failover instance resource agent, HA Resource Parameters
- System V init , Starting and Stopping the Cluster Software
T
- table
- command line tools, Command Line Administration Tools
- tables
- HA resources, parameters, HA Resource Parameters
- timeout failover, Cluster Service Resource Check and Failover Timeout
- Tomcat 5 resource agent, HA Resource Parameters
- troubleshooting
- diagnosing and correcting problems in a cluster, Diagnosing and Correcting Problems in a Cluster, Diagnosing and Correcting Problems in a Cluster
- types
- cluster resource, Considerations for Configuring HA Services
U
- upgrading, RHEL 4 to RHEL 5, Upgrading A Red Hat Cluster from RHEL 4 to RHEL 5
V
- virtual machine fencing, Fence Device Parameters
- virtual machine resource service, HA Resource Parameters
- virtual machines, in a cluster, Configuring Virtual Machines in a Clustered Environment
- Vixel SAN switch, Fence Device Parameters
- VMware (SOAP interface), Fence Device Parameters
W
- WTI power switch, Fence Device Parameters