Dieser Inhalt ist in der von Ihnen ausgewählten Sprache nicht verfügbar.
Chapter 7. Test failover
7.1. Failover ASCS due to node crash
Before the crash, ASCS was running on s4node1 while ERS was running on s4node2.
On s4node2, run the following command to monitor the status changes in the cluster:
[root@s4node2 ~]# crm_mon -Arf
Crash s4node1 by running the following command. Please note that connection to s4node1 will be lost after the command.
[root@s4node1 ~]# echo c > /proc/sysrq-trigger
On s4node2, monitor the failover process. After failover, cluster should be in such state, with ASCS running on s4node3, and ERS remaining on s4node2.
[root@s4node2 ~]# pcs status ... Resource Group: s4h_ASCS20_group s4h_fs_ascs20 (ocf::heartbeat:Filesystem): Started s4node1 s4h_vip_ascs20 (ocf::heartbeat:IPaddr2): Started s4node1 s4h_ascs20 (ocf::heartbeat:SAPInstance): Started s4node1 Resource Group: s4h_ERS29_group s4h_fs_ers29 (ocf::heartbeat:Filesystem): Started s4node2 s4h_vip_ers29 (ocf::heartbeat:IPaddr2): Started s4node2 s4h_ers29 (ocf::heartbeat:SAPInstance): Started s4node2 ...
7.2. ERS remains on current node
Bring s4node1 back online. ERS should remain on the current node instead of moving back to s4node1.
7.3. Test ERS crash
Similarly, test crash the node where ERS is running. The ERS group should failover to the spare node while ASCS remains intact on its current node. After the crashed node is back, the ERS group should not move back.
7.4. Your cost-optimized SAP S/4HANA Cluster environment can look like the one below
[root@s4node1 ~]# pcs status Cluster name: SAP-S4-HANA …. Node List: * Online: [ s4node1 s4node2 ] …. Full List of Resources: * s4-fence (stonith:fence_rhevm): Started s4node1 * Clone Set: fs_sapmnt-clone [fs_sapmnt]: * Started: [ s4node1 s4node2 ] * Clone Set: fs_sap_trans-clone [fs_sap_trans]: * Started: [ s4node1 s4node2 ] * Clone Set: fs_sap_SYS-clone [fs_sap_SYS]: * Started: [ s4node1 s4node2 ] * Resource Group: s4h_ASCS20_group: * s4h_lvm_ascs20 (ocf::heartbeat:LVM-activate): Started s4node1 * s4h_fs_ascs20 (ocf::heartbeat:Filesystem): Started s4node1 * s4h_ascs20 (ocf::heartbeat:SAPInstance): Started s4node1 * s4h_vip_ascs20 (ocf::heartbeat:IPaddr2): Started s4node1 * Resource Group: s4h_ERS29_group: * s4h_lvm_ers29 (ocf::heartbeat:LVM-activate): Started s4node2 * s4h_fs_ers29 (ocf::heartbeat:Filesystem): Started s4node2 * s4h_ers29 (ocf::heartbeat:SAPInstance): Started s4node2 * s4h_vip_ers29 (ocf::heartbeat:IPaddr2): Started s4node2 * Clone Set: SAPHanaTopology_S4D_00-clone [SAPHanaTopology_S4D_00]: * Started: [ s4node1 s4node2 ] * Clone Set: SAPHana_S4D_00-clone [SAPHana_S4D_00] (promotable): * Masters: [ s4node2 ] * Slaves: [ s4node1 ] * vip_S4D_00 (ocf::heartbeat:IPaddr2): Started s4node2