Administration and troubleshooting - SAP NetWeaver on Amazon
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China.

Administration and troubleshooting

Monitor the status of the cluster

You can check the status of the cluster by running the following command.

[root@hahost01 ~]# pcs status Cluster name: rhelha Stack: corosync Current DC: hahost01 (version 1.1.19-8.el7_6.5-c3c624ea3d) - partition with quorum Last updated: Mon May 17 01:56:42 2021 Last change: Mon May 17 01:54:41 2021 by root via crm_resource on hahost01 2 nodes configured 7 resources configured Online: [ hahost01 hahost02 ] Full list of resources: clusterfence (stonith:fence_aws): Started hahost02 Resource Group: rsc_ASCS00_group rsc_fs_ascs00 (ocf::heartbeat:Filesystem): Started hahost01 rsc_vip_ascs00 (ocf::heartbeat:aws-vpc-move-ip): Started hahost01 rsc_ascs00 (ocf::heartbeat:SAPInstance): Started hahost01 Resource Group: rsc_ERS10_group rsc_fs_ers10 (ocf::heartbeat:Filesystem): Started hahost02 rsc_vip_ers10 (ocf::heartbeat:aws-vpc-move-ip): Started hahost02 rsc_ers10 (ocf::heartbeat:SAPInstance): Started hahost02 Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled

Cluster administration

To manually migrate the cluster resources from one node to another, run the following command.

[root@hahost01 ~]# pcs resource move rsc_ascs00 hahost02

You can verify the status of resource migration by running the following command.

[root@hahost01 ~]# pcs status Cluster name: rhelha Stack: corosync Current DC: hahost01 (version 1.1.19-8.el7_6.5-c3c624ea3d) - partition with quorum Last updated: Mon May 17 01:56:42 2021 Last change: Mon May 17 01:54:41 2021 by root via crm_resource on hahost01 2 nodes configured 7 resources configured Online: [ hahost01 hahost02 ] Full list of resources: clusterfence (stonith:fence_aws): Started hahost02 Resource Group: rsc_ASCS00_group rsc_fs_ascs00 (ocf::heartbeat:Filesystem): Started hahost02 rsc_vip_ascs00 (ocf::heartbeat:aws-vpc-move-ip): Started hahost02 rsc_ascs00 (ocf::heartbeat:SAPInstance): Started hahost02 Resource Group: rsc_ERS10_group rsc_fs_ers10 (ocf::heartbeat:Filesystem): Started hahost01 rsc_vip_ers10 (ocf::heartbeat:aws-vpc-move-ip): Started hahost01 rsc_ers10 (ocf::heartbeat:SAPInstance): Started hahost01 Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled

With the invocation of each pcs resource move command, the cluster creates location constraints that move the resource. These constraints must be removed to enable automated failover in the future.

To check the constraints created by the move, run the following command.

[root@hahost01 ~]# pcs constraint list --full Location Constraints: Resource: rsc_ascs00 Enabled on: hahost02 (score:INFINITY) (role: Started) (id:cli-prefer-HA1_ASCS00) Ordering Constraints: Start rsc_ASCS00_group then start rsc_ERS10_group (kind:Optional) (non-symmetrical) (id:order- rsc_ASCS00_group-rsc_ERS10_group -Optional) Colocation Constraints: rsc_ERS10_group with rsc_ASCS00_group (score:-5000) (id:colocation- rsc_ERS10_group-rsc_ASCS00_group --5000)

To remove the location constraints created by the move, run the following command.

[root@hahost01 ~]# pcs resource clear rsc_ascs00 hahost02

Resource cleanup activities

You can run the following command to clean any failed actions.

pcs resource cleanup <resource> --node <node-name> [root@hahost02 ~]# pcs resource cleanup rsc_ers10 --node hahost02 Cleaned up rsc_fs_ers10 on hahost02 Cleaned up rsc_vip_ers10 on hahost02 Cleaned up rsc_ers10 on hahost02 Waiting for 1 replies from the CRMd. OK

Checking the logs

Start troubleshooting by checking the logs at /var/log/messages. You can find additional information in the following logs:

  • Cluster logs – updated in the corosync.log, located at /var/log/cluster/corosync/log.

  • Pacemaker logs – updated in the pacemaker.log, located at /var/log/pacemaker.