Administration and troubleshooting - SAP NetWeaver on Amazon
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China.

Administration and troubleshooting

Monitor the status of the cluster

You can check the status of the cluster by running the following command.

hahost01:~ # crm status Cluster Summary: * Stack: corosync * Current DC: hahost01 (version 2.0.xxxxxxxxxxxx) - partition with quorum * Last updated: * Last change: by root via crm_resource on hahost01 * 2 nodes configured * 7 resource instances configured Node List: * Online: [ hahost01 hahost02 ] Full List of Resources: * res_AWS_STONITH (stonith:external/ec2): Started hahost01 * Resource Group: grp_HA1_ASCS00: * rsc_IP_HA1_ASCS00 (ocf::suse:aws-vpc-move-ip): Started hahost01 * rsc_FS_HA1_ASCS00 (ocf::heartbeat:Filesystem): Started hahost01 * rsc_SAP_HA1_ASCS00 (ocf::heartbeat:SAPInstance): Started hahost01 * Resource Group: grp_HA1_ERS10: * rsc_IP_HA1_ERS10 (ocf::suse:aws-vpc-move-ip): Started hahost02 * rsc_FS_HA1_ERS10 (ocf::heartbeat:Filesystem): Started hahost02 * rsc_SAP_HA1_ERS10 (ocf::heartbeat:SAPInstance): Started hahost02

Cluster administration

To manually migrate the cluster resources from one node to another, run the following command.

hahost01:~ # crm resource move rsc_SAP_HA1_ASCS00 force

You can verify the status of resource migration by running the following command.

hahost01:~ # crm status Cluster Summary: * Stack: corosync * Current DC: hahost01 (version 2.0.xxxxxxxxxxxx) - partition with quorum * Last updated: * Last change: by root via crm_resource on hahost01 * 2 nodes configured * 7 resource instances configured Node List: * Online: [ hahost01 hahost02 ] Full List of Resources: * res_AWS_STONITH (stonith:external/ec2): Started hahost02 * Resource Group: grp_HA1_ASCS00: * rsc_IP_HA1_ASCS00 (ocf::suse:aws-vpc-move-ip): Started hahost02 * rsc_FS_HA1_ASCS00 (ocf::heartbeat:Filesystem): Started hahost02 * rsc_SAP_HA1_ASCS00 (ocf::heartbeat:SAPInstance): Started hahost02 * Resource Group: grp_HA1_ERS10: * rsc_IP_HA1_ERS10 (ocf::suse:aws-vpc-move-ip): Started hahost01 * rsc_FS_HA1_ERS10 (ocf::heartbeat:Filesystem): Started hahost01 * rsc_SAP_HA1_ERS10 (ocf::heartbeat:SAPInstance): Started hahost01

With the invocation of each crm resource move command, the cluster creates location constraints that move the resource. These constraints must be removed to enable automated failover in the future.

To remove the location constraints that are created by the move, run the following command.

hahost01:~ # crm resource clear rsc_SAP_HA1_ASCS00

Resource cleanup activities

You can run the following command to clean any failed actions.

hahost01:~ # crm resource cleanup rsc_SAP_HA1_ASCS00 hahost01 Cleaned up rsc_IP_HA1_ASCS00 on hahost01 Cleaned up rsc_FS_HA1_ASCS00 on hahost01 Cleaned up rsc_SAP_HA1_ASCS00 on hahost01

Checking the logs

Start troubleshooting by checking the logs at /var/log/messages. You can find additional information in the following logs:

  • Cluster logs – updated in the corosync.log, located at /var/log/cluster.

  • Pacemaker logs – updated in the pacemaker.log, located at /var/log/pacemaker.