Administration and troubleshooting
Monitor the status of the cluster
You can check the status of the cluster by running the following command.
hahost01:~ # crm status Cluster Summary: * Stack: corosync * Current DC: hahost01 (version 2.0.xxxxxxxxxxxx) - partition with quorum * Last updated: * Last change: by root via crm_resource on hahost01 * 2 nodes configured * 7 resource instances configured Node List: * Online: [ hahost01 hahost02 ] Full List of Resources: * res_AWS_STONITH (stonith:external/ec2): Started hahost01 * Resource Group: grp_HA1_ASCS00: * rsc_IP_HA1_ASCS00 (ocf::suse:aws-vpc-move-ip): Started hahost01 * rsc_FS_HA1_ASCS00 (ocf::heartbeat:Filesystem): Started hahost01 * rsc_SAP_HA1_ASCS00 (ocf::heartbeat:SAPInstance): Started hahost01 * Resource Group: grp_HA1_ERS10: * rsc_IP_HA1_ERS10 (ocf::suse:aws-vpc-move-ip): Started hahost02 * rsc_FS_HA1_ERS10 (ocf::heartbeat:Filesystem): Started hahost02 * rsc_SAP_HA1_ERS10 (ocf::heartbeat:SAPInstance): Started hahost02
Cluster administration
To manually migrate the cluster resources from one node to another, run the following command.
hahost01:~ # crm resource move rsc_SAP_HA1_ASCS00 force
You can verify the status of resource migration by running the following command.
hahost01:~ # crm status Cluster Summary: * Stack: corosync * Current DC: hahost01 (version 2.0.xxxxxxxxxxxx) - partition with quorum * Last updated: * Last change: by root via crm_resource on hahost01 * 2 nodes configured * 7 resource instances configured Node List: * Online: [ hahost01 hahost02 ] Full List of Resources: * res_AWS_STONITH (stonith:external/ec2): Started hahost02 * Resource Group: grp_HA1_ASCS00: * rsc_IP_HA1_ASCS00 (ocf::suse:aws-vpc-move-ip): Started hahost02 * rsc_FS_HA1_ASCS00 (ocf::heartbeat:Filesystem): Started hahost02 * rsc_SAP_HA1_ASCS00 (ocf::heartbeat:SAPInstance): Started hahost02 * Resource Group: grp_HA1_ERS10: * rsc_IP_HA1_ERS10 (ocf::suse:aws-vpc-move-ip): Started hahost01 * rsc_FS_HA1_ERS10 (ocf::heartbeat:Filesystem): Started hahost01 * rsc_SAP_HA1_ERS10 (ocf::heartbeat:SAPInstance): Started hahost01
With the invocation of each
crm
resource move command, the cluster creates location constraints that move
the resource. These constraints must be removed to enable automated failover in the
future.
To remove the location constraints that are created by the move, run the following command.
hahost01:~ # crm resource clear rsc_SAP_HA1_ASCS00
Resource cleanup activities
You can run the following command to clean any failed actions.
hahost01:~ # crm resource cleanup rsc_SAP_HA1_ASCS00 hahost01 Cleaned up rsc_IP_HA1_ASCS00 on hahost01 Cleaned up rsc_FS_HA1_ASCS00 on hahost01 Cleaned up rsc_SAP_HA1_ASCS00 on hahost01
Checking the logs
Start troubleshooting by checking the logs at /var/log/messages
. You can find additional information in the following logs:
Cluster logs – updated in the
corosync.log
, located at/var/log/cluster
.
Pacemaker logs – updated in the
pacemaker.log
, located at/var/log/pacemaker
.