Administration and troubleshooting
Monitor the status of the cluster
You can check the status of the cluster by running the following command.
[root@hahost01 ~]# pcs status Cluster name: rhelha Stack: corosync Current DC: hahost01 (version 1.1.19-8.el7_6.5-c3c624ea3d) - partition with quorum Last updated: Mon May 17 01:56:42 2021 Last change: Mon May 17 01:54:41 2021 by root via crm_resource on hahost01 2 nodes configured 7 resources configured Online: [ hahost01 hahost02 ] Full list of resources: clusterfence (stonith:fence_aws): Started hahost02 Resource Group: rsc_ASCS00_group rsc_fs_ascs00 (ocf::heartbeat:Filesystem): Started hahost01 rsc_vip_ascs00 (ocf::heartbeat:aws-vpc-move-ip): Started hahost01 rsc_ascs00 (ocf::heartbeat:SAPInstance): Started hahost01 Resource Group: rsc_ERS10_group rsc_fs_ers10 (ocf::heartbeat:Filesystem): Started hahost02 rsc_vip_ers10 (ocf::heartbeat:aws-vpc-move-ip): Started hahost02 rsc_ers10 (ocf::heartbeat:SAPInstance): Started hahost02 Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled
Cluster administration
To manually migrate the cluster resources from one node to another, run the following command.
[root@hahost01 ~]# pcs resource move rsc_ascs00 hahost02
You can verify the status of resource migration by running the following command.
[root@hahost01 ~]# pcs status Cluster name: rhelha Stack: corosync Current DC: hahost01 (version 1.1.19-8.el7_6.5-c3c624ea3d) - partition with quorum Last updated: Mon May 17 01:56:42 2021 Last change: Mon May 17 01:54:41 2021 by root via crm_resource on hahost01 2 nodes configured 7 resources configured Online: [ hahost01 hahost02 ] Full list of resources: clusterfence (stonith:fence_aws): Started hahost02 Resource Group: rsc_ASCS00_group rsc_fs_ascs00 (ocf::heartbeat:Filesystem): Started hahost02 rsc_vip_ascs00 (ocf::heartbeat:aws-vpc-move-ip): Started hahost02 rsc_ascs00 (ocf::heartbeat:SAPInstance): Started hahost02 Resource Group: rsc_ERS10_group rsc_fs_ers10 (ocf::heartbeat:Filesystem): Started hahost01 rsc_vip_ers10 (ocf::heartbeat:aws-vpc-move-ip): Started hahost01 rsc_ers10 (ocf::heartbeat:SAPInstance): Started hahost01 Daemon Status: corosync: active/enabled pacemaker: active/enabled pcsd: active/enabled
With the invocation of
each pcs
resource move command, the cluster creates location constraints
that move the resource. These constraints must be removed to enable automated failover
in the future.
To check the constraints created by the move, run the following command.
[root@hahost01 ~]# pcs constraint list --full Location Constraints: Resource: rsc_ascs00 Enabled on: hahost02 (score:INFINITY) (role: Started) (id:cli-prefer-HA1_ASCS00) Ordering Constraints: Start rsc_ASCS00_group then start rsc_ERS10_group (kind:Optional) (non-symmetrical) (id:order- rsc_ASCS00_group-rsc_ERS10_group -Optional) Colocation Constraints: rsc_ERS10_group with rsc_ASCS00_group (score:-5000) (id:colocation- rsc_ERS10_group-rsc_ASCS00_group --5000)
To remove the location constraints created by the move, run the following command.
[root@hahost01 ~]# pcs resource clear rsc_ascs00 hahost02
Resource cleanup activities
You can run the following command to clean any failed actions.
pcs resource cleanup <resource> --node <node-name> [root@hahost02 ~]# pcs resource cleanup rsc_ers10 --node hahost02 Cleaned up rsc_fs_ers10 on hahost02 Cleaned up rsc_vip_ers10 on hahost02 Cleaned up rsc_ers10 on hahost02 Waiting for 1 replies from the CRMd. OK
Checking the logs
Start troubleshooting by checking the logs at /var/log/messages
. You can find additional information in the following logs:
Cluster logs – updated in the
corosync.log
, located at/var/log/cluster/corosync/log
.
Pacemaker logs – updated in the
pacemaker.log
, located at/var/log/pacemaker
.