Configuring high availability for SAP S/4HANA (ASCS and ERS) in a Red Hat Enterprise Linux High Availability Add-On cluster in a multizone region environment
The following information describes the configuration of ABAP SAP Central Services (ASCS) and Enqueue Replication Server (ERS) in a Red Hat Enterprise Linux (RHEL) High Availability Add-On cluster. The cluster uses virtual server instances in IBM® Power® Virtual Server as cluster nodes.
This example configuration applies to the second generation of the Standalone Enqueue Server, also called ENSA2.
Starting with the release of SAP S/4HANA 1809, ENSA2 is installed by default, and can be configured in a two-node or multi-node cluster. This example uses the ENSA2 setup for a two-node RHEL HA Add-On cluster. If the ASCS service fails in a two-node cluster, it restarts on the node where the ERS instance is running. The lock entries for the SAP application are then restored from the copy of the lock table in the ERS instance. When an administrator activates the failed cluster node, the ERS instance moves to the other node (anti-collocation) to protect its copy of the lock table.
It is recommended that you install the SAP database instance and other SAP application server instances on virtual server instances outside the two-node cluster for ASCS and ERS.
Before you begin
Review the general requirements, product documentation, support articles, and SAP notes listed in Implementing high availability for SAP applications on IBM Power Virtual Server References.
Prerequisites
-
This information describes a setup that uses NFS mounted storage for the instance directories.
- The ASCS instance uses the mount point
/usr/sap/<SID>/ASCS<INSTNO>
. - The ERS instance uses the mount point
/usr/sap/<SID>/ERS<INSTNO>
. - Both instances use the
/sapmnt/<SID>
mount point with shared read and write access. - Other shared file systems such as saptrans
/usr/sap/trans
might be needed.
Make sure that a highly available NFS server is configured to serve these shares. The NFS server must not be installed on a virtual server that is part of the ENSA2 cluster. This document does not describe the steps for setting up file storage or creating cluster file system resources.
- The ASCS instance uses the mount point
-
The virtual hostnames for the ASCS and ERS instances must meet the requirements as documented in Hostnames of SAP ABAP Platform servers.
-
The subnets and the virtual IP addresses for the ASCS and ERS instances must not exist in the Power Virtual Server workspaces. They are configured as cluster resources. However, you must add the virtual IP addresses and virtual hostnames for the ASCS and ERS instances to the Domain Name Service (DNS) and to the
/etc/hosts
file on all cluster nodes.
Preparing nodes to install ASCS and ERS instances
The following information describes how to prepare the nodes for installing the SAP ASCS and ERS instances.
Preparing environment variables
To simplify the setup, prepare the following environment variables for user root
on both cluster nodes. These environment variables are used with later operating system commands in this information.
On both nodes, set the following environment variables.
# General settings
export CLUSTERNAME="SAP_S01" # Cluster name
export NODE1=<HOSTNAME_1> # Virtual server instance 1 hostname
export NODE2=<HOSTNAME_2> # Virtual server instance 2 hostname
export SID=<SID> # SAP System ID (uppercase)
export sid=<sid> # SAP System ID (lowercase)
# ASCS instance
export ASCS_INSTNO=<INSTNO> # ASCS instance number
export ASCS_NET=<Subnet name> # Name for the ASCS subnet in IBM Cloud
export ASCS_CIDR=<CIDR of subnet> # CIDR of the ASCS subnet containing the service IP address
export ASCS_VH=<virtual hostname> # ASCS virtual hostname
export ASCS_IP=<IP address> # ASCS virtual IP address
# ERS instance
export ERS_INSTNO=<INSTNO> # ERS instance number
export ERS_NET=<Subnet name> # Name for the ERS subnet in IBM Cloud
export ERS_CIDR=<CIDR of subnet> # CIDR of the ERS subnet containing the service IP address
export ERS_VH=<virtual hostname> # ERS virtual hostname
export ERS_IP=<IP address> # ERS virtual IP address
# Other multizone region settings
export CLOUD_REGION=<CLOUD_REGION> # Multizone region name
export APIKEY="APIKEY or path to file" # API key of the ServiceID for the resource agent
export API_TYPE="private or public" # Use private or public API endpoints
export IBMCLOUD_CRN_1=<IBMCLOUD_CRN_1> # Workspace 1 CRN
export IBMCLOUD_CRN_2=<IBMCLOUD_CRN_2> # Workspace 2 CRN
export POWERVSI_1=<POWERVSI_1> # Virtual server 1 instance id
export POWERVSI_2=<POWERVSI_2> # Virtual server 2 instance id
export JUMBO="true or false" # Enable Jumbo frames
# NFS settings
export NFS_SERVER="NFS server" # Hostname or IP address of the highly available NFS server
export NFS_SHARE="NFS server directory" # Exported file system directory on the NFS server
export NFS_OPTIONS="rw,sec=sys" # Sample NFS client mount options
The following is an example of how to set the extra environment variables that are required for a multizone region implementation.
# General settings
export CLUSTERNAME="SAP_S01" # Cluster name
export NODE1="cl-s01-1" # Virtual service instance 1 hostname
export NODE2="cl-s01-2" # Virtual server instance 2 hostname
export SID="S01" # SAP System ID (uppercase)
export sid="s01" # SAP System ID (lowercase)
# ASCS instance
export ASCS_INSTNO="21" # ASCS instance number
export ASCS_NET="s01-ascs-net" # Name for the ASCS subnet in IBM Cloud
export ASCS_CIDR="10.40.21.100/30" # CIDR of the ASCS subnet containing the service IP address
export ASCS_VH="s01ascs" # ASCS virtual hostname
export ASCS_IP="10.40.21.102" # ASCS virtual IP address
# ERS instance
export ERS_INSTNO="22" # ERS instance number
export ERS_NET="s01-ers-net" # Name for the ERS subnet in IBM Cloud
export ERS_CIDR="10.40.22.100/30" # CIDR of the ERS subnet containing the service IP address
export ERS_VH="s01ers" # ERS virtual hostname
export ERS_IP="10.40.22.102" # ERS virtual IP address
# Other multizone region settings
export CLOUD_REGION="eu-de"
export IBMCLOUD_CRN_1="crn:v1:bluemix:public:power-iaas:eu-de-2:a/a1b2c3d4e5f60123456789a1b2c3d4e5:a1b2c3d4-0123-4567-89ab-a1b2c3d4e5f6::"
export IBMCLOUD_CRN_2="crn:v1:bluemix:public:power-iaas:eu-de-1:a/a1b2c3d4e5f60123456789a1b2c3d4e5:e5f6a1b2-cdef-0123-4567-a1b2c3d4e5f6::"
export POWERVSI_1="a1b2c3d4-0123-890a-f012-0123456789ab"
export POWERVSI_2="e5f6a1b2-4567-bcde-3456-cdef01234567"
export APIKEY="@/root/.apikey.json"
export API_TYPE="private"
export JUMBO="true"
# NFS settings
export NFS_SERVER="cl-nfs" # Hostname or IP address of the highly available NFS server
export NFS_SHARE="/sapS01" # Exported file system directory on the NFS server
export NFS_OPTIONS="rw,sec=sys" # Sample NFS client mount options
Creating mount points for the instance file systems
On both nodes, run the following command to create the mount points for the instance file systems.
mkdir -p /usr/sap/${SID}/{ASCS${ASCS_INSTNO},ERS${ERS_INSTNO}} /sapmnt/${SID}
Installing and setting up the RHEL HA Add-On cluster
Install and set up the RHEL HA Add-On cluster according to Implementing a RHEL HA Add-On cluster on IBM Power Virtual Server in a Multizone Region Environment.
Configure and test the cluster fencing as described in Creating the fencing device.
Preparing cluster resources before the SAP installation
Make sure that the RHEL HA Add-On cluster is running on both virtual server instances and that node fencing has been tested.
Preparing to install the ASCS instance on NODE1
On NODE1, run the following command to create a Filesystem cluster resource that mounts the ASCS instance directory.
pcs resource create ${sid}_fs_ascs${ASCS_INSTNO} Filesystem \
device="${NFS_SERVER}:${NFS_SHARE}/ASCS" \
directory=/usr/sap/${SID}/ASCS${ASCS_INSTNO} \
fstype=nfs \
options="${NFS_OPTIONS}" \
force_unmount=safe \
op start interval=0 timeout=60 \
op stop interval=0 timeout=120 \
--group ${sid}_ascs${ASCS_INSTNO}_group
On NODE1, run the following command to create a powervs-subnet
cluster resource for the ASCS virtual IP address.
pcs resource create ${sid}_vip_ascs${ASCS_INSTNO} powervs-subnet \
api_key=${APIKEY} \
api_type=${API_TYPE} \
cidr=${ASCS_CIDR} \
ip=${ASCS_IP} \
crn_host_map="${NODE1}:${IBMCLOUD_CRN_1};${NODE2}:${IBMCLOUD_CRN_2}" \
vsi_host_map="${NODE1}:${POWERVSI_1};${NODE2}:${POWERVSI_2}" \
jumbo=${JUMBO} \
region=${CLOUD_REGION} \
subnet_name=${ASCS_NET} \
route_table=5${ASCS_INSTNO} \
op start timeout=720 \
op stop timeout=300 \
op monitor interval=60 timeout=30 \
--group ${sid}_ascs${ASCS_INSTNO}_group
Preparing to install the ERS instance on NODE2
On NODE1, run the following command to create a Filesystem cluster resource to mount the ERS instance directory.
pcs resource create ${sid}_fs_ers${ERS_INSTNO} Filesystem \
device="${NFS_SERVER}:${NFS_SHARE}/ERS" \
directory=/usr/sap/${SID}/ERS${ERS_INSTNO} \
fstype=nfs \
options="${NFS_OPTIONS}" \
force_unmount=safe \
op start interval=0 timeout=60 \
op stop interval=0 timeout=120 \
--group ${sid}_ers${ERS_INSTNO}_group
On NODE1, run the following command to create a powervs-subnet
cluster resource for the ERS virtual IP address.
pcs resource create ${sid}_vip_ers${ERS_INSTNO} powervs-subnet \
api_key=${APIKEY} \
api_type=${API_TYPE} \
cidr=${ERS_CIDR} \
ip=${ERS_IP} \
crn_host_map="${NODE1}:${IBMCLOUD_CRN_1};${NODE2}:${IBMCLOUD_CRN_2}" \
vsi_host_map="${NODE1}:${POWERVSI_1};${NODE2}:${POWERVSI_2}" \
jumbo=${JUMBO} \
region=${CLOUD_REGION} \
subnet_name=${ERS_NET} \
route_table=5${ERS_INSTNO} \
op start timeout=720 \
op stop timeout=300 \
op monitor interval=60 timeout=30 \
--group ${sid}_ers${ERS_INSTNO}_group
Verifying the cluster configuration
On NODE1, run the following command to verify the cluster configuration at this stage.
pcs status --full
Sample output:
# pcs status --full
Cluster name: SAP_S01
Status of pacemakerd: 'Pacemaker is running' (last updated 2024-11-20 14:04:05 +01:00)
Cluster Summary:
* Stack: corosync
* Current DC: cl-s01-2 (2) (version 2.1.5-9.el9_2.4-a3f44794f94) - partition with quorum
* Last updated: Wed Nov 20 14:04:06 2024
* Last change: Wed Nov 20 13:51:19 2024 by hacluster via crmd on cl-s01-2
* 2 nodes configured
* 8 resource instances configured
Node List:
* Node cl-s01-1 (1): online, feature set 3.16.2
* Node cl-s01-2 (2): online, feature set 3.16.2
Full List of Resources:
* fence_node1 (stonith:fence_ibm_powervs): Started cl-s01-2
* fence_node2 (stonith:fence_ibm_powervs): Started cl-s01-2
* Clone Set: fs_sapmnt-clone [fs_sapmnt]:
* fs_sapmnt (ocf:heartbeat:Filesystem): Started cl-s01-1
* fs_sapmnt (ocf:heartbeat:Filesystem): Started cl-s01-2
* Resource Group: s01_ascs21_group:
* s01_fs_ascs21 (ocf:heartbeat:Filesystem): Started cl-s01-1
* s01_vip_ascs21 (ocf:heartbeat:powervs-subnet): Started cl-s01-1
* Resource Group: s01_ers22_group:
* s01_fs_ers22 (ocf:heartbeat:Filesystem): Started cl-s01-1
* s01_vip_ers22 (ocf:heartbeat:powervs-subnet): Started cl-s01-1
Migration Summary:
Tickets:
PCSD Status:
cl-s01-1: Online
cl-s01-2: Online
Daemon Status:
corosync: active/disabled
pacemaker: active/disabled
pcsd: active/enabled
Make sure that the ${sid}_ascs${ASCS_INSTNO}_group
cluster resource group runs on NODE1 and the ${sid}_ers${ERS_INSTNO}_group
cluster resource group runs on NODE2. If necessary, use the pcs resource move <resource_group_name>
command to move the resource group to the correct node.
Changing the ownership of the ASCS and ERS mount points
The ASCS and ERS mount points must be owned by the sidadm user. You must define the required users and groups and set the mount point ownership before you can start the instance installation.
On both nodes, use the following steps to set the required owner.
-
Start the SAP Software Provisioning Manager (SWPM) to create the operating system users and groups.
<swpm>/sapinst
In the SWPM web interface, use the path System Rename > Preparations > Operating System Users and Group. Note the user and group IDs and make sure that they are the same on both nodes.
-
Change the ownership of the mount points.
chown -R ${sid}adm:sapsys /sapmnt/${SID} /usr/sap/${SID}
Installing the ASCS and ERS instances
Use SWPM to install both instances.
-
Install ASCS and ERS instances on the cluster nodes.
- On NODE1, use the virtual hostname
${ASCS_VH}
that is associated with the ASCS virtual IP address and install an ASCS instance.
<swpm>/sapinst SAPINST_USE_HOSTNAME=${ASCS_VH}
- On NODE2, use the virtual hostname
${ERS_VH}
that is associated with the ERS virtual IP address and install an ERS instance.
<swpm>/sapinst SAPINST_USE_HOSTNAME=${ERS_VH}
- On NODE1, use the virtual hostname
-
Install all other SAP application instances outside the cluster.
Preparing the ASCS and ERS instances for cluster integration
Use the following steps to prepare the SAP instances for the cluster integration.
Disabling the automatic start of the SAP instance agents for ASCS and ERS
You must disable the automatic start of the sapstartsrv
instance agents for both ASCS and ERS instances after a reboot.
Verifying the SAP instance agent integration type
Recent versions of the SAP instance agent sapstartsrv
provide native systemd
support on Linux. For more information, refer to the the SAP notes that are listed at SAP Notes.
On both nodes, check the content of the /usr/sap/sapservices
file.
cat /usr/sap/sapservices
In the systemd
format, the lines start with systemctl
entries.
Example:
systemctl --no-ask-password start SAPS01_01 # sapstartsrv pf=/usr/sap/S01/SYS/profile/S01_ASCS01_cl-sap-scs
If the entries for ASCS and ERS are in systemd
format, continue with the steps in Disabling systemd services of the ASCS and the ERS SAP instance.
In the classic
format, the lines start with LD_LIBRARY_PATH
entries.
Example:
LD_LIBRARY_PATH=/usr/sap/S01/ASCS01/exe:$LD_LIBRARY_PATH;export LD_LIBRARY_PATH;/usr/sap/S01/ASCS01/exe/sapstartsrv pf=/usr/sap/S01/SYS/profile/S01_ASCS01_cl-sap-scs -D -u s01adm
If the entries for ASCS and ERS are in classic
format, then modify the /usr/sap/sapservices
file to prevent the automatic start of the sapstartsrv
instance agent for both ASCS and ERS instances after a reboot.
On both nodes, remove or comment out the sapstartsrv
entries for both ASCS and ERS in the SAP services file.
sed -i -e 's/^LD_LIBRARY_PATH=/#LD_LIBRARY_PATH=/' /usr/sap/sapservices
Example:
#LD_LIBRARY_PATH=/usr/sap/S01/ASCS01/exe:$LD_LIBRARY_PATH;export LD_LIBRARY_PATH;/usr/sap/S01/ASCS01/exe/sapstartsrv pf=/usr/sap/S01/SYS/profile/S01_ASCS01_cl-sap-scs -D -u s01adm
Proceed to Installing permanent SAP license keys.
Disabling systemd services of the ASCS and the ERS instances
On both nodes, disable the instance agent for the ASCS.
systemctl disable --now SAP${SID}_${ASCS_INSTNO}.service
On both nodes, disable the instance agent for the ERS.
systemctl disable --now SAP${SID}_${ERS_INSTNO}.service
Disabling systemd
restart of a crashed ASCS or ERS instance
Systemd
has its own mechanisms for restarting a crashed service. In a high availability setup, only the HA cluster is responsible for managing the SAP ASCS and ERS instances. Create systemd drop-in files
on both
cluster nodes to prevent systemd
from restarting a crashed SAP instance.
On both nodes, create the directories for the drop-in files.
mkdir /etc/systemd/system/SAP${SID}_${ASCS_INSTNO}.service.d
mkdir /etc/systemd/system/SAP${SID}_${ERS_INSTNO}.service.d
On both nodes, create the drop-in files for ASCS and ERS.
cat >> /etc/systemd/system/SAP${SID}_${ASCS_INSTNO}.service.d/HA.conf << EOT
[Service]
Restart=no
EOT
cat >> /etc/systemd/system/SAP${SID}_${ERS_INSTNO}.service.d/HA.conf << EOT
[Service]
Restart=no
EOT
Restart=no
must be in the [Service]
section, and the drop-in files must be available on all cluster nodes.
On both nodes, reload the systemd
unit files.
systemctl daemon-reload
Installing permanent SAP license keys
When the SAP ASCS instance is installed on a Power Virtual Server instance, the SAP license mechanism relies on the partition UUID. For more information, see SAP note 2879336 - Hardware key based on unique ID.
On both nodes, run the following command as user <sid>adm
to identify the HARDWARE KEY
of the node.
sudo -i -u ${sid}adm -- sh -c 'saplikey -get'
Sample output:
$ sudo -i -u ${sid}adm -- sh -c 'saplikey -get'
saplikey: HARDWARE KEY = H1428224519
Note the HARDWARE KEY
of each node.
You need both hardware keys to request two different SAP license keys. Check the following SAP notes for more information about requesting SAP license keys:
Installing SAP resource agents
Install the required software packages. The resource-agents-sap
includes the SAPInstance cluster resource agent for managing the SAP instances.
Unless sap_cluster_connector
is configured for the SAP instance, the RHEL HA Add-On cluster considers any state change of the instance as an issue. If other SAP tools such as sapcontrol
are used to manage the instance,
then sap_cluster_connector
grants permission to control SAP instances that are running inside the cluster. If the SAP instances are managed by only cluster tools, the implementation of sap_cluster_connector
is not
necessary.
Install the packages for the resource agent and the SAP Cluster Connector library. For more information, see How to enable the SAP HA Interface for SAP ABAP application server instances managed by the RHEL HA Add-On
On both nodes, run the following commands.
If needed, use subscription-manager
to enable the SAP NetWeaver repository. The RHEL for SAP Subscriptions and Repositories documentation describes how to enable the required repositories.
subscription-manager repos --enable="rhel-8-for-ppc64le-sap-netweaver-e4s-rpms"
Install the required packages.
dnf install -y resource-agents-sap sap-cluster-connector
Configuring SAP Cluster Connector
Add user ${sid}adm
to the haclient
group.
On both nodes, run the following command.
usermod -a -G haclient ${sid}adm
Adapting the SAP instance profiles
Modify the start profiles of all SAP instances that are managed by SAP tools outside the cluster. Both ASCS and ERS instances can be controlled by the RHEL HA Add-On cluster and its resource agents. Adjust the SAP instance profiles to prevent an automatic restart of instance processes.
On NODE1, navigate to the SAP profile directory.
cd /sapmnt/${SID}/profile
Change all occurrences of Restart_Program
to Start_Program
in the instance profile of both ASCS and ERS.
sed -i -e 's/Restart_Program_\([0-9][0-9]\)/Start_Program_\1/' ${SID}_ASCS${ASCS_INSTNO}_${ASCS_VH}
sed -i -e 's/Restart_Program_\([0-9][0-9]\)/Start_Program_\1/' ${SID}_ERS${ERS_INSTNO}_${ERS_VH}
Add the following two lines at the end of the SAP instance profile to configure sap_cluster_connector
for the ASCS and ERS instances.
service/halib = $(DIR_EXECUTABLE)/saphascriptco.so
service/halib_cluster_connector = /usr/bin/sap_cluster_connector
Configuring the ASCS and ERS cluster resources
Up to this point, the following are assumed:
- A RHEL HA Add-On cluster is running on both virtual server instances and node fencing has been tested.
- A cloned Filesystem cluster resource is configured to mount the sapmnt share.
- Two Filesystem cluster resources are configured to mount the ASCS and ERS instance file systems.
- Two powervs-subnet cluster resources are configured for the virtual IP addresses of the ASCS and ERS instances.
- The ASCS instance is installed and active on NODE1.
- THE ERS instance is installed and active on NODE2.
- All steps according to Prepare ASCS and ERS instances for the cluster integration are complete.
Configuring the ASCS cluster resource group
On NODE1, run the following commands to create a cluster resource for managing the ASCS instance.
pcs resource create ${sid}_ascs${ASCS_INSTNO} SAPInstance \
InstanceName="${SID}_ASCS${ASCS_INSTNO}_${ASCS_VH}" \
START_PROFILE=/sapmnt/${SID}/profile/${SID}_ASCS${ASCS_INSTNO}_${ASCS_VH} \
AUTOMATIC_RECOVER=false \
meta resource-stickiness=5000 \
migration-threshold=1 failure-timeout=60 \
op monitor interval=20 on-fail=restart timeout=60 \
op start interval=0 timeout=600 \
op stop interval=0 timeout=600 \
--group ${sid}_ascs${ASCS_INSTNO}_group
The meta resource-stickiness=5000
option is used to balance the failover constraint with ERS so that the resource stays on the node where it started and doesn't migrate uncontrollably in the cluster.
Add a resource stickiness to the group to ensure that the ASCS instance stays on the node.
pcs resource meta ${sid}_ascs${ASCS_INSTNO}_group \
resource-stickiness=3000
Configuring the ERS cluster resource group
On NODE2, run the following command to create a resource for managing the ERS instance.
pcs resource create ${sid}_ers${ERS_INSTNO} SAPInstance \
InstanceName="${SID}_ERS${ERS_INSTNO}_${ERS_VH}" \
START_PROFILE=/sapmnt/${SID}/profile/${SID}_ERS${ERS_INSTNO}_${ERS_VH} \
AUTOMATIC_RECOVER=false \
IS_ERS=true \
op monitor interval=20 on-fail=restart timeout=60 \
op start interval=0 timeout=600 \
op stop interval=0 timeout=600 \
--group ${sid}_ers${ERS_INSTNO}_group
Configuring the cluster constraints
On NODE1, run the following command to create the cluster constraints.
A colocation constraint prevents resource groups ${sid}_ascs${ASCS_INSTNO}_group
and ${sid}_ers${ERS_INSTNO}_group
from being active on the same node whenever possible. If only a single node is available, the stickiness
value of -5000
ensures that they run on the same node.
pcs constraint colocation add \
${sid}_ers${ERS_INSTNO}_group with ${sid}_ascs${ASCS_INSTNO}_group -- -5000
An order constraint controls that ${sid}_ascs${ASCS_INSTNO}_group
starts before ${sid}_ers${ERS_INSTNO}_group
.
pcs constraint order start \
${sid}_ascs${ASCS_INSTNO}_group then stop ${sid}_ers${ERS_INSTNO}_group \
symmetrical=false \
kind=Optional
The following two order constraints ensure that the SAPMNT file system mounts before ${sid}_ascs${ASCS_INSTNO}_group
and ${sid}_ers${ERS_INSTNO}_group
start.
pcs constraint order fs_sapmnt-clone then ${sid}_ascs${ASCS_INSTNO}_group
pcs constraint order fs_sapmnt-clone then ${sid}_ers${ERS_INSTNO}_group
Conclusion
This completes the ENSA2 cluster implementation in a multizone region environemnt.
You should now proceed with testing the cluster, similar to the tests described in Testing an SAP ENSA2 cluster.
The following is a sample output of the pcs status
command for a completed ENSA2 cluster in a multi-zone region implementation.
Cluster name: SAP_S01
Status of pacemakerd: 'Pacemaker is running' (last updated 2024-11-22 09:42:15 +01:00)
Cluster Summary:
* Stack: corosync
* Current DC: cl-s01-1 (version 2.1.5-9.el9_2.4-a3f44794f94) - partition with quorum
* Last updated: Fri Nov 22 09:42:15 2024
* Last change: Fri Nov 22 09:06:18 2024 by root via cibadmin on cl-s01-1
* 2 nodes configured
* 10 resource instances configured
Node List:
* Online: [ cl-s01-1 cl-s01-2 ]
Full List of Resources:
* fence_node1 (stonith:fence_ibm_powervs): Started cl-s01-1
* fence_node2 (stonith:fence_ibm_powervs): Started cl-s01-2
* Clone Set: fs_sapmnt-clone [fs_sapmnt]:
* Started: [ cl-s01-1 cl-s01-2 ]
* Resource Group: s01_ascs21_group:
* s01_fs_ascs21 (ocf:heartbeat:Filesystem): Started cl-s01-1
* s01_vip_ascs21 (ocf:heartbeat:powervs-subnet): Started cl-s01-1
* s01_ascs21 (ocf:heartbeat:SAPInstance): Started cl-s01-1
* Resource Group: s01_ers22_group:
* s01_fs_ers22 (ocf:heartbeat:Filesystem): Started cl-s01-2
* s01_vip_ers22 (ocf:heartbeat:powervs-subnet): Started cl-s01-2
* s01_ers22 (ocf:heartbeat:SAPInstance): Started cl-s01-2
Daemon Status:
corosync: active/disabled
pacemaker: active/disabled
pcsd: active/enabled