Top
PRIMECLUSTER  Installation and Administration Guide 4.4
FUJITSU Software

5.1.2 Setting up the Shutdown Facility

This section describes the setup procedure of the shutdown facility for the PRIMERGY, PRIMEQUEST, and virtual machine environment (KVM environment).

The setup procedure for the shutdown facility is different depending on the model/configuration.

The following shows the shutdown agents required for each hardware model/configuration.

Table 5.2 Shutdown agent for PRIMERGY

Server model

IPMI
(SA_ipmi)

Blade
(SA_blade)

kdump
(SA_lkcd)

RX series
TX series

Y

-

Y

BX series
(For use in combination with ServerView Resource Orchestrator Virtual Edition)

Y (*1)

-

Y

BX series
(For not use in combination with ServerView Resource Orchestrator Virtual Edition)

-

Y

Y

(*1) The combination of user and password for BMC or iRMC that is used in the shutdown facility must be the same on all blades.

Table 5.3 Shutdown agent for PRIMEQUEST

Server model

MMB

Panic
(SA_mmbp)

Reset
(SA_mmbr)

PRIMEQUEST

1000/2000 series

Y

Y

Table 5.4 Shutdown agent for virtual machine environment (KVM environment)

Server model


Cluster configuration

libvirt

Panic
(SA_libvirtgp)

Reset
(SA_libvirtgr)

PRIMERGY

Guests in a unit

Y

Y

Guests between units

Y

Y

PRIMEQUEST

1000/2000 series

Guests in a unit

Y

Y

Guests between units

Y

Y

When using the host OS failover function in virtual machine environment (KVM environment), set the following shutdown agents. The shutdown agent that are set on the guest OS are the same as those used in the virtual machine function.

Table 5.5 Shutdown agent for the host OS failover function in virtual machine environment (KVM environment)

Server model


Cluster configuration


Cluster node

Shutdown agent for the server model

libvirt

vmchkhost

Panic
(SA_libvirtgp)

Reset
(SA_libvirtgr)

Checking the status
(SA_vmchkhost)

PRIMERGY

Guests between units

Host OS

Y
(See Table 5.2)

-

-

-

Guest OS

-

Y

Y

Y

PRIMEQUEST
1000/2000
series

Guests between units

Host OS

Y
(See Table 5.3)

-

-

-

Guest OS

-

Y

Y

Y

See

For details on the shutdown facility, see the following manuals:

  1. "2.3.1.7 PRIMECLUSTER SF" in "PRIMECLUSTER Concepts Guide"

  2. "Chapter 7 Shutdown Facility" in "PRIMECLUSTER Cluster Foundation (CF) Configuration and Administration Guide"

5.1.2.1 Survival Priority

If the cluster partition occurred due to a fault in the cluster interconnect, all the nodes would still be in the state of accessing the user resources. For details on the cluster partition, see "1.2.2.1 Protecting data integrity" in "PRIMECLUSTER Concepts Guide."

In order to guarantee the data consistency in the user resources, SF must determine the node groups of which nodes remain to survive and which nodes need to be forcibly stopped.

The weight assigned to each node group is referred to as "Survival priority" in PRIMECLUSTER.

The greater the weight of the node, the higher the survival priority. Conversely, the less the weight of the node, the lower the survival priority. If the multiple node groups have the same survival priority, the node group that includes the node with the alphabetical earliest node name will survive.

Survival priority can be calculated based on the following formula:

Survival priority = SF node weight + ShutdownPriority of userApplication

Note

When SF calculates the survival priority, each node will send its survival priority to the remote node via the administrative LAN. If any communication problem of the administrative LAN occurs, the survival priority will not be able to reach. In this case, the survival priority will be calculated only by the SF node weight.

SF node weight (Weight):

Weight of node. Default value = 1. Set this value while configuring the shutdown facility.

userApplication ShutdownPriority:

Set this attribute when userApplication is created. For details on how to change the settings, see "11.1 Changing the Operation Attributes of a userApplication."

See

For details on the ShutdownPriority attribute of userApplication, see "12.1 Attributes available to the user" in "PRIMECLUSTER Reliant Monitor Services (RMS) with Wizard Tools Configuration and Administration Guide."

Survival scenarios

The typical scenarios that are implemented are shown below:

[Largest node group survival]
  • Set the weight of all the nodes to 1 (default).

  • Set the ShutdownPriority attribute of every userApplication to 0 (default).

[Specific node survival]
  • Set the "weight" of the node to survive to a value more than double the total weight of the other nodes.

  • Set the ShutdownPriority attribute of every userApplication to 0 (default).

In the following example, node1 is to survive:

[Specific userApplication survival]
  • Set the "weight" of all the nodes to 1 (default).

  • Set the ShutdownPriority attribute of userApplication whose operation is to continue to a value more than double the total of the ShutdownPriority attributes of the other user applications and the weights of all the nodes.

  • Set the ShutdownPriority attribute within the range of 1 to 20.


In the following example, the node for which app1 is operating is to survive:

[Node survival in a specific order of node]
  • Set the "weight" of the node to survive to a value more than double the total weight of the other nodes which have lower priority.

  • Set the ShutdownPriority attribute of every userApplication to 0 (default).

In the following example, node1, node2, node3, and node4 are to survive in this order:

[Node survival in a specific userApplication order]
  • Set the "weight" of all the nodes to 1 (default).

  • Set the value that is power of 2 (1, 2, 4, 8, 16, ...) to the ShutdownPriority attribute of userApplication if its operation must be continued.

  • Calculate the minimum value to be set to the ShutdownPriority attribute using the following formula. The value must be power of 2 (1, 2, 4, 8, 16, ...) and equal to or larger than the calculated value.

    The number of configuration node - 1
    Example: In 2-node configuration, (2 - 1) = 1. The minimum settable value to ShutdownPriority attribute is 1.
    Example: In 3-node configuration, (3 - 1) = 2. The minimum settable value to ShutdownPriority attribute is 2.
    Example: In 4-node configuration, (4 - 1) = 3. The minimum settable value to ShutdownPriority attribute is 4.

The following example shows the survival priority of nodes on which userApplication runs. Sequentially app1, app2, and app3 are prioritized.

[Host OS failover function]
  • Set the "weight" of nodes to a power-of-two value (1,2,4,8,16,...) in ascending order of survival priority in each cluster system.

  • The "weight" set to a guest OS should have the same order relation with a corresponding host OS.

    For example, when setting a higher survival priority to host1 than host2 between host OSes, set a higher survival priority to node1 (corresponding to host1) than node2-4 (corresponding to host2) between guest OSes.

  • Set the ShutdownPriority attribute of every userApplication to 0 (default).

In the following example, node1, node2, node3, and node4 are to survive in this order:

5.1.2.2 Setup Flow for Shutdown Facility

5.1.2.2.1 Setup Flow in PRIMERGY RX/TX Series

For the setup flow for the shutdown facility in PRIMERGY RX/TX series, take the following steps.

  1. Checking the shutdown agent information

  2. Setting up the shutdown daemon

  3. Configuring the IPMI shutdown agent

  4. Configuring the kdump shutdown agent

  5. Starting up the shutdown facility

  6. Test for forced shutdown of cluster nodes

For the detailed setup procedure, refer to "5.1.2.3 Setup Procedure for Shutdown Facility in PRIMERGY."

5.1.2.2.2 Setup Flow in PRIMERGY BX Series

When using in combination with ServerView Resource Orchestrator Virtual Edition

When using in combination with ServerView Resource Orchestrator Virtual Edition, for the setup flow for the shutdown facility in PRIMERGY BX series, take the following steps.

  1. Checking the shutdown agent information

  2. Setting up the shutdown daemon

  3. Configuring the IPMI shutdown agent

  4. Configuring the kdump shutdown agent

  5. Starting up the shutdown facility

  6. Test for forced shutdown of cluster nodes

For the detailed setup procedure, refer to "5.1.2.3 Setup Procedure for Shutdown Facility in PRIMERGY."

When not using in combination with ServerView Resource Orchestrator Virtual Edition

When not using in combination with ServerView Resource Orchestrator Virtual Edition, for the setup flow for the shutdown facility in PRIMERGY BX series, take the following steps.

  1. Checking the shutdown agent information

  2. Setting up the shutdown daemon

  3. Configuring the Blade shutdown agent

  4. Configuring the kdump shutdown agent

  5. Starting up the shutdown facility

  6. Test for forced shutdown of cluster nodes

For the detailed setup procedure, refer to "5.1.2.3 Setup Procedure for Shutdown Facility in PRIMERGY."

5.1.2.2.3 Setup Flow in PRIMEQUEST

For the setup flow for the shutdown facility in PRIMEQUEST, take the following steps.

  1. Checking the shutdown agent information

  2. Configuring the MMB shutdown agent

  3. Setting up the shutdown daemon

  4. Starting the MMB asynchronous monitoring daemon

  5. Setting the I/O completion wait time(for using other than ETERNUS disk array as the shared disk)

  6. Starting up the shutdown facility

  7. Test for forced shutdown of cluster nodes

For the detailed setup procedure, refer to "5.1.2.4 Setup Procedure for Shutdown Facility in PRIMEQUEST."

5.1.2.2.4 Setup Flow in Virtual Machine Environment

When using the host OS failover function

When using the host OS failover function in a virtual machine environment, for the setup flow for the shutdown facility, take the following steps.

  1. Setting up the shutdown facility on the host OS in PRIMERGY/PRIMEQUEST

  2. Checking the shutdown agent information in a virtual machine environment

  3. Configuring the libvirt shutdown agent

  4. Configuring the vmchkhost shutdown agent

  5. Starting up the shutdown facility

  6. Setting up the host OS failover function on the host OS (PRIMEQUEST only)

  7. Test for forced shutdown of cluster nodes

For the detailed setup procedure, see the following.

5.1.2.3 Setup Procedure for Shutdown Facility in PRIMERGY

5.1.2.4 Setup Procedure for Shutdown Facility in PRIMEQUEST

5.1.2.5 Setup Procedure for Shutdown Facility in Virtual Machine Environment

When not using the host OS failover function

When not using the host OS failover function in a virtual machine environment, for setup flow for the shutdown facility, take the following steps.

  1. Checking the shutdown agent information in a virtual machine environment

  2. Configuring the libvirt shutdown agent

  3. Starting up the shutdown facility

  4. Test for forced shutdown of cluster nodes

For the detailed setup procedure, refer to "5.1.2.5 Setup Procedure for Shutdown Facility in Virtual Machine Environment."

5.1.2.3 Setup Procedure for Shutdown Facility in PRIMERGY

This section describes the procedure for setting up the shutdown agent in PRIMERGY.

For shutdown facility, set the supported server type to be used.

Note

When creating a redundant administrative LAN used in the shutdown facility by using GLS, set as below.

  • For taking over the IP address between nodes

    Configure GLS by using the logical IP address takeover function of the NIC switching mode.

    For shutdown facility, specify a physical IP address instead of a logical IP address.

  • For not taking over the IP address between nodes

    Configure GLS by using the physical IP address takeover function of the NIC switching mode.

5.1.2.3.1 Checking the Shutdown Agent Information

RX/TX series

Check the following settings in BMC (Baseboard Management Controller) or iRMC (integrated Remote Management Controller) necessary for setting the IPMI shutdown agent.

Also, check the following.

BX series (When using in combination with ServerView Resource Orchestrator Virtual Edition)

Necessary settings are the same as the settings of RX/TX series. Refer to RX/TX series.

BX series (When not using in combination with ServerView Resource Orchestrator Virtual Edition)

Check the following settings for the management blade necessary for setting the Blade shutdown agent.

5.1.2.3.2 Setting up the Shutdown Daemon

Create /etc/opt/SMAW/SMAWsf/rcsd.cfg on all the nodes as shown below.

Create the rcsd.cfg file by a root user and change the permission to 600.

RX/TX series

CFNameX,weight=weight,admIP=myadmIP:agent=SA_ipmi,timeout=timeout
CFNameX,weight=weight,admIP=myadmIP:agent=SA_ipmi,timeout=timeout
CFNameX          : Specify the CF node name of the cluster host.
weight           : Specify the weight of the SF node.
myadmIP          : Specify the IP address of the administrative LAN
                   used in the shutdown facility of the cluster host.
                   It is not the IP address of iRMC or the management blade.
                   Available IP addresses are IPv4 and IPv6 addresses.
                   IPv6 link local addresses are not available.
                   When specifying an IPv6 address, enclose it in brackets "[ ]".
                   (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
                   If you specify a host name, make sure it is listed in /etc/hosts.
SA_ipmi          : Specify the IPMI shutdown agent.
timeout          : Specify the timeout duration (seconds) of the IPMI shutdown agent.    
                   For the IPMI shutdown agent, specify 25 seconds.

Example:

node1,weight=1,admIP=10.20.30.100:agent=SA_ipmi,timeout=25
node2,weight=1,admIP=10.20.30.101:agent=SA_ipmi,timeout=25

BX series (When using in combination with ServerView Resource Orchestrator Virtual Edition)

Necessary settings are the same as the settings of RX/TX series. Refer to RX/TX series.

BX series (When not using in combination with ServerView Resource Orchestrator Virtual Edition)

CFNameX,weight=weight,admIP=myadmIP:agent=SA_blade,timeout=timeout
CFNameX,weight=weight,admIP=myadmIP:agent=SA_blade,timeout=timeout
CFNameX          : Specify the CF node name of the cluster host.
weight           : Specify the weight of the SF node.
myadmIP          : Specify the IP address of the administrative LAN
                   used in the shutdown facility of the cluster host.
                   It is not the IP address of iRMC or the management blade.
                   Available IP addresses are IPv4 and IPv6 addresses.
                   IPv6 link local addresses are not available.
                   When specifying an IPv6 address, enclose it in brackets "[ ]".
                   (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
                   If you specify a host name, make sure it is listed in /etc/hosts.
SA_blade         : Specify the Blade shutdown agent.
timeout          : Specify the timeout duration (seconds) of the Blade shutdown agent.    
                   For the Blade shutdown agent, specify 20 seconds.

Example:

node1,weight=1,admIP=10.20.30.100:agent=SA_blade,timeout=20
node2,weight=1,admIP=10.20.30.101:agent=SA_blade,timeout=20

Note

  • For using STP (Spanning Tree Protocol) in the administrative LAN used in the shutdown facility, it is necessary to set the timeout value to the current value plus (+) 50 (seconds), taking into account the time STP needs to create the tree and an extra cushion. This setting increases the time required for failover.

  • The contents of the rcsd.cfg file must be the same on all the nodes. If different, it does not work.

Information

When the "/etc/opt/SMAW/SMAWsf/rcsd.cfg" file is to be created, the "/etc/opt/SMAW/SMAWsf/rcsd.cfg.template" file can be used as a prototype.

5.1.2.3.3 Setting up IPMI Shutdown Agent

In RX/TX series, or when using in combination with ServerView Resource Orchestrator Virtual Edition in BX series, for the server with the BMC (Baseboard Management Controller) or iRMC (integrated Remote Management Controller) installed, configure the IPMI shutdown agent.

You must configure the IPMI shutdown agent before you configure the kdump shutdown agent.

  1. Starting the IPMI service

    [RHEL6]

    Execute the following command on all the nodes to check the startup status of the IPMI service.

    # /sbin/service ipmi status

    Execute the following command on all the nodes in which the IPMI service is not activated to start the IPMI service.

    # /sbin/service ipmi start
    Starting ipmi drivers: [ OK ]

    [RHEL7]

    Execute the following command on all the nodes to check the startup status of the IPMI service.

    # /usr/bin/systemctl status ipmi.service
    ipmi.service - IPMI Driver
    Loaded: loaded (/usr/lib/systemd/system/ipmi.service; disabled)
    Active: inactive (dead)

    Execute the following command on all the nodes in which the "inactive" appeared in the "Active:" field and the IPMI service is not activated to start the IPMI service.

    # /usr/bin/systemctl start ipmi.service
  2. Setting the run level of the IPMI service

    [RHEL6]

    Execute the following command on all the nodes to read the IPMI service on startup.

    # /sbin/chkconfig --level 2345 ipmi on

    [RHEL7]

    Execute the following command on all the nodes to read the IPMI service on startup.

    # /usr/bin/systemctl enable ipmi.service
  3. Encrypting the password

    Execute the sfcipher command to encrypt passwords of a user for the shutdown facility.

    Example: If the password specified when setting up IPMI (BMC and iRMC) is "bmcpwd$"

    # sfcipher -c
    Enter User's Password:  <- enter bmcpwd$ 
    Re-enter User's Password:  <- enter bmcpwd$ 
    /t1hXYb/Wno=

    Note: It is not necessary to insert '\' in front of the special characters specified as the password in 4.3A30 or later.

    For information on how to use the sfcipher command, see the "sfcipher" manual page.

    Note

    For the passwords specified when setting up IPMI (BMC and iRMC), seven-bit ASCII characters are available.
    Among them, do not use the following characters as they may cause a problem.

    >  <  "  /  \  =  !  ?  ;  ,  &
  4. Setting the shutdown agent

    Create /etc/opt/SMAW/SMAWsf/SA_ipmi.cfg on all the nodes as shown below.

    Create the SA_ipmi.cfg file by a root user and change the permission to 600.

    • For IPv4 address

      CFName1  ip-address:user:passwd {cycle | leave-off} 
      CFName2  ip-address:user:passwd {cycle | leave-off} 
    • For IPv6 address

      CFName1 [ip-address]:user:passwd {cycle | leave-off} 
      CFName2 [ip-address]:user:passwd {cycle | leave-off} 
    CFNameX          : Specify the CF node name of the cluster host.
    ip-address       : Specify the Ip address for IPMI (BMC or iRMC)
                       in the server where a cluster host is operating.
                       Available IP addresses are IPv4 and IPv6 addresses.
                       IPv6 link local addresses are not available.
                       When specifying the IPv6 address, enclose it in brackets "[ ]".
                       (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
    user             : Specify the user defined when IPMI (BMC or iRMC) was setup.
    passwd           : Password defined when IPMI (BMC or iRMC) was setup.
                       Specify the password encrypted in step 3.
    cycle            : Reboot the node after forcibly stopping the node.
    leave-off        : Power-off the node after forcibly stopping the node.
    

    Example 1:

    When the IP address of iRMC of node1 is 10.20.30.50 and the IP address of iRMC of node2 is 10.20.30.51

    node1  10.20.30.50:root:/t1hXYb/Wno= cycle
    node2 10.20.30.51:root:/t1hXYb/Wno= cycle

    Example 2:

    When the IP address of iRMC of node1 is 1080:2090:30a0:40b0:50c0:60d0:70e0:80f0 and the IP address of iRMC of node2 is 1080:2090:30a0:40b0:50c0:60d0:70e0:80f1

    node1 [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0]:root:/t1hXYb/Wno= cycle
    node2 [1080:2090:30a0:40b0:50c0:60d0:70e0:80f1]:root:/t1hXYb/Wno= cycle

    Information

    When the "/etc/opt/SMAW/SMAWsf/SA_ipmi.cfg" file is to be created, the "/etc/opt/SMAW/SMAWsf/SA_ipmi.cfg.template" file can be used as a prototype.

    Note

    • Check if the setting contents of the /etc/opt/SMAW/SMAWsf/SA_ipmi.cfg file are correct. If there is an error in the setting contents, the shutdown facility cannot be performed normally.

    • Check if the IP address (ip-address) of IPMI (BMC or iRMC) corresponding to the cluster host's CF node name (CFNameX) of the /etc/opt/SMAW/SMAWsf/SA_ipmi.cfg file are set. If there is an error in the setting, a different node may be forcibly stopped.

    • The contents of the SA_ipmi.cfg file must be the same on all the nodes. If different, it does not work.

5.1.2.3.4 Setting up Blade Shutdown Agent

When not using in combination with ServerView Resource Orchestrator Virtual Edition in BX series, configure the Blade shutdown agent. You must configure the Blade shutdown agent before you configure the kdump shutdown agent.

Create /etc/opt/SMAW/SMAWsf/SA_blade.cfg on all the nodes as shown below.

Create SA_blade.cfg file by a root user and change the permission to 600.

Cluster configuration within a single chassis
management-blade-ip IPaddress 
community-string SNMPcommunity
CFName1 slot-no {cycle | leave-off}
CFName2 slot-no
{cycle | leave-off}
IPaddress          : Specify the IP address of the management blade.
                     Available IP addresses are IPv4 and IPv6 addresses.
                     IPv6 link local addresses are not available.
                     When specifying the IPv6 address, enclose it in brackets "[ ]".
                     (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
SNMPcommunity      : Specify the SNMP community of the management blade.
CFNameX            : Specify the CF node name of the cluster host.
slot-no            : Specify the slot No. of the server blade where a cluster
                     host is operating.
cycle              : Reboot the node after forcibly stopping the node.
leave-off          : Power-off the node after forcibly stopping the node.

Example :

When the IP address of the management blade of node1 and node2 is 10.20.30.50, the slot number of node1 is 1 and the slot number of node2 is 2.

management-blade-ip 10.20.30.50
community-string public
node1 1 cycle
node2 2 cycle
Cluster configuration across multiple chassis
community-string SNMPcommunity
management-blade-ip IPaddress1 
CFName1 slot-no {cycle | leave-off}
management-blade-ip IPaddress2
CFName2 slot-no
{cycle | leave-off}
IPaddressX         : Specify the IP address of the management blade
                     in a chassis where a cluster host of CFNameX exists.
                     Available IP addresses are IPv4 and IPv6 addresses.
                     IPv6 link local addresses are not available.
                     When specifying the IPv6 address, enclose it in brackets "[ ]".
                     (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
                     Make sure to write it before CFNameX.
SNMPcommunity      : Specify the SNMP community of the management blade.
CFNameX            : Specify the CF node name of the cluster host.
slot-no            : Specify the slot No. of the server blade
                     where a cluster host is operating.
cycle              : Reboot the node after forcibly stopping the node.
leave-off          : Power-off the node after forcibly stopping the node.

Note

SNMP community name of the management blade must be same in all the chassis.

Example:

When the IP address of the management blade of node1 is 10.20.30.50, and the slot number of node1 is 1.
Moreover, when the IP address of the management blade of node2 is 10.20.30.51, and the slot number of node2 is 2.

community-string public
management-blade-ip  10.20.30.50
node1 1 cycle
management-blade-ip  10.20.30.51
node2 2 cycle

Information

When the "/etc/opt/SMAW/SMAWsf/SA_blade.cfg" file is to be created, the "/etc/opt/SMAW/SMAWsf/SA_blade.cfg.template" file can be used as a prototype.

Note

  • Check if the setting contents of the /etc/opt/SMAW/SMAWsf/SA_blade.cfg file are correct. If there is an error in the setting contents, the shutdown facility cannot be performed normally.

  • Check if the IP address (IPaddress) of the management blade and the slot number (slot-no) of the server blade corresponding to the cluster host's CF node name (CFNameX) of the /etc/opt/SMAW/SMAWsf/SA_blade.cfg file are set. If there is an error in the setting, a different node may be forcibly stopped.

  • The contents of SA_blade.cfg file must be same on all the nodes. If different, it does not work.

5.1.2.3.5 Setting up kdump Shutdown Agent

Configure the kdump shutdown agent after configuring IPMI shutdown agent or Blade shutdown agent.

Perform the following procedures.

  1. Initializing the configuration file for the kdump

    Execute the following command on any one of the cluster nodes.

    # /etc/opt/FJSVcllkcd/bin/panicinfo_setup

    If the following message is output, the setting file (rcsd.cfg) of the shutdown daemon has an error. Correct the file.

    panicinfo_setup: ERROR: Reading the Shutdown Facility configuration failed.

    If the following message is output, the setting file (SA_ipmi.cfg or SA_blade.cfg) of the shutdown agent has an error. Correct the file.

    panicinfo_setup: ERROR: Reading the Shutdown Agent configuration failed.

    In the environment where panicinfo_setup has already been executed, the following massage is output.

    panicinfo_setup: WARNING: /etc/panicinfo.conf file already exists.
    (I)nitialize, (C)opy or (Q)uit (I/C/Q) ?

    In the case, input "I".

    Note

    To execute the command, CF and CF services (CFSH and CFCP) must be activated. For details, see "5.1.1 Setting Up CF and CIP."

  2. Setting crash dump collection

    • In RX/TX series, or when using in combination with ServerView Resource Orchestrator Virtual Edition in BX series

      1. Change /etc/opt/FJSVcllkcd/etc/SA_lkcd.tout as follows on all the nodes.

        Before change

        PANICINFO_TIMEOUT 5
        RSB_PANIC 0

        After change

        PANICINFO_TIMEOUT 10
        RSB_PANIC 3
      2. Change the timeout value of SA_lkcd in the /etc/opt/SMAW/SMAWsf/rcsd.cfg file as follows on all the nodes.

        Before change

        agent=SA_lkcd,timeout=20

        After change

        agent=SA_lkcd,timeout=25
    • When not using in combination with ServerView Resource Orchestrator Virtual Edition in BX series

      Change RSB_PANIC of /etc/opt/FJSVcllkcd/etc/SA_lkcd.tout as follows on all the nodes.

      Before change

      RSB_PANIC 0

      After change

      RSB_PANIC 2
5.1.2.3.6 Starting up the Shutdown Facility

Start or restart the shutdown facility on all the nodes.

  1. Starting the shutdown facility

    Check that the shutdown facility has been started on all the nodes.

    # sdtool -s

    If the shutdown facility has already been started, execute the following commands to restart the shutdown facility on all the nodes.

    # sdtool -e
    # sdtool -b

    If the shutdown facility has not been started, execute the following command to start the shutdown facility on all the nodes.

    # sdtool -b
  2. Checking the status of the shutdown facility

    Check the status of the shutdown facility on all the nodes.

    # sdtool -s

Information

Display results of the sdtool -s command
  • If InitFailed is displayed in Init State, it means that a problem occurred during initialization of that shutdown agent.

  • If TestFailed is displayed in Test State, it means that a problem occurred while the agent was testing whether or not the node displayed in the Cluster Host field could be stopped. Some sort of problem probably occurred in the software, hardware, or network resources being used by that agent.

  • If Unknown is displayed in Shut State or Init State, it means that SF has not yet executed node stop, path testing, or SA initialization. Unknown is displayed temporarily in Test State and Init State until the actual status can be confirmed.

  • If TestFailed or InitFailed is displayed, check /var/log/messages. After the failure-causing problem is resolved and SF is restarted, the status display changes to InitWorked or TestWorked.

Note

If TestFailed is displayed in Test State when "sdtool -s" is executed after the shutdown facility was started, possible causes are as follows:
  • The shutdown agent is incorrectly set.

  • The IPMI shutdown agent is used without the user password of the shutdown facility encrypted.

Take the following procedure:

  1. Execute the following command on all the nodes to stop the shutdown facility.

    # sdtool -e
  2. Review the settings of shutdown facility.

  3. Execute the following command on any node to apply changes of the configuration file.

    # /etc/opt/FJSVcllkcd/bin/panicinfo_setup

    After the following message is displayed, select "I."

    panicinfo_setup: WARNING: /etc/panicinfo.conf file already exists.
    (I)nitialize, (C)opy or (Q)uit (I/C/Q) ?
  4. Execute the following command on all the nodes to start the shutdown facility.

    # sdtool -b
  5. Execute the following command on all the nodes and make sure that the shutdown facility operates normally.

    # sdtool -s
5.1.2.3.7 Test for Forced Shutdown of Cluster Nodes

After setting up the shutdown facility, conduct a test for forced shutdown of cluster nodes to check that the correct nodes can be forcibly stopped.

For the detail of the test for forced shutdown of cluster nodes, refer to "1.4 Test."

5.1.2.4 Setup Procedure for Shutdown Facility in PRIMEQUEST

This section describes the setup procedure for the shutdown facility in PRIMEQUEST.

Note

When creating a redundant administrative LAN used in the shutdown facility by using GLS, set as below.

  • For taking over the IP address between nodes

    Configure GLS by using the logical IP address takeover function of the NIC switching mode.

    For shutdown facility, specify a physical IP address instead of a logical IP address.

  • For not taking over the IP address between nodes

    Configure GLS by using the physical IP address takeover function of the NIC switching mode.

5.1.2.4.1 Checking the Shutdown Agent Information

MMB check items

Check the following settings for MMB blade necessary for setting Blade shutdown agent.

Also check that following settings are enabled for the user confirmed above:

Check the settings for the user who uses RMCP to control the MMB. Log in to MMB Web-UI, and check the settings from the "Remote Server Management" window of the "Network Configuration" menu.

If the above settings have not been set, set up the MMB so that the above settings are set.

Note

The MMB units have two types of users:

  • User who controls all MMB units

  • User who uses RMCP to control the MMB

The user to be checked here is the user who uses RMCP to control the MMB. Be sure to check the correct type of user.

Checking the time to wait until I/O to the shared disk is completed (when using other than the ETERNUS disk array as the shared disk)

When using any disks other than the ETERNUS disk array as the shared disk, to prevent the data error when the node is down due to a panic or other causes, set the time until I/O to the shared disk is completed.

To set the wait time, panic the node during I/O to the shared disk. After that, check the time until I/O to the shared disk is completed.

5.1.2.4.2 Setting up the MMB Shutdown Agent

Set up the MMB shutdown agent according to the procedure described below.

Take this procedure after taking the procedure described in "5.1.1 Setting Up CF and CIP."

  1. Execute the "clmmbsetup -a" command on all the nodes, and register the MMB information.

    For instructions on using the "clmmbsetup" command, see the "clmmbsetup" manual page.

    # /etc/opt/FJSVcluster/bin/clmmbsetup -a mmb-user
    Enter User's Password:
    Re-enter User's Password:

    For mmb-user and User's Password, enter the following values that were checked in "5.1.2.4.1 Checking the Shutdown Agent Information."

    mmb-user

    User's name for controlling the MMB with RMCP

    User's Password

    User's password for controlling the MMB with RMCP.

    Note

    For the passwords specified when setting MMB, seven-bit ASCII characters are available.

    Among them, do not use the following characters as they may cause a problem.

    >  <  "  /  \  =  !  ?  ;  ,  &
  2. Execute the "clmmbsetup -l" command on all the nodes, and check the registered MMB information.

    If the registered MMB information was not output on all the nodes in Step 1, start over from Step 1.

    # /etc/opt/FJSVcluster/bin/clmmbsetup -l
    cluster-host-name  user-name
    -----------------------------------
    node1              mmb-user
    node2              mmb-user
5.1.2.4.3 Setting up the Shutdown Daemon

On all the nodes, create /etc/opt/SMAW/SMAWsf/rcsd.cfg with the following information.

Create the rcsd.cfg file using root user access privileges and change the permission of the file to 600.

CFNameX,weight=weight,admIP=myadmIP:agent=SA_mmbp,timeout=timeout:agent=SA_mmbr,timeout=timeout
CFNameX,weight=weight,admIP=myadmIP:agent=SA_mmbp,timeout=timeout:agent=SA_mmbr,timeout=timeout
CFNameX        :  Specify the CF node name of the cluster host.
weight         :  Specify the weight of the SF node.
myadmIP        :  Specify the IP address of the administrative LAN that is used 
                  by the shutdown facility of the cluster host.
                  It is not the IP address of MMB.
                  Available IP addresses are IPv4 and IPv6 addresses.
                  IPv6 link local addresses are not available.
                  When specifying the IPv6 address, enclose it in brackets "[ ]".
                  (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
                  If you specify a host name, please make sure it is listed in /etc/hosts.
SA_mmbp        :  Make sure to specify this shutdown agent that panics the node via MMB.
SA_mmbr        :  Make sure to specify this shutdown agent that resets the node via MMB.
timeout        :  Specify the timeout duration (seconds) of the shutdown agent.
                  Specify 20 seconds for "SA_mmbp" and "SA_mmbr".

Example:

node1,weight=2,admIP=fuji2:agent=SA_mmbp,timeout=20:agent=SA_mmbr,timeout=20
node2,weight=2,admIP=fuji3:agent=SA_mmbp,timeout=20:agent=SA_mmbr,timeout=20

Note

  • For the shutdown agents to be specified in the rcsd.cfg file, set both the SA_mmbp and SA_mmbr shutdown agents in that order.

  • Set the same contents in the rcsd.cfg file on all the nodes. Otherwise, a malfunction may occur.

Information

When creating the /etc/opt/SMAW/SMAWsf/rcsd.cfg file, you can use the /etc/opt/SMAW/SMAWsf/rcsd.cfg.mmb.template file as a template.

5.1.2.4.4 Starting the MMB Asynchronous Monitoring Daemon

Start the MMB asynchronous monitoring daemon.

Check that the MMB asynchronous monitoring daemon has been started on all the nodes.

# /etc/opt/FJSVcluster/bin/clmmbmonctl

If "The devmmbd daemon exists." is displayed, the MMB asynchronous monitoring daemon has been started.

If "The devmmbd daemon does not exist." is displayed, the MMB asynchronous monitoring daemon has not been started. Execute the following command to start the MMB asynchronous monitoring daemon.

# /etc/opt/FJSVcluster/bin/clmmbmonctl start
5.1.2.4.5 Setting I/O Completion Wait Time

When using any disks other than the ETERNUS disk array as the shared disk, to prevent the data error when the node is down due to a panic or other causes, set the time until I/O to the shared disk is completed.

Execute the command in any node that is part of the cluster system, and set the wait time until I/O completion (WaitForIOComp) during failover triggered by a node failure (panic, etc.).

For details about the "cldevparam" command, see the "cldevparam" manual page.

# /etc/opt/FJSVcluster/bin/cldevparam -p WaitForIOComp value
value        :  Specify the wait time until I/O completion.
                Specify the time checked by the procedure described in
                "5.1.2.4.1 Checking the Shutdown Agent Information."

After setting the wait time, execute the following command to check if the specified value is set.

# /etc/opt/FJSVcluster/bin/cldevparam -p WaitForIOComp
value

Note

If an I/O completion wait time is set, the failover time when a node failure (panic, etc.) occurs increases by that amount of time.

5.1.2.4.6 Starting the Shutdown Facility

On all the nodes, start or restart the shutdown facility

  1. Starting the shutdown facility

    Check that the shutdown facility has been started on all the nodes.

    # sdtool -s

    If the shutdown facility has already been started, execute the following commands on all the nodes to restart the shutdown facility.

    # sdtool -e
    # sdtool -b

    If the shutdown facility has not been started, execute the following command on all the nodes to start the shutdown facility.

    # sdtool -b
  2. Checking the status of the shutdown facility

    Check the status of the shutdown facility on all the nodes.

    # sdtool -s

Information

Display results of the sdtool -s command

  • If InitFailed is displayed in Init State, it means that a problem occurred during initialization of that shutdown agent.

  • If TestFailed is displayed in Test State, it means that a problem occurred while the agent was testing whether or not the node displayed in the Cluster Host field could be stopped. Some sort of problem probably occurred in the software, hardware, or network resources being used by that agent.

  • If Unknown is displayed in Shut State or Init State, it means that SF has not yet executed node stop, path testing, or SA initialization. Unknown is displayed temporarily in Test State and Init State until the actual status can be confirmed.

  • If TestFailed or InitFailed is displayed, check /var/log/messages. After the failure-causing problem is resolved and SF is restarted, the status display changes to InitWorked or TestWorked.

Note

  • If TestFailed is displayed in Test State and the message 7210 is output to /var/log/messages at the same time when "sdtool -s" is executed after the shutdown facility was started, possible causes are as follows:

    Make sure each setting is correctly set.

    7210 An error was detected in MMB. (node:nodename mmb_ipaddress1:mmb_ipaddress1 mmb_ipaddress2:mmb_ipaddress2 
    node_ipaddress1:node_ipaddress1 node_ipaddress2:node_ipaddress2 status:status detail:detail)
    • PSA/Svmco is not installed or not set.

    • A node is not restarted after installing Svmco manually.

    • Incorrect PSA/Svmco settings

      Example: An incorrect IP address (such as MMB IP address) is set to the IP address of the administrative LAN.

    • Necessary firewall to activate PSA/Svmco is not set.

    • Incorrect MMB settings

      Example 1: An incorrect IP address is set.

      Example 2: Both the virtual IP address of MMB and the physical IP address of MMB are not set.

  • If "sdtool -s" is executed immediately after the OS is started, TestFailed may be displayed in Test State for the local node. However, this state is displayed because the snmptrapd daemon is still being activated and does not indicate a malfunction. If "sdtool -s" is executed 10 minutes after the shutdown facility is started, TestWorked is displayed in Test State.

    In the following example, TestFailed is displayed in Test State for the local node (node1).

    # sdtool -s
    Cluster Host    Agent         SA State      Shut State  Test State  Init State
    ------------    -----         --------      ----------  ----------  ----------
    node1           SA_mmbp.so    Idle          Unknown     TestFailed  InitWorked
    node1           SA_mmbr.so    Idle          Unknown     TestFailed  InitWorked
    node2           SA_mmbp.so    Idle          Unknown     TestWorked  InitWorked
    node2           SA_mmbr.so    Idle          Unknown     TestWorked  InitWorked

    The following messages may be displayed in the Syslog right after the OS is started by same reason as previously described.

    3084: Monitoring another node has been stopped.
    SA SA_mmbp.so to test host nodename failed
    SA SA_mmbr.so to test host nodename failed

    These messages are also displayed because the snmptrapd daemon is being activated and does not indicate a malfunction. The following message is displayed in the Syslog 10 minutes after the shutdown facility is started.

    3083: Monitoring another node has been started.
  • If "sdtool -s" is executed when MMB asynchronous monitoring daemon is started for the first time, TestFailed may be displayed. This is a normal behavior because the settings are synchronizing between nodes. If "sdtool -s" is executed 10 minutes after the shutdown facility is started, TestWorked is displayed in Test State.

  • If nodes are forcibly stopped by the SA_mmbr shutdown agent, the following messages may be displayed in the Syslog. These are displayed because it takes time to stop the nodes and do not indicate a malfunction.

    Fork SA_mmbp.so(PID pid) to shutdown host nodename
            :
    SA SA_mmbp.so to shutdown host nodename failed
            :
    Fork SA_mmbr.so(PID pid) to shutdown host nodename
            :
    SA SA_mmbr.so to shutdown host nodename failed
            :
    MA SA_mmbp.so reported host nodename leftcluster, state MA_paniced_fsnotflushed
            :
    MA SA_mmbr.so reported host nodename leftcluster, state MA_paniced_fsnotflushed
            :
    Fork SA_mmbp.so(PID pid) to shutdown host nodename
            :
    SA SA_mmbp.so to shutdown host nodename succeeded

    If "sdtool -s" is executed after the messages above were displayed, KillWorked is displayed in Shut State for SA_mmbp.so. Then, KillFailed is displayed in Shut State for SA_mmbr.so.

    The following is the example of "sdtool -s" when the nodes (from node1 to node2) were forcibly stopped and the messages above were displayed.

    # sdtool -s
    Cluster Host    Agent         SA State      Shut State  Test State  Init State
    ------------    -----         --------      ----------  ----------  ----------
    node1           SA_mmbp.so    Idle          Unknown     TestWorked  InitWorked
    node1           SA_mmbr.so    Idle          Unknown     TestWorked  InitWorked
    node2           SA_mmbp.so    Idle          KillWorked  TestWorked  InitWorked
    node2           SA_mmbr.so    Idle          KillFailed  TestWorked  InitWorked

    To recover KillFailed displayed by "sdtool -s," perform the following procedure.

    # sdtool -e
    # sdtool -b
    # sdtool -s Cluster Host Agent SA State Shut State Test State Init State ------------ ----- -------- ---------- ---------- ---------- node1 SA_mmbp.so Idle Unknown TestWorked InitWorked node1 SA_mmbr.so Idle Unknown TestWorked InitWorked node2 SA_mmbp.so Idle Unknown TestWorked InitWorked node2 SA_mmbr.so Idle Unknown TestWorked InitWorked
5.1.2.4.7 Test for Forced Shutdown of Cluster Nodes

After setting up the shutdown facility, conduct a test for forced shutdown of cluster nodes to check that the correct nodes can be forcibly stopped.

For the detail of the test for forced shutdown of cluster nodes, refer to "1.4 Test."

5.1.2.5 Setup Procedure for Shutdown Facility in Virtual Machine Environment

This section describes the setup procedure of the shutdown facility in the virtual machine environment.

Note

When creating a redundant administrative LAN used in the shutdown facility by using GLS, set as below.

  • For taking over the IP address between nodes

    Configure GLS by using the logical IP address takeover function of the NIC switching mode.

    For shutdown facility, specify a physical IP address instead of a logical IP address.

  • For not taking over the IP address between nodes

    Configure GLS by using the physical IP address takeover function of the NIC switching mode.

5.1.2.5.1 Checking the Shutdown Agent Information

To forcibly stop the domain in the guest OS by using the shutdown facility in KVM environment, log in to the host OS via SSH.

Check in advance the following settings that are necessary for setting the shutdown facility.

For information on the user and password for logging in to the host OS, check the following information set up by the procedures described in the following sections:

Also take the following steps to check that the setting of the sudo command is already completed.

This is necessary for the confirmed user to execute the command as the root user.

  1. Execute the visudo command on all the nodes.

  2. Check that the following setting is described in the setting file displayed by executing the visudo command.

    <User ID>    ALL=(root) NOPASSWD: ALL

    If this setting information is missing, describe it to the file.

5.1.2.5.2 Setting up libvirt Shutdown Agent

Set up the libvirt shutdown agent.

Take the following steps.

Note

Be sure to perform the following operations from 1. to 3. on all guest OSes (nodes).

  1. Encrypt the password.

    Execute the sfcipher command to encrypt the password that was checked in "5.1.2.5.1 Checking the Shutdown Agent Information."

    For details on how to use the sfcipher command, see the manual page of "sfcipher."

    # sfcipher -c
    Enter User's Password:
    Re-enter User's Password:
    D0860AB04E1B8FA3
  2. Set up the panicky shutdown agent (SA_libvirtgp) and reset shutdown agent (SA_libvirtgr).

    Set up the panicky shutdown agent (SA_libvirtgp) and reset shutdown agent (SA_libvirtgr).

    Create the /etc/opt/SMAW/SMAWsf/SA_libvirtgp.cfg and the /etc/opt/SMAW/SMAWsf/SA_libvirtgr.cfg as below.

    Create the /etc/opt/SMAW/SMAWsf/SA_libvirtgp.cfg and the /etc/opt/SMAW/SMAWsf/SA_libvirtgr.cfg by using the root user privilege, and change the permission of the file to 600.

    CFNameX domainX ip-address user passwd
    CFNameX domainX ip-address user passwd
    CFNameX    : Specify the CF node name of the cluster host.
    domainX    : Specify the guest OS domain name.
                 Specify the domain name checked in 
                 "5.1.2.5.1 Checking the Shutdown Agent Information."
    ip-address : Specify the IP address of the host OS.
                 Specify the IP address of the host OS checked in 
                 "5.1.2.5.1 Checking the Shutdown Agent Information."
                 Available IP addresses are IPv4 and IPv6 addresses.
                 IPv6 link local addresses are not available.
    user       : User to log in to the host OS. 
                 Specify the user checked in 
                 "5.1.2.5.1 Checking the Shutdown Agent Information."
    passwd     : Password of the user specified by "user".
                 Specify the encrypted password that you have checked in 1.

    Example:

    When the guest OS domain name of node1 is domain1, and the IP address of the host OS on which node1 operates is 10.20.30.50.
    Moreover, when the guest OS domain name of node2 is domain2, and the IP address of the host OS on which node2 operates is 10.20.30.51.

    • /etc/opt/SMAW/SMAWsf/SA_libvirtgp.cfg

      node1 domain1 10.20.30.50 user D0860AB04E1B8FA3
      node2 domain2 10.20.30.51 user D0860AB04E1B8FA3
    • /etc/opt/SMAW/SMAWsf/SA_libvirtgr.cfg

      node1 domain1 10.20.30.50 user D0860AB04E1B8FA3
      node2 domain2 10.20.30.51 user D0860AB04E1B8FA3

    Note

    • Check if the setting contents of the /etc/opt/SMAW/SMAWsf/SA_libvirtgp.cfg file and the /etc/opt/SMAW/SMAWsf/SA_libvirtgr.cfg file are correct. If there is an error in the setting contents, the shutdown facility cannot be performed normally.

    • Check if the domain name (domainX) of the guest OS and the IP address (ip-address) of the host OS corresponding to the cluster host's CF node name (CFNameX) of the /etc/opt/SMAW/SMAWsf/SA_libvirtgp.cfg file and the /etc/opt/SMAW/SMAWsf/SA_libvirtgr.cfg file are set. If there is an error in the setting, a different node may be forcibly stopped.

    • The contents of the SA_libvirtgp.cfg, SA_libvirtgr.cfg, and rcsd.cfg files of all guest OSes (nodes) should be identical. If not, a malfunction will occur.

  3. Log in to the host OS

    The shutdown facility accesses the host OS with SSH. Therefore, you need to authenticate yourself (create the RSA key) in advance, which is required when using SSH for the first time.
    On all guest OSes (nodes), log in to each host OS IP address (ip-address) set in the step 2. using each set user.

    Execute the command as the root user access privilege.

    # ssh -l user XXX.XXX.XXX.XXX
    The authenticity of host 'XXX.XXX.XXX.XXX (XXX.XXX.XXX.XXX)' can't be established.
    RSA key fingerprint is xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx.
    Are you sure you want to continue connecting (yes/no)? yes <- "Enter yes."
    #
5.1.2.5.3 Setting Up vmchkhost Shutdown Agent

When using the host OS failover function, set up the vmchkhost shutdown agent.

Perform this setting after setting up the libvirt shutdown agent.

Note

Be sure to perform the following operations from 2. to 3. on all guest OSes (nodes).

  1. Set up the libvirt shutdown agent and check the information of the host OS.

    Check the following information that are set in the libvirt shutdown agent:

    • IP address for the host OS

    • User for logging in to the host OS

    • Encrypted user password for logging in to the host OS

    Also check that the following information for the host OS.

    • CF node name

  2. Set up the vmchkhost shutdown agent.

    Create /etc/opt/SMAW/SMAWsf/SA_vmchkhost.cfg as described in the following.

    Create the SA_vmchkhost.cfg using the root user access privilege and change the permission of the file to 600.

    guest-cfnameX host-cfnameX ip-address user password
    guest-cfnameX host-cfnameX ip-address user password
    guest-cfnameX      : CF node name of the guest OS (cluster node).
    host-cfnameX       : CF node name of the host OS.
                         Specify the CF node name checked in step 1.
    ip-address         : An IP address of the host OS.
                         Specify the IP address checked in step 1.
    user               : User to log in to the host OS.
                         Specify the user checked in step 1.
    password           : Password of the user specified by "user".
                         Specify the encrypted password checked in 1. 

    Example:

    When the CF node name of the host OS on which node1 (CF node name of the guest OS) operates is hostos1, the IP address of the host OS is 10.20.30.50, the CF node name of the host OS on which node2 (CF node name of the guest OS) operates is hostos2, and the IP address of the host OS is 10.20.30.51.

    node1 hostos1 10.20.30.50 user D0860AB04E1B8FA3
    node2 hostos2 10.20.30.51 user D0860AB04E1B8FA3

    Note

    • Check if the setting contents of the /etc/opt/SMAW/SMAWsf/SA_vmchkhost.cfg file are correct. If there is an error in the setting contents, the shutdown facility cannot be performed normally.

    • Check if the CF node name of the host OS (host-cfnameX) and the IP address of the host OS (ip-address) corresponding to the CF node name (guest-cfnameX) of the guest OS (clutser host) of the /etc/opt/SMAW/SMAWsf/SA_vmchkhost.cfg file are set. If there is an error in the setting, the shutdown facility cannot be performed normally.

    • The contents of the SA_vmchkhost.cfg file of all guest OSes (nodes) should be identical. If not, a malfunction will occur.

  3. Log in to the host OS

    The shutdown facility accesses the host OS with SSH. Therefore, you need to authenticate yourself (create the RSA key) in advance, which is required when using SSH for the first time.

    Check that you have already authenticated yourself (created the RSA key) when setting up the libvirt shutdown agent.

5.1.2.5.4 Setting up the Shutdown Daemon

On all the nodes, create /etc/opt/SMAW/SMAWsf/rcsd.cfg with the following information.

Create the rcsd.cfg file using root user access privileges and change the permission of the file to 600.

CFNameX        :  Specify the CF node name of the cluster host.
weight         :  Specify the weight of the SF node.
myadmIP        :  Specify the IP address of the administrative LAN that used by the Shutdown 
                  Facility of the cluster host.
                  It is not the IP address of iRMC or the management blade.
                  Available IP addresses are IPv4 and IPv6 addresses.
                  IPv6 link local addresses are not available.
                  When specifying the IPv6 address, enclose it in brackets "[ ]".
                  (Example: [1080:2090:30a0:40b0:50c0:60d0:70e0:80f0])
                  If you specify a host name, please make sure it is listed in /etc/hosts.
SA_libvirtgp   :  Make sure to set this shutdown agent that panics the guest OS.
SA_mmbr        :  Make sure to set this shutdown agent that resets the guest OS.
SA_vmchkhost   :  Shutdown agent for the host OS failover function. 
timeout        :  Specify the timeout duration (seconds) of the shutdown agent.
                  Specify 35 seconds for SA_libvirtgp, SA_libvirtgr, and SA_vmchkhost.

Example1: When using the host OS failover function

node1,weight=2,admIP=fuji2:agent=SA_libvirtgp,timeout=35:agent=SA_libvirtgr,timeout=35:agent=SA_vmchkhost,timeout=35
node2,weight=1,admIP=fuji3:agent=SA_libvirtgp,timeout=35:agent=SA_libvirtgr,timeout=35:agent=SA_vmchkhost,timeout=35

Example 2: When not using the host OS failover function

node1,weight=2,admIP=fuji2:agent=SA_libvirtgp,timeout=35:agent=SA_libvirtgr,timeout=35
node2,weight=1,admIP=fuji3:agent=SA_libvirtgp,timeout=35:agent=SA_libvirtgr,timeout=35

Note

  • SA_libvirtgp shutdown agent must be set first followed by SA_libvirtgr, and then set SA_vmchkhost as the last of all in the rcsd.cfg file.

  • Set the same contents in the rcsd.cfg file on all the nodes. Otherwise, a malfunction may occur.

Information

When creating the /etc/opt/SMAW/SMAWsf/rcsd.cfg file, you can use the /etc/opt/SMAW/SMAWsf/rcsd.cfg.mmb.template file as a template.

5.1.2.5.5 Starting the Shutdown Facility

Start or restart the shutdown facility on all the nodes.

  1. Starting the shutdown facility

    Check that the shutdown facility has been started on all the nodes.

    # sdtool -s

    If the shutdown facility has already been started, execute the following commands to restart the shutdown facility on all the nodes.

    # sdtool -e
    # sdtool -b

    If the shutdown facility has not been started, execute the following command to start the shutdown facility on all the nodes.

    # sdtool -b
  2. Checking the status of the shutdown facility

    Check the status of the shutdown facility on all the nodes.

    # sdtool -s

Information

Display results of the sdtool -s command
  • If InitFailed is displayed in Init State, it means that a problem occurred during initialization of that shutdown agent.

  • If TestFailed is displayed in Test State, it means that a problem occurred while the agent was testing whether or not the node displayed in the Cluster Host field could be stopped. Some sort of problem probably occurred in the software, hardware, or network resources being used by that agent.

  • If Unknown is displayed in Shut State or Init State, it means that SF has not yet executed node stop, path testing, or SA initialization. Unknown is displayed temporarily in Test State and Init State until the actual status can be confirmed.

  • If TestFailed or InitFailed is displayed, check the following files:

    • /var/log/messages

    • /etc/sysconfig/libvirt-guests

    For /etc/sysconfig/libvirt-guests, check whether the following settings are made:

    After the failure-causing problem is resolved and SF is restarted, the status display changes to InitWorked or TestWorked.

5.1.2.5.6 Setting up the Host OS Failover Function to the Host OS (PRIMEQUEST only)

When using the host OS failover function in PRIMEQUEST, for linking with MMB asynchronous monitoring function, configure the host OS failover function to the host OS.

Set up this setting after setting libvirt shutdown agent and vmchkhost shutdown agent.

Note

Be sure to perform the following operations from 3. to 7. on all the host OSes (nodes).

  1. Check the setting information.

    The host OS failover function in PRIMEQUEST, when detecting an host OS error by MMB asynchronous monitoring function, logs in to a guest OS (a cluster node) using SSH and then notifies the shutdown facility of the host OS error.

    For setting the host OS failover function, confirm the following necessary information previously.

    • IP address of the guest OS

    • Domain name of the guest OS

    • Cluster name of the guest OS

    • CF node name of the guest OS

  2. Create the user (when logging in to the guest OS not as a root user).

    When the host OS failover function logs in to the guest OS not as a root user, a user for logging in is created. Perform the following procedure on all the guest OS.

    (1) Create the login user.

    Set the user password with seven-bit ASCII characters except the following characters.

    >  <  "  /  \  =  !  ?  ;  ,  &

    (2) Set the sudo command so that the created user can execute the command as a root user.

    Execute the visudo command by using the root command. Describe the following setting in the displayed setting file.

    <User created in (1)>    ALL=(root) NOPASSWD: ALL
  3. Encrypt the password.

    Execute the sfcpher command to encrypt passwords for login to the guest OS as a root user.

    For details on how to use the sfcipher command, see the manual page of "sfcipher."

    # sfcipher -c
    Enter User's Password:
    Re-enter User's Password:
    D0860AB04E1B8FA3
  4. Create /etc/opt/FJSVcluster/etc/kvmguests.conf.

    Create /etc/opt/FJSVcluster/etc/kvmguests.conf with the following contents.

    Create the kvmguests.conf file using the root user access privilege and change the permission of the file to 600.

    When multiple guest OSes (the cluster nodes) are operating on a host OS that configures the cluster, describe all the guest OSes configured the host OS failover function in this file.

    guest-name host-cfname guest-clustername guest-cfname guest_IP guest_user guest_passwd
              :
    • Enter the information of one node in one line.

    • Delimit each item with a single space.

    • The kvmguests.conf file must be the same on all cluster nodes.

    guest-name         :Specify the domain name of the guest OS.
    host-cfname        :Specify the CF node name of the host OS in which "guest-name" is running.
                        If you execute "cftool -l" on the host OS in which "guest-name" is running,
                        you can confirm the CF node name of the node.
    guest-clustername  :Specify the cluster name of the guest OS.
                        If you execute "cftool -c" on the guest OS, you can confirm the cluster
                        name of the node.
    guest-cfname       :Specify the CF node name of the guest OS.
                        If you execute "cftool -l" on the guest OS, you can confirm the CF node
                        name of the node.
    guest_IP           :Specify the IP address of the guest OS.
                        Available IP address formats are IPv4 and IPv6 addresses.
                        IPv6 link local addresses are not available.
    guest_user         :Specify the user for logging in to the guest OS using SSH.
                        Specify the fixed root or the user created in step 2.
    guest_passwd       :Specify the user password for logging in to the guest OS.
                        Specify the password encrypted in step 3.

    Example: In a two-node configuration between guest OSes, two cluster systems are configured

    guest11 cfhost1 cluster1 cfguest11  10.20.30.50 root D0860AB04E1B8FA3
    guest12 cfhost2 cluster1 cfguest12  10.20.30.51 root D0860AB04E1B8FA3
    guest21 cfhost1 cluster2 cfguest21  10.20.30.60 root D0860AB04E1B8FA3
    guest22 cfhost2 cluster2 cfguest12  10.20.30.61 root D0860AB04E1B8FA3
  5. Confirm the log in to the guest OS

    The host OS failover function in PRIMEQUEST accesses the guest OS with SSH. Therefore, you need to authenticate yourself (create the RSA key) in advance, which is required when using SSH for the first time.

    Check that you can connect to all the guest OSes (nodes) which are specified to /etc/opt/FJSVcluster/etc/kvmguests.conf via SSH as a root user.

    # ssh -l user1 XXX.XXX.XXX.XXX
    The authenticity of host 'XXX.XXX.XXX.XXX (XXX.XXX.XXX.XXX)' can't be established.
    RSA key fingerprint is xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx:xx.
    Are you sure you want to continue connecting (yes/no)? yes <- Enter "yes."
  6. Check the setting in /etc/opt/FJSVcluster/etc/kvmguests.conf.

    Execute the sfkvmtool command on all the host OSes to make sure that the settings in /etc/opt/FJSVcluster/etc/kvmguests.conf are correct.

    If the settings are correct, the following message is output.

    # /opt/SMAW/SMAWsf/bin/sfkvmtool -c
    NOTICE: The check of configuration file succeeded.

    If a message other than above is output, review the setting in /etc/opt/FJSVcluster/etc/kvmguests.conf.

  7. Start the shutdown facility

    Check that the shutdown facility has already been started on all the nodes.

    # sdtool -s

    If the shutdown facility has already been started, execute the following on all the nodes to restart it.

    # sdtool -e
    # sdtool -b

    If the shutdown facility has not been started, execute the following on all the nodes to start it.

    # sdtool -b
5.1.2.5.7 Test for Forced Shutdown of Cluster Nodes

After setting up the shutdown facility, conduct a test for forced shutdown of cluster nodes to check that the correct nodes can be forcibly stopped.

For the detail of the test for forced shutdown of cluster nodes, refer to "1.4 Test."

Note

After shutting down a node (a guest OS) forcibly by SA_libvirtgp, the guest OS may be a temporary stopped state. (For example, when there is no space in /var/crash on the host OS.) In the case, forcibly shutdown the guest OS by the virsh destroy command.