Top
PRIMECLUSTER  Installation and Administration Guide4.5
FUJITSU Software

17.3.1 Performing Live Migration of the Cluster on a Control Domain

17.3.1.1 Operation before Performing Live Migration

This section describes the operation before performing the Live Migration in an Oracle VM Server for SPARC Environment.

Note

  1. Changing the cluster configuration (control domain)

    Change the cluster configuration before performing the Live Migration.

    Execute the following command on the cluster nodes of the source control domain.

    # /etc/opt/FJSVcluster/bin/clovmmigrate -p source-ldom
    source-ldom

    The target guest domain name to be migrated

    If you execute this command, the following cluster configurations are changed on all nodes in the cluster between guests that are specified for source-ldom.

    • The timeout value of the CF cluster interconnect (current setting value [default 10 seconds] to 600 seconds)

    • Stopping the shutdown facility

    Note

    If the live migration was canceled or failed then the guest domain stopped, implement the following recovery procedure. If the guest domain does not stop, perform "17.3.1.2 Operation after Performing Live Migration" first and then implement "17.3.1.1 Operation before Performing Live Migration."

    1. Execute the cftool -n command on a node that has not been migrated, then check that the target guest domain has become DOWN. After executing the clovmmigrate command, it takes about 600 seconds before the status changes to DOWN.

    2. Start the target guest domain to be migrated.

    3. Perform "17.3.1.2 Operation after Performing Live Migration."

    When the above recovery procedure is not performed and the guest domain is started which stopped before the guest domain to be migrated becomes DOWN status, the start of SF, CRM and RMS fail on the target guest domain to be migrated. Stop the guest domain again and perform the above recovery procedure.

17.3.1.2 Operation after Performing Live Migration

This section describes the operation after performing the Live Migration in an Oracle VM Server for SPARC Environment.

Note

  • A cluster system will not be switched until this operation is completed.

  • Perform step 1 on a control domain.

  • Perform step 2 on a guest domain.

  • Before performing this operation, the configuration information of the saved logical domain needs to be less than six or less. Execute the following command to check it. Also, check that "config_tmp" does not exist in the saved configuration information name of the logical domains.

    # ldm list-spconfig

    If the configuration information of the saved logical domain is 7 or more, execute the following command to delete the configuration information to make them 6 or less.

    # ldm remove-spconfig configuration name

    If "config_tmp" exists in the configuration name of the logical domain, execute the following command to delete it.

    # ldm remove-spconfig config_tmp

    For details on the configuration information of the logical domain, see "Operations and Commands Related to Logical Domain Configurations" in "Fujitsu SPARC M12 and Fujitsu M10/SPARC M10 Domain Configuration Guide."

  1. Changing the cluster configuration (control domain)

    Change the cluster configuration after performing the Live Migration.

    Execute the following command on the cluster nodes of the source control domain:

    # /etc/opt/FJSVcluster/bin/clovmmigrate -u source-ldom target-host
    source-ldom

    The guest domain name to be migrated

    target-host

    The host name registered to the IP address or the /etc/inet/hosts file of the destination control domain

    Even if a guest domain is not migrated by such as cancelling the Migration, you need to specify a control domain. To do so, specify the host name registered to the IP address on the source control domain or the /etc/inet/hosts file.

    If you execute this command, the following cluster configurations are changed on all nodes in the cluster between guests that are specified for source-ldom. In addition, the configuration information of the logical domains is saved both by the source control domain and the destination control domain.

    • The timeout value of the CF cluster interconnect (600 seconds to the setting value before Migration [default 10 seconds])

    • Change the setting of the shutdown facility (IP address of XSCF-LAN#0, IP address of XSCF-LAN#1, and the SF weight)

    • Starting the shutdown facility

  2. Checking the state of the shutdown facility (guest domain)

    Execute the following commands on each node of a guest domain to check if the cluster is configured correctly after the Live Migration.

    # /etc/opt/FJSVcluster/bin/clsnmpsetup -l
    # /opt/SMAW/bin/sdtool -s

    Note

    If TestFailed or InitFailed is displayed, the setting of the shutdown facility may not be changed.
    If the message 7245 is output in the /var/adm/messages file, the first SSH connection to XSCF has not completed yet. For details, see "PRIMECLUSTER Messages."

    For all other circumstances, go back to step 1 and try again.

    Example: When the Migration is performed for guest 2 in the two-node cluster between guest domains ("Figure 17.1 Cluster configuration example")

    guest2 # /etc/opt/FJSVcluster/bin/clsnmpsetup -l
    device-name cluster-host-name  PPAR-ID  domain-name  IP-address1  IP-address2  user-name  connection-type
    -------------------------------------------------------------------------------------------------------------------
    xscf        cfguest1             0        guest1       10.20.30.71  10.20.40.71  xuser   ssh
    xscf        cfguest2             1        guest2       10.20.30.73  10.20.40.73  xuser   ssh
    guest2 #                         ^^                    ^^^^^^^^^^^^^^^^^^^^^^^^
    guest2 #                                                The target XSCF IP address to be migrated
    guest2 #                        The target PPAR-ID to be migrated
    guest2 # /opt/SMAW/bin/sdtool -s
    Cluster Host      Agent                SA State      Shut State  Test State  Init State
    ------------      -----                --------      ----------  ----------  ----------
    cfguest1          SA_xscfsnmpg0p.so    Idle          Unknown     TestWorked  InitWorked
    cfguest1          SA_xscfsnmpg1p.so    Idle          Unknown     TestWorked  InitWorked
    cfguest1          SA_xscfsnmpg0r.so    Idle          Unknown     TestWorked  InitWorked
    cfguest1          SA_xscfsnmpg1r.so    Idle          Unknown     TestWorked  InitWorked
    cfguest2          SA_xscfsnmpg0p.so    Idle          Unknown     TestWorked  InitWorked
    cfguest2          SA_xscfsnmpg1p.so    Idle          Unknown     TestWorked  InitWorked
    cfguest2          SA_xscfsnmpg0r.so    Idle          Unknown     TestWorked  InitWorked
    cfguest2          SA_xscfsnmpg1r.so    Idle          Unknown     TestWorked  InitWorked
    guest2 #

Note

After performing the Migration, the following message which indicates the time is not synchronized between the cluster nodes may be printed in the switchlog or /var/adm/messages file.

(WRP, 34) Cluster host <host> is no longer in time sync with local node. Sane operation of RMS can no longer be guaranteed.
Further out-of-sync messages will appear in the syslog.

If this situation continues, the following message may be periodically printed in the /var/adm/messages file.

(WRP, 35) Cluster host <host> is no longer in time sync with local node. Sane operation of RMS can no longer be guaranteed.

This message stops once the time is synchronized. For details on the messages, see "PRIMECLUSTER Messages."