Sun Microsystems, Inc.  Oracle System Handbook - ISO 7.0 May 2018 Internal/Partner Edition
   Home | Current Systems | Former STK Products | EOL Systems | Components | General Info | Search | Feedback

Asset ID: 1-79-1934130.1
Update Date:2016-08-10
Keywords:

Solution Type  Predictive Self-Healing Sure

Solution  1934130.1 :   How To Recover SuperCluster Components that were Backed Up with osc-config-backup  


Related Items
  • Oracle SuperCluster M7 Hardware
  •  
  • Solaris Operating System
  •  
  • Oracle SuperCluster T5-8 Half Rack
  •  
  • Oracle SuperCluster T5-8 Full Rack
  •  
Related Categories
  • PLA-Support>Eng Systems>Exadata/ODA/SSC>SPARC SuperCluster>DB: SuperCluster_EST
  •  




In this Document
Purpose
Scope
Details
 Overview
 Possible recovery scenarios:
 Requirements
 Known Issues and Workarounds
 Recover Domains Using ZFS streams
References


Applies to:

Oracle SuperCluster M7 Hardware - Version All Versions to All Versions [Release All Releases]
Solaris Operating System - Version 11.3 and later
Oracle SuperCluster T5-8 Full Rack - Version All Versions to All Versions [Release All Releases]
Oracle SuperCluster T5-8 Half Rack - Version All Versions to All Versions [Release All Releases]
Oracle Solaris on SPARC (64-bit)

Purpose

This article describes how to recover SuperCluster components that were backed up with the osc-config-backup tool. It also highlights critical issues and workarounds which SuperCluster administrators should be aware of before attempting a recovery scenario.

Scope

The recovery process is only currently supported on T5-8 SuperCluster and M7 SuperCluster. Detailed instructions and examples that you can use to recover SuperCluster components previously backed up by osc-config-backup can be found in the Oracle SuperCluster Configuration Backup Utility - Recovery Guide.

Please note that data storage on the Exadata Storage Servers is not covered by osc-config-backup, and this article does not cover storage server recovery. Use the Engineered Systems Backup Utility (ESBU) to back up and recover this data. Refer to the article titled RMAN Backup From SPARC SuperCluster to Sun ZFS Backup Appliance (Doc ID 1517107.1)

For further details and best practices, refer to the whitepaper titled Oracle Optimized Solution for Backup and Recovery of Oracle SuperCluster.

Details

Overview

When osc-config-backup is used to backup SuperCluster components prior to a failure, you can use the osc-config-backup data to recover SuperCluster components should the need arise. During the backup process, osc-config-backup copies critical component data to a central location on the internal ZFS Storage Appliance which in the event of an incident can be used to recover from a number of components.

Components that can be recovered using osc-config-backup data:

  • Domains – Configuration, ZFS pools (rpool, bpool if present, and u01-pool for Database Domains)
  • Solaris zones – Configuration, ZFS pools (rpool and u01-pool for Database Zones)
  • IB switches – Configuration
  • Ethernet management switch – Configuration
  • ZFS storage appliance – Configuration
  • SuperCluster – Configuration information (ssc-cu.zip file)
  • Explorer – Data from each Database and Application Domain

Possible recovery scenarios:

These recovery scenarios cover some situations with which customers could potentially find themselves having to recover from. In most situations where a single component failure has occured in the rack, no loss of service should occur assuming the configuration has been implemended with maximum availability in mind. In some situations the loss of an entire SuperCluster platform might not have too much impact on a business as it's becoming increasing common for customer's to implement a secondary DR SuperCluster platform.

  • Restoring an Oracle Solaris Zone
  • Rolling Back an I/O Domain
  • Restore a Domain Configuration
  • Restore an IB Switch
  • Restore the Ethernet Management Switch
  • Restore an Oracle ILOM Configuration

For complex situations or if customers are unsure of the recovery situation then please raise a support ticket for further assistance.

Requirements

To perform the recovery procedures in this article, ensure that these requirements are met:

SuperCluster components must have been backed up with the osc-config-backup tool prior to the disaster. Refer to the article titled How To Back Up the Oracle SuperCluster Configuration with osc-config-backup (Doc ID 1934129.1).

  • You must have access to all of the data that was collected by osc-config-backup. This includes Explorer data, OS images for domains and zones, IB switch and Ethernet management switch configuration data.
  • The ZFS storage appliance contains the osc-config-backup data and must be functional, or you must have access to your ZFS storage appliance backups.
  • All hardware components of the SuperCluster that you plan to recover must be functional with no faults.
  • You must have access to SuperCluster's Explorer data that was collected prior to the disaster. You might need to access the Explorer data from your ZFS storage appliance backup.
  • You must have a proxy system, such as a laptop, with network access to the SuperCluster ZFS storage appliance and to the node you are restoring. The proxy system must be able to ssh to SuperCluster.

Known Issues and Workarounds

If you have opted to not backup the SuperCluster platform using iscsi luns then the only available recovery process is to use the zfs stream dumps collected within the /osc-config-backup mountpoint. A domain can be fully recovered by both methods but the zfs stream recovery process is slightly more involved that following the documented iscsi recovery methodology.

Recover Domains Using ZFS streams

This additional step only applies to T5-8 SuperCluster platforms and you only need to perform this task if you were not able to restore the OS via an iSCSI backup or you opted not to take iSCSI backups in the first place. This proceedure assumes the following conditions:

1/ Boot the domain using a CDROM image into single user mode

{0} ok boot cdrom -s

NOTE: Username=root password=solaris

2/ Based on Explorer data, create the pools using the first two internal drives (HDD0 labeled as solaris, and HDD1). Run the format utility to identify the drive's cntndnsn numbers (such as c0t5000CCA02523F5F4d0), then quit the utility.

root@solaris:~# format
0. c0t5000CCA02523F5F4d0 <HITACHI-H109090SESUN900G-A31A cyl 38553 alt 2 hd 24 sec 1900> solaris
/scsi_vhci/disk@g5000cca02523f5f4
1. c0t5000CCA02523F968d0 <HITACHI-H109090SESUN900G-A31A cyl 38553 alt 2 hd 24 sec 1900> solaris
/scsi_vhci/disk@g5000cca02523f968

NOTE: if the physical drives have been replaced then please ensure the partition table matches the original drive. Also be aware that I/O domains on T5-8 SuperCluster only have a single disk supplied by the primary control domain from the ZFS-SA.

Run this command for Oracle Solaris 11 domains:

root@solaris# zpool create -f rpool mirror c0t5000CCA02523F5F4d0s0 c0t5000CCA02523F968d0s0

Run this command for Oracle Solaris 10 domains.

# zpool create -f -o failmode=continue -R /a -m legacy -o cachefile=/etc/zfs/zpool.cache rpool mirror c0t5000CCA02523F5F4d0s0 c0t5000CCA02523F968d0s0 

Verify the zpools:

root@solaris:~# zpool list
NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
rpool 278G 85K 278G 0% 1.00x ONLINE -

root@solaris:~# zpool status
pool: rpool
state: ONLINE
scan: none requested
config:

NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
mirror-0 ONLINE 0 0 0
c0t5000CCA02523F5F4d0s0 ONLINE 0 0 0
c0t5000CCA02523F968d0s0 ONLINE 0 0 0

errors: No known data errors

Create any additional pools you have on this domain.

3/ Mount the ZFS backup directory on the guest domain.

Assign a temporary network address on the same management subnet of the ZFS storage appliance on the guest domain.

For example, if the ZFS storage appliance management IP address is 10.163.238.110, you can assign an IP address of 10.163.238.10 to the guest domain.

root@solaris~# ipadm create-ip net3
root@solaris~# ipadm create-addr -T static -a local=10.10.10.2/24 net3/v4 x
root@solaris~# mount -o vers=3 10.10.10.1:/export/osc-config-backup/backup /backup

4/ Restore the rpool from the ZFS snapshots.

root@solaris~# cd /backup/domain/<domainname>
root@solaris:/backup/domain/<domainname># ls <domainname>.rpool.backup.2014.07.18.11.17.zfs
root@solaris:/backup/domain/<domainname># cat <domainname>.rpool.backup.2014.07.18.11.17.zfs | zfs receive -Fdu rpool
root@solaris:/backup/domain/<domainname># zpool list
NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
rpool 278G 11.9G 266G 4% 1.00x ONLINE -

5/ Activate the BE and boot from the rpool disk.

Oracle Solaris 11 example:

root@solaris:/backup/domain/<domainname># beadm list
be_find_current_be: failed to find current BE name
BE Active Mountpoint Space Policy Created
-- ------ ---------- ----- ------ -------
S11.1_SRU7.5_idrs - - 17.10G static 2014-02-18 09:26
solaris - - 160.68M static 2014-02-18 09:32

root@solaris:/backup/domain/<domainname># beadm activate S11.1_SRU7.5_idrs
be_find_current_be: failed to find current BE name

root@solaris:/backup/domain/<domainname># beadm list
be_find_current_be: failed to find current BE name
BE Active Mountpoint Space Policy Created
-- ------ ---------- ----- ------ -------
S11.1_SRU7.5_idrs R - 17.11G static 2014-02-18 09:26
solaris - - 160.68M static 2014-02-18 09:32
Oracle Solaris 10 example:

root@solaris:/backup/domain/<domainname># zpool set bootfs=rpool/ROOT/5.10 rpool
root@solaris:/backup/domain/<domainname># installboot -F zfs /usr/platform/'uname -i'/lib/fs/zfs/bootblk /dev/rdsk/c0t5000CCA02523F5F4d0s0
root@solaris:/backup/domain/<domainname># reboot

The guest domain is up and running.

6/ Create swap and dump file systems.

Example:

root@<domainname>:~# zfs create -V 48G rpool/swap
root@<domainname>:~# zfs create -V 64G rpool/dump
root@<domainname>:~# zfs list rpool/dump
NAME USED AVAIL REFER MOUNTPOINT
rpool/dump 66.0G 212G 16K -

root@<domainname>:~# zfs list rpool/swap
NAME USED AVAIL REFER MOUNTPOINT
rpool/swap 49.5G 196G 16K -

root@<domainname>:~# swap -a /dev/zvol/dsk/rpool/swap
operating system crash dump was previously disabled --
invoking dumpadm(1M) -d swap to select new dump device
dumpadm: no swap devices could be configured as the dump device

root@<domainname>:~# dumpadm -d /dev/zvol/dsk/rpool/dump
Dump content: kernel pages
Dump device: /dev/zvol/dsk/rpool/dump (dedicated)
Savecore directory: /var/crash
Savecore enabled: yes
Save compressed: on

The guest domain is now restored and operational.

 

References

<NOTE:1934129.1> - How To Back Up the Oracle SuperCluster Configuration with osc-config-backup
<NOTE:1567979.1> - Oracle SuperCluster Supported Software Versions - All Hardware Types

Attachments
This solution has no attachment
  Copyright © 2018 Oracle, Inc.  All rights reserved.
 Feedback